Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
5,715
8,179,257,849
IssuesEvent
2018-08-28 15:50:51
Microsoft/ChakraCore
https://api.github.com/repos/Microsoft/ChakraCore
closed
Include function name in built-in function toString (Intl functions)
Compatibility Dev Experience Intl
See: https://github.com/Microsoft/ChakraCore/issues/3204#issuecomment-313806817 The current spec (and proposed, see [proposal](https://tc39.github.io/Function-prototype-toString-revision/#sec-function.prototype.tostring) also linked from the comment above) require that a the toString of a built-in function object matches the NativeFunction grammar, which states that the function name is optional (can be blank) or "the portion of the returned String that would be matched by IdentifierName must be the initial value of the name property of func" Currently, we omit the name. But I would maintain that it would be a better developer experience if we included the name. Additionally, all the Intl constructors report their name while all (AFAICT) functions omit their name from the toString. @bterlson: opinions? /cc @atulkatti
True
Include function name in built-in function toString (Intl functions) - See: https://github.com/Microsoft/ChakraCore/issues/3204#issuecomment-313806817 The current spec (and proposed, see [proposal](https://tc39.github.io/Function-prototype-toString-revision/#sec-function.prototype.tostring) also linked from the comment above) require that a the toString of a built-in function object matches the NativeFunction grammar, which states that the function name is optional (can be blank) or "the portion of the returned String that would be matched by IdentifierName must be the initial value of the name property of func" Currently, we omit the name. But I would maintain that it would be a better developer experience if we included the name. Additionally, all the Intl constructors report their name while all (AFAICT) functions omit their name from the toString. @bterlson: opinions? /cc @atulkatti
non_process
include function name in built in function tostring intl functions see the current spec and proposed see also linked from the comment above require that a the tostring of a built in function object matches the nativefunction grammar which states that the function name is optional can be blank or the portion of the returned string that would be matched by identifiername must be the initial value of the name property of func currently we omit the name but i would maintain that it would be a better developer experience if we included the name additionally all the intl constructors report their name while all afaict functions omit their name from the tostring bterlson opinions cc atulkatti
0
66,755
8,967,346,621
IssuesEvent
2019-01-29 02:56:18
SAP/fundamental-react
https://api.github.com/repos/SAP/fundamental-react
opened
Add examples of importing components from the dist/ directory to docs
documentation enhancement
### Description After doing some research on tree shaking, I think the best way forward is to provide examples of how to import individual components from the `dist` folder if consumers want to reduce their library size. ex: ```javascript import {Alert} from 'react-fundamentals/dist/Alert/Alert' ``` ### Versions **fundamental-react:** --- _**NOTE:** Where applicable, please include uncropped screen captures._ _**DISCLAIMER:** After triaging an issue, the fundamental-react team will see if it can be reproduced or confirmed. If more information is needed, the fundamental-react team will contact the author. Any issues awaiting responses from the author for more than 7 days will be closed. The author can re-open the issue at a later time if they can present the requested information._
1.0
Add examples of importing components from the dist/ directory to docs - ### Description After doing some research on tree shaking, I think the best way forward is to provide examples of how to import individual components from the `dist` folder if consumers want to reduce their library size. ex: ```javascript import {Alert} from 'react-fundamentals/dist/Alert/Alert' ``` ### Versions **fundamental-react:** --- _**NOTE:** Where applicable, please include uncropped screen captures._ _**DISCLAIMER:** After triaging an issue, the fundamental-react team will see if it can be reproduced or confirmed. If more information is needed, the fundamental-react team will contact the author. Any issues awaiting responses from the author for more than 7 days will be closed. The author can re-open the issue at a later time if they can present the requested information._
non_process
add examples of importing components from the dist directory to docs description after doing some research on tree shaking i think the best way forward is to provide examples of how to import individual components from the dist folder if consumers want to reduce their library size ex javascript import alert from react fundamentals dist alert alert versions fundamental react note where applicable please include uncropped screen captures disclaimer after triaging an issue the fundamental react team will see if it can be reproduced or confirmed if more information is needed the fundamental react team will contact the author any issues awaiting responses from the author for more than days will be closed the author can re open the issue at a later time if they can present the requested information
0
19,300
25,466,482,337
IssuesEvent
2022-11-25 05:16:26
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[IDP] [PM] Able to create new admin account in the following scenario
Bug Blocker P0 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
**Pre-condition:** mfa should be enabled for the PM **Steps:** 1. Login to PM 2. Click on 'Admins' tab 3. Click on 'Add new admin' button 4. Try to add admin without entering phone number field and Verify **AR:** Able to create new admin account in the PM **ER:** New admin should not created in the PM, without completing the phone number field **Note:** Phone number field should be a mandatory field in the application, if mfa is enabled for the PM
3.0
[IDP] [PM] Able to create new admin account in the following scenario - **Pre-condition:** mfa should be enabled for the PM **Steps:** 1. Login to PM 2. Click on 'Admins' tab 3. Click on 'Add new admin' button 4. Try to add admin without entering phone number field and Verify **AR:** Able to create new admin account in the PM **ER:** New admin should not created in the PM, without completing the phone number field **Note:** Phone number field should be a mandatory field in the application, if mfa is enabled for the PM
process
able to create new admin account in the following scenario pre condition mfa should be enabled for the pm steps login to pm click on admins tab click on add new admin button try to add admin without entering phone number field and verify ar able to create new admin account in the pm er new admin should not created in the pm without completing the phone number field note phone number field should be a mandatory field in the application if mfa is enabled for the pm
1
12,286
14,815,210,032
IssuesEvent
2021-01-14 06:49:02
Figma-Linux/figma-linux
https://api.github.com/repos/Figma-Linux/figma-linux
closed
'Recent files' tabs opening constantly, rendering the program useless.
Kind: Renderer Process Priority: High Status: To Do Type: Bug
* App version: 0.6.2 * Operating System (Platform and Version): Arch Linux on kernel 5.5.5-arch1-1 * Type of installed package (Snap, AppImage, deb, rpm, pacman): AUR/pacman **Bug description** Upon attempting to load `figma-linux` from `/opt/figma-linux/figma-linux` following a successful login, the program window appears but constantly tries to open new 'recent files' tabs, so I'm unable to use the program. Not too sure where to start with trying to rectify this; reinstalling does not resolve this problem. **Screenshots** ![image](https://user-images.githubusercontent.com/12870244/75104003-5e275180-5657-11ea-8232-bfcda30b3646.png)
1.0
'Recent files' tabs opening constantly, rendering the program useless. - * App version: 0.6.2 * Operating System (Platform and Version): Arch Linux on kernel 5.5.5-arch1-1 * Type of installed package (Snap, AppImage, deb, rpm, pacman): AUR/pacman **Bug description** Upon attempting to load `figma-linux` from `/opt/figma-linux/figma-linux` following a successful login, the program window appears but constantly tries to open new 'recent files' tabs, so I'm unable to use the program. Not too sure where to start with trying to rectify this; reinstalling does not resolve this problem. **Screenshots** ![image](https://user-images.githubusercontent.com/12870244/75104003-5e275180-5657-11ea-8232-bfcda30b3646.png)
process
recent files tabs opening constantly rendering the program useless app version operating system platform and version arch linux on kernel type of installed package snap appimage deb rpm pacman aur pacman bug description upon attempting to load figma linux from opt figma linux figma linux following a successful login the program window appears but constantly tries to open new recent files tabs so i m unable to use the program not too sure where to start with trying to rectify this reinstalling does not resolve this problem screenshots
1
9,229
12,260,114,997
IssuesEvent
2020-05-06 17:44:22
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
Timeout in unit test `tests.cmdline.commands.test_processes.py:test_pause_play_kill` when running locally
priority/quality-of-life topic/processes topic/testing topic/verdi type/bug
The test will timeout: ``` (aiida_dev) sph@bastion:~/code/aiida/env/dev/aiida-core$ pytest tests/cmdline/commands/test_process.py -sv -k test_pause_play_kill ============================================================================================== test session starts ============================================================================================== platform linux -- Python 3.7.6, pytest-5.4.1, py-1.5.4, pluggy-0.13.1 -- /home/sph/.virtualenvs/aiida_dev/bin/python3.7 cachedir: .pytest_cache rootdir: /home/sph/code/aiida/env/dev/aiida-core, inifile: pytest.ini plugins: regressions-1.0.6, timeout-1.3.4, asyncio-0.10.0, datadir-1.3.1, cov-2.8.1 collected 10 items / 9 deselected / 1 selected tests/cmdline/commands/test_process.py::TestVerdiProcessDaemon::test_pause_play_kill ['/home/sph/.virtualenvs/aiida_dev/bin/verdi', '-p', 'test_profile', 'devel', 'run_daemon'] FAILED =================================================================================================== FAILURES ==================================================================================================== __________________________________________________________________________________ TestVerdiProcessDaemon.test_pause_play_kill __________________________________________________________________________________ self = <tests.cmdline.commands.test_process.TestVerdiProcessDaemon testMethod=test_pause_play_kill> def test_pause_play_kill(self): """ Test the pause/play/kill commands """ # pylint: disable=no-member from aiida.orm import load_node calc = self.runner.submit(test_processes.WaitProcess) start_time = time.time() while calc.process_state is not plumpy.ProcessState.WAITING: if time.time() - start_time >= self.TEST_TIMEOUT: > self.fail('Timed out waiting for process to enter waiting state') E AssertionError: Timed out waiting for process to enter waiting state ``` Note that this happens both when specifying an explicit test profile as well as when letting the test manager generate one on the fly.
1.0
Timeout in unit test `tests.cmdline.commands.test_processes.py:test_pause_play_kill` when running locally - The test will timeout: ``` (aiida_dev) sph@bastion:~/code/aiida/env/dev/aiida-core$ pytest tests/cmdline/commands/test_process.py -sv -k test_pause_play_kill ============================================================================================== test session starts ============================================================================================== platform linux -- Python 3.7.6, pytest-5.4.1, py-1.5.4, pluggy-0.13.1 -- /home/sph/.virtualenvs/aiida_dev/bin/python3.7 cachedir: .pytest_cache rootdir: /home/sph/code/aiida/env/dev/aiida-core, inifile: pytest.ini plugins: regressions-1.0.6, timeout-1.3.4, asyncio-0.10.0, datadir-1.3.1, cov-2.8.1 collected 10 items / 9 deselected / 1 selected tests/cmdline/commands/test_process.py::TestVerdiProcessDaemon::test_pause_play_kill ['/home/sph/.virtualenvs/aiida_dev/bin/verdi', '-p', 'test_profile', 'devel', 'run_daemon'] FAILED =================================================================================================== FAILURES ==================================================================================================== __________________________________________________________________________________ TestVerdiProcessDaemon.test_pause_play_kill __________________________________________________________________________________ self = <tests.cmdline.commands.test_process.TestVerdiProcessDaemon testMethod=test_pause_play_kill> def test_pause_play_kill(self): """ Test the pause/play/kill commands """ # pylint: disable=no-member from aiida.orm import load_node calc = self.runner.submit(test_processes.WaitProcess) start_time = time.time() while calc.process_state is not plumpy.ProcessState.WAITING: if time.time() - start_time >= self.TEST_TIMEOUT: > self.fail('Timed out waiting for process to enter waiting state') E AssertionError: Timed out waiting for process to enter waiting state ``` Note that this happens both when specifying an explicit test profile as well as when letting the test manager generate one on the fly.
process
timeout in unit test tests cmdline commands test processes py test pause play kill when running locally the test will timeout aiida dev sph bastion code aiida env dev aiida core pytest tests cmdline commands test process py sv k test pause play kill test session starts platform linux python pytest py pluggy home sph virtualenvs aiida dev bin cachedir pytest cache rootdir home sph code aiida env dev aiida core inifile pytest ini plugins regressions timeout asyncio datadir cov collected items deselected selected tests cmdline commands test process py testverdiprocessdaemon test pause play kill failed failures testverdiprocessdaemon test pause play kill self def test pause play kill self test the pause play kill commands pylint disable no member from aiida orm import load node calc self runner submit test processes waitprocess start time time time while calc process state is not plumpy processstate waiting if time time start time self test timeout self fail timed out waiting for process to enter waiting state e assertionerror timed out waiting for process to enter waiting state note that this happens both when specifying an explicit test profile as well as when letting the test manager generate one on the fly
1
83,873
24,163,050,799
IssuesEvent
2022-09-22 13:09:44
eclipse-dataspaceconnector/DataSpaceConnector
https://api.github.com/repos/eclipse-dataspaceconnector/DataSpaceConnector
closed
Move the documentation generator `tooling/module-processor` into the plugin repo
documentation build
All code situation in `tooline/module-processor*` should be moved to the [plugin repo](https://github.com/eclipse-dataspaceconnector/GradlePlugins). This issue will contain of two PRs, one to move the code _into_ the plugin repo, and one to _delete the code_ from the EDC repo.
1.0
Move the documentation generator `tooling/module-processor` into the plugin repo - All code situation in `tooline/module-processor*` should be moved to the [plugin repo](https://github.com/eclipse-dataspaceconnector/GradlePlugins). This issue will contain of two PRs, one to move the code _into_ the plugin repo, and one to _delete the code_ from the EDC repo.
non_process
move the documentation generator tooling module processor into the plugin repo all code situation in tooline module processor should be moved to the this issue will contain of two prs one to move the code into the plugin repo and one to delete the code from the edc repo
0
16,682
9,487,326,838
IssuesEvent
2019-04-22 16:32:20
SixLabors/ImageSharp
https://api.github.com/repos/SixLabors/ImageSharp
closed
Optimize memory consumption of ResizeProcessor
area:memory area:performance area:transforms up-for-grabs
Currently we are using `O(NewWidth * SourceHeight)` memory for doing resize operations, however it's possible to do it consuming only `O(NewWidth * KernelDiameter)` memory, without significant negative effect on speed. We need to implement a "sliding window" algorithm for this. It will most likely fix issues like #597. **UPDATE** ~The issue is now part of the epic #733.~ **UPDATE 2** The ideas in #733 are abandoned now, we need to implement this with the current architecture.
True
Optimize memory consumption of ResizeProcessor - Currently we are using `O(NewWidth * SourceHeight)` memory for doing resize operations, however it's possible to do it consuming only `O(NewWidth * KernelDiameter)` memory, without significant negative effect on speed. We need to implement a "sliding window" algorithm for this. It will most likely fix issues like #597. **UPDATE** ~The issue is now part of the epic #733.~ **UPDATE 2** The ideas in #733 are abandoned now, we need to implement this with the current architecture.
non_process
optimize memory consumption of resizeprocessor currently we are using o newwidth sourceheight memory for doing resize operations however it s possible to do it consuming only o newwidth kerneldiameter memory without significant negative effect on speed we need to implement a sliding window algorithm for this it will most likely fix issues like update the issue is now part of the epic update the ideas in are abandoned now we need to implement this with the current architecture
0
2,615
5,394,349,710
IssuesEvent
2017-02-27 02:44:02
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
the password to vmware should not be traced to logs in post-processor.go
bug post-processor/vsphere
In file `post-processor/vsphere/post-processor.go, line 149 the call: ``` go log.Printf("Starting ovftool with parameters: %s", strings.Join(args, " ")) ``` traces password to logs. It should be handled in similar way as it is in `builder/vmware/iso/step_export.go`, line 23.
1.0
the password to vmware should not be traced to logs in post-processor.go - In file `post-processor/vsphere/post-processor.go, line 149 the call: ``` go log.Printf("Starting ovftool with parameters: %s", strings.Join(args, " ")) ``` traces password to logs. It should be handled in similar way as it is in `builder/vmware/iso/step_export.go`, line 23.
process
the password to vmware should not be traced to logs in post processor go in file post processor vsphere post processor go line the call go log printf starting ovftool with parameters s strings join args traces password to logs it should be handled in similar way as it is in builder vmware iso step export go line
1
14,544
17,661,164,691
IssuesEvent
2021-08-21 14:35:20
OpenCodeFoundation/eSchool
https://api.github.com/repos/OpenCodeFoundation/eSchool
closed
Build pipeline for certificate processing service
service:certificate-processing
Also, Update the readme file to show build status for the service
1.0
Build pipeline for certificate processing service - Also, Update the readme file to show build status for the service
process
build pipeline for certificate processing service also update the readme file to show build status for the service
1
20,062
4,488,456,613
IssuesEvent
2016-08-30 07:22:35
saltstack/salt
https://api.github.com/repos/saltstack/salt
closed
Salt wheel key documentation improvements
Core Documentation Feature Fixed Pending Verification TEAM Core
**Background:** I was interested in coding up accepting minion keys. I found https://docs.saltstack.com/en/latest/ref/wheel/all/salt.wheel.key.html, but I wanted to know if my `accept` went through or not (since there was a risk my key was rejected). I looked at https://github.com/saltstack/salt/blob/develop/salt/wheel/key.py#L49 and discovered that it actually returns a result of some kind. However, the PyDoc doesn't document that. **Proposal:** To document the return results on the functions in https://github.com/saltstack/salt/blob/develop/salt/wheel/key.py.
1.0
Salt wheel key documentation improvements - **Background:** I was interested in coding up accepting minion keys. I found https://docs.saltstack.com/en/latest/ref/wheel/all/salt.wheel.key.html, but I wanted to know if my `accept` went through or not (since there was a risk my key was rejected). I looked at https://github.com/saltstack/salt/blob/develop/salt/wheel/key.py#L49 and discovered that it actually returns a result of some kind. However, the PyDoc doesn't document that. **Proposal:** To document the return results on the functions in https://github.com/saltstack/salt/blob/develop/salt/wheel/key.py.
non_process
salt wheel key documentation improvements background i was interested in coding up accepting minion keys i found but i wanted to know if my accept went through or not since there was a risk my key was rejected i looked at and discovered that it actually returns a result of some kind however the pydoc doesn t document that proposal to document the return results on the functions in
0
2,131
4,971,890,734
IssuesEvent
2016-12-05 19:58:16
matz-e/lobster
https://api.github.com/repos/matz-e/lobster
closed
Move Dashboard Monitoring to a config setting
enhancement monitoring processing
Avoids having to query the SiteDB at every Lobster start, and allows us to give the option to override the user name… also brings us closer to #440?
1.0
Move Dashboard Monitoring to a config setting - Avoids having to query the SiteDB at every Lobster start, and allows us to give the option to override the user name… also brings us closer to #440?
process
move dashboard monitoring to a config setting avoids having to query the sitedb at every lobster start and allows us to give the option to override the user name… also brings us closer to
1
84,005
10,346,558,101
IssuesEvent
2019-09-04 15:29:42
python/mypy
https://api.github.com/repos/python/mypy
closed
Document how to suppress colorized output and summary line
documentation priority-0-high
We need to document `--no-coror-output` and `--no-error-summary`. This is a follow-up for https://github.com/python/mypy/pull/7425
1.0
Document how to suppress colorized output and summary line - We need to document `--no-coror-output` and `--no-error-summary`. This is a follow-up for https://github.com/python/mypy/pull/7425
non_process
document how to suppress colorized output and summary line we need to document no coror output and no error summary this is a follow up for
0
4,716
7,552,538,559
IssuesEvent
2018-04-19 00:54:11
UnbFeelings/unb-feelings-docs
https://api.github.com/repos/UnbFeelings/unb-feelings-docs
closed
[Não conformidade] Análises do que será medido não existe
Processo Qualidade invalid
@UnbFeelings/process Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o) fora auditada o [plano de medições](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o). ### Descrição Foi identificado que as análises do que será medido estão incompletas e/ou não existem. #### Recomendações Com base na Política de Não Conformidades utilizando a matriz GUT, obteve-se uma pontuação de 80 pontos, o que se encaixa em problema grave, assim o prazo para resolução da Não conformidade é de 1 dias #### Detalhes **Auditor**: Marcelo Augusto **Técnica de Audição**: Checklist **Tipo:** Processo **Prazo:** 18/04/2018
1.0
[Não conformidade] Análises do que será medido não existe - @UnbFeelings/process Perante critérios definidos para as [Auditorias](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o) fora auditada o [plano de medições](https://github.com/UnbFeelings/unb-feelings-GQA/wiki/Crit%C3%A9rios-de-Avalia%C3%A7%C3%A3o-e-T%C3%A9cnicas-de-Auditoria#plano-de-medi%C3%A7%C3%A3o). ### Descrição Foi identificado que as análises do que será medido estão incompletas e/ou não existem. #### Recomendações Com base na Política de Não Conformidades utilizando a matriz GUT, obteve-se uma pontuação de 80 pontos, o que se encaixa em problema grave, assim o prazo para resolução da Não conformidade é de 1 dias #### Detalhes **Auditor**: Marcelo Augusto **Técnica de Audição**: Checklist **Tipo:** Processo **Prazo:** 18/04/2018
process
análises do que será medido não existe unbfeelings process perante critérios definidos para as fora auditada o descrição foi identificado que as análises do que será medido estão incompletas e ou não existem recomendações com base na política de não conformidades utilizando a matriz gut obteve se uma pontuação de pontos o que se encaixa em problema grave assim o prazo para resolução da não conformidade é de dias detalhes auditor marcelo augusto técnica de audição checklist tipo processo prazo
1
20,058
26,544,851,729
IssuesEvent
2023-01-19 22:48:08
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Provide a "SED-like" mechanism to "find and replace" characters in a log record.
enhancement processor/transform
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] As an OTel native logging user, I need a way to "find and replace" characters that may exist in the original log record that I don't want/need in the logging backend. This can be useful to anonymize info, but also to make log records more usable in logging backends. My current example is removing ANSI colour codes from my log files... ``` [2022-01-04 01:51:50] [36mINFO[39m "GET /ads.txt" [36m301[39m 10ms ``` **Describe the solution you'd like** A clear and concise description of what you want to happen. I want to define a SED command processor to find and replace the ANSI colour codes, or any other characters in the log record, as I see fit. Something like: ``` SEDCMD-removecolorcodes = s/\x1B\[([0-9]{1,2}(;[0-9]{1,2})?)? ``` **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. I looked through existing OTel processors but I only see REGEX based ones, and none that make it easy to achieve "find and replace" in existing log pipelines **Additional context** Add any other context or screenshots about the feature request here. Sometimes kubernetes or application logs will use colour codes to make pretty printing when viewed in a terminal, or may include strings that need to be anonymized or replaced to make the log more usable by the user in the logging backend. ![image](https://user-images.githubusercontent.com/31549209/151220075-f3f427b9-a1a5-4c29-8289-17c8ead571f6.png)
1.0
Provide a "SED-like" mechanism to "find and replace" characters in a log record. - **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] As an OTel native logging user, I need a way to "find and replace" characters that may exist in the original log record that I don't want/need in the logging backend. This can be useful to anonymize info, but also to make log records more usable in logging backends. My current example is removing ANSI colour codes from my log files... ``` [2022-01-04 01:51:50] [36mINFO[39m "GET /ads.txt" [36m301[39m 10ms ``` **Describe the solution you'd like** A clear and concise description of what you want to happen. I want to define a SED command processor to find and replace the ANSI colour codes, or any other characters in the log record, as I see fit. Something like: ``` SEDCMD-removecolorcodes = s/\x1B\[([0-9]{1,2}(;[0-9]{1,2})?)? ``` **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. I looked through existing OTel processors but I only see REGEX based ones, and none that make it easy to achieve "find and replace" in existing log pipelines **Additional context** Add any other context or screenshots about the feature request here. Sometimes kubernetes or application logs will use colour codes to make pretty printing when viewed in a terminal, or may include strings that need to be anonymized or replaced to make the log more usable by the user in the logging backend. ![image](https://user-images.githubusercontent.com/31549209/151220075-f3f427b9-a1a5-4c29-8289-17c8ead571f6.png)
process
provide a sed like mechanism to find and replace characters in a log record is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when as an otel native logging user i need a way to find and replace characters that may exist in the original log record that i don t want need in the logging backend this can be useful to anonymize info but also to make log records more usable in logging backends my current example is removing ansi colour codes from my log files get ads txt describe the solution you d like a clear and concise description of what you want to happen i want to define a sed command processor to find and replace the ansi colour codes or any other characters in the log record as i see fit something like sedcmd removecolorcodes s describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered i looked through existing otel processors but i only see regex based ones and none that make it easy to achieve find and replace in existing log pipelines additional context add any other context or screenshots about the feature request here sometimes kubernetes or application logs will use colour codes to make pretty printing when viewed in a terminal or may include strings that need to be anonymized or replaced to make the log more usable by the user in the logging backend
1
21,778
30,290,507,772
IssuesEvent
2023-07-09 08:09:13
X-Sharp/XSharpPublic
https://api.github.com/repos/X-Sharp/XSharpPublic
closed
Preprocessor bug (XBase++ dialect)
bug Preprocessor
**Describe the bug** Preprocessor translates incorrectly UDC **Example 1. To Reproduce .prg** ``` // with iClass() {A, B, C} => {iClass():A, iClass():B, iClass():C} #xtranslate with <expr> {<v1>[, <vN>]} => {<expr>:<v1> [, <expr>:<vN>]} procedure Main() local arr := with Example() {A, B, C} AEval(arr, {|c| Qout(c) }) return class Example exported: inline access class method A(); return "A" inline access class method B(); return "B" inline access class method C(); return "C" endclass ``` **Expected behavior (xBase++ exe)** Output ``` A B C ``` **Actual behavior (X# exe)** ``` A B ``` **X# .ppo** ``` procedure Main() local arr := {Example():A , Example():B} AEval(arr, {|c| Qout(c) }) return class Example exported: inline access class method A(); return "A" inline access class method B(); return "B" inline access class method C(); return "C" endclass ``` If you change the array like this ``` local arr := with Example() {; A, B, C; } ``` it will be a compiler error ``` error XS9002: Parser: unexpected input 'with' ``` **Additional context** X# Compiler version 2.16.0.5 (public) -dialect:xBase++ -xpp1 -lb -memvar -vo1 -vo3 -vo5 -vo10 -vo15 -vo16
1.0
Preprocessor bug (XBase++ dialect) - **Describe the bug** Preprocessor translates incorrectly UDC **Example 1. To Reproduce .prg** ``` // with iClass() {A, B, C} => {iClass():A, iClass():B, iClass():C} #xtranslate with <expr> {<v1>[, <vN>]} => {<expr>:<v1> [, <expr>:<vN>]} procedure Main() local arr := with Example() {A, B, C} AEval(arr, {|c| Qout(c) }) return class Example exported: inline access class method A(); return "A" inline access class method B(); return "B" inline access class method C(); return "C" endclass ``` **Expected behavior (xBase++ exe)** Output ``` A B C ``` **Actual behavior (X# exe)** ``` A B ``` **X# .ppo** ``` procedure Main() local arr := {Example():A , Example():B} AEval(arr, {|c| Qout(c) }) return class Example exported: inline access class method A(); return "A" inline access class method B(); return "B" inline access class method C(); return "C" endclass ``` If you change the array like this ``` local arr := with Example() {; A, B, C; } ``` it will be a compiler error ``` error XS9002: Parser: unexpected input 'with' ``` **Additional context** X# Compiler version 2.16.0.5 (public) -dialect:xBase++ -xpp1 -lb -memvar -vo1 -vo3 -vo5 -vo10 -vo15 -vo16
process
preprocessor bug xbase dialect describe the bug preprocessor translates incorrectly udc example to reproduce prg with iclass a b c iclass a iclass b iclass c xtranslate with procedure main local arr with example a b c aeval arr c qout c return class example exported inline access class method a return a inline access class method b return b inline access class method c return c endclass expected behavior xbase exe output a b c actual behavior x exe a b x ppo procedure main local arr example a example b aeval arr c qout c return class example exported inline access class method a return a inline access class method b return b inline access class method c return c endclass if you change the array like this local arr with example a b c it will be a compiler error error parser unexpected input with additional context x compiler version public dialect xbase lb memvar
1
11,717
14,547,534,709
IssuesEvent
2020-12-15 23:10:34
pacificclimate/quail
https://api.github.com/repos/pacificclimate/quail
closed
TN10p
process
## Description This function takes a climdexInput object as input and computes the monthly or annual percent of values below the 10th percentile of baseline daily minimum temperature. ## Function to wrap [`climdex.tn10p`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L941)
1.0
TN10p - ## Description This function takes a climdexInput object as input and computes the monthly or annual percent of values below the 10th percentile of baseline daily minimum temperature. ## Function to wrap [`climdex.tn10p`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L941)
process
description this function takes a climdexinput object as input and computes the monthly or annual percent of values below the percentile of baseline daily minimum temperature function to wrap
1
6,016
8,822,391,503
IssuesEvent
2019-01-02 09:12:30
linnovate/root
https://api.github.com/repos/linnovate/root
closed
deleted entities multiple select bug
2.0.6 Fixed Process bug
when selecting the all option in multiple select mode, it selects only the one that was used to enter the mode in the first place and after uncheking one of the deleted items, the item gets deleted from the list until refreshing the page. in addition, you cant add watchers in select all mode. like #1396 ![image](https://user-images.githubusercontent.com/38312178/50156437-8428c780-02d7-11e9-8dd9-5fab29cc9c6b.png)
1.0
deleted entities multiple select bug - when selecting the all option in multiple select mode, it selects only the one that was used to enter the mode in the first place and after uncheking one of the deleted items, the item gets deleted from the list until refreshing the page. in addition, you cant add watchers in select all mode. like #1396 ![image](https://user-images.githubusercontent.com/38312178/50156437-8428c780-02d7-11e9-8dd9-5fab29cc9c6b.png)
process
deleted entities multiple select bug when selecting the all option in multiple select mode it selects only the one that was used to enter the mode in the first place and after uncheking one of the deleted items the item gets deleted from the list until refreshing the page in addition you cant add watchers in select all mode like
1
14,404
17,457,651,626
IssuesEvent
2021-08-06 05:24:32
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Android] Sign-out button is not working in particular scenario in Standalone app
Bug P1 Android Process: Tested QA
Steps:- 1. Login into application 2. Navigate to Hamburger menu 3. Click on Sign-out button 4. In the Pop-up click on cancel button 5. Again Click on Sign-out and verify A/R:- Sign-out button is not working once clicked on Cancel button E/R:- Sign-out button should work anytime and from any screen
1.0
[Android] Sign-out button is not working in particular scenario in Standalone app - Steps:- 1. Login into application 2. Navigate to Hamburger menu 3. Click on Sign-out button 4. In the Pop-up click on cancel button 5. Again Click on Sign-out and verify A/R:- Sign-out button is not working once clicked on Cancel button E/R:- Sign-out button should work anytime and from any screen
process
sign out button is not working in particular scenario in standalone app steps login into application navigate to hamburger menu click on sign out button in the pop up click on cancel button again click on sign out and verify a r sign out button is not working once clicked on cancel button e r sign out button should work anytime and from any screen
1
7,890
11,054,603,369
IssuesEvent
2019-12-10 13:48:16
cetic/tsorage
https://api.github.com/repos/cetic/tsorage
closed
Generalize the value type
ingestion processing
For the moment only numeric values are stored in the database, as floats. This should be extended in order to consider integer values, states (strings or encoding integers), 2D or 3D positions, or any arbitrary type. In particular, supporting arbitrary nested structure would be great. Types to consider for the moment: - [X] numeric (change floats to doubles) - [ ] text (for representing states) - [ ] 2 numerics (for instance, for representing a position ; demonstrate the ability to handle arbitrary structures)
1.0
Generalize the value type - For the moment only numeric values are stored in the database, as floats. This should be extended in order to consider integer values, states (strings or encoding integers), 2D or 3D positions, or any arbitrary type. In particular, supporting arbitrary nested structure would be great. Types to consider for the moment: - [X] numeric (change floats to doubles) - [ ] text (for representing states) - [ ] 2 numerics (for instance, for representing a position ; demonstrate the ability to handle arbitrary structures)
process
generalize the value type for the moment only numeric values are stored in the database as floats this should be extended in order to consider integer values states strings or encoding integers or positions or any arbitrary type in particular supporting arbitrary nested structure would be great types to consider for the moment numeric change floats to doubles text for representing states numerics for instance for representing a position demonstrate the ability to handle arbitrary structures
1
19,179
25,287,249,210
IssuesEvent
2022-11-16 20:24:39
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
opened
[Entry] Fachschaft Lehrertum
entry webform delete-after-processing
Unter verschiedenen Begriffen nicht gefunden: Fachschaft Lehrertum Fachschaft Lehramt Fachschaft Marsstraße Erwartetes Ergebnis: Marsstraße 20-22, Raum 130 (2907.01.130)
1.0
[Entry] Fachschaft Lehrertum - Unter verschiedenen Begriffen nicht gefunden: Fachschaft Lehrertum Fachschaft Lehramt Fachschaft Marsstraße Erwartetes Ergebnis: Marsstraße 20-22, Raum 130 (2907.01.130)
process
fachschaft lehrertum unter verschiedenen begriffen nicht gefunden fachschaft lehrertum fachschaft lehramt fachschaft marsstraße erwartetes ergebnis marsstraße raum
1
194,534
15,434,723,995
IssuesEvent
2021-03-07 05:05:25
devosoft/Empirical
https://api.github.com/repos/devosoft/Empirical
closed
"Test Documentation Build" job on GitHub Actions reports success even if building the docs failed
Goal: Bug Fix Prereq: Minimal (low-hanging-fruit?) Topic: Documentation
I think we can fix this by editing the docker container to return the exit code of the last thing it ran. Documentation on how to do that is [here](https://docs.github.com/en/actions/creating-actions/setting-exit-codes-for-actions#setting-a-failure-exit-code-in-a-docker-container-action).
1.0
"Test Documentation Build" job on GitHub Actions reports success even if building the docs failed - I think we can fix this by editing the docker container to return the exit code of the last thing it ran. Documentation on how to do that is [here](https://docs.github.com/en/actions/creating-actions/setting-exit-codes-for-actions#setting-a-failure-exit-code-in-a-docker-container-action).
non_process
test documentation build job on github actions reports success even if building the docs failed i think we can fix this by editing the docker container to return the exit code of the last thing it ran documentation on how to do that is
0
37,778
15,372,604,946
IssuesEvent
2021-03-02 11:28:11
KWARC/FoMID
https://api.github.com/repos/KWARC/FoMID
opened
Semantic Search
Services
Should largely work out of the box using MathWebSearch, but we might need to figure out how to generate an index from FLOMDoc-XHTML, that maybe also makes use of *semantic* information?
1.0
Semantic Search - Should largely work out of the box using MathWebSearch, but we might need to figure out how to generate an index from FLOMDoc-XHTML, that maybe also makes use of *semantic* information?
non_process
semantic search should largely work out of the box using mathwebsearch but we might need to figure out how to generate an index from flomdoc xhtml that maybe also makes use of semantic information
0
16,601
21,657,686,785
IssuesEvent
2022-05-06 15:38:24
HausDAO/daohaus-monorepo
https://api.github.com/repos/HausDAO/daohaus-monorepo
closed
Block Wednesdays from Meetings
process meetings
To help us focus and have time to get shit done we are trying to avoid Magesmith related meetings on Wednesdays. Hopefully this benefits the following opportunities for improvement: - Time balance and supporting need for more focused work - we all should make sure devs have everything they need - Get a clearer idea of the needs from design team for the upcoming few weeks - PMs now to run interference on any potential distractions for devs - turn off discord when in the flow This is from the areas we would like to improve from the [Retro: April 15, 2022](https://github.com/HausDAO/daohaus-monorepo/wiki/Retro:-April-15,-2022#what-would-we-like-to-change)
1.0
Block Wednesdays from Meetings - To help us focus and have time to get shit done we are trying to avoid Magesmith related meetings on Wednesdays. Hopefully this benefits the following opportunities for improvement: - Time balance and supporting need for more focused work - we all should make sure devs have everything they need - Get a clearer idea of the needs from design team for the upcoming few weeks - PMs now to run interference on any potential distractions for devs - turn off discord when in the flow This is from the areas we would like to improve from the [Retro: April 15, 2022](https://github.com/HausDAO/daohaus-monorepo/wiki/Retro:-April-15,-2022#what-would-we-like-to-change)
process
block wednesdays from meetings to help us focus and have time to get shit done we are trying to avoid magesmith related meetings on wednesdays hopefully this benefits the following opportunities for improvement time balance and supporting need for more focused work we all should make sure devs have everything they need get a clearer idea of the needs from design team for the upcoming few weeks pms now to run interference on any potential distractions for devs turn off discord when in the flow this is from the areas we would like to improve from the
1
57,872
6,559,463,020
IssuesEvent
2017-09-07 04:07:28
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
Type parameters checks are inconsistent and incomplete
A-resolve C-bug E-needstest
The way we currently check type parameters is inconsistent and error prone. The code below compiles and it shouldn't. The `resolve` phase currently ignores these checks and leaves them up to `typeck`, `collect` and `check`, which as demonstrated by bugs like this one and #20302 makes this check unsound. ``` rust fn is_copy<T: ::std::marker<i32>::Copy>() {} fn main() {} ```
1.0
Type parameters checks are inconsistent and incomplete - The way we currently check type parameters is inconsistent and error prone. The code below compiles and it shouldn't. The `resolve` phase currently ignores these checks and leaves them up to `typeck`, `collect` and `check`, which as demonstrated by bugs like this one and #20302 makes this check unsound. ``` rust fn is_copy<T: ::std::marker<i32>::Copy>() {} fn main() {} ```
non_process
type parameters checks are inconsistent and incomplete the way we currently check type parameters is inconsistent and error prone the code below compiles and it shouldn t the resolve phase currently ignores these checks and leaves them up to typeck collect and check which as demonstrated by bugs like this one and makes this check unsound rust fn is copy copy fn main
0
7,067
10,219,424,479
IssuesEvent
2019-08-15 18:32:55
toggl/mobileapp
https://api.github.com/repos/toggl/mobileapp
closed
Create a manual PR template (to copy and paste) for "new copy" translations
process
This PR should be very similar to the PR templated introduced by #4838, except that it will be used for new translations.
1.0
Create a manual PR template (to copy and paste) for "new copy" translations - This PR should be very similar to the PR templated introduced by #4838, except that it will be used for new translations.
process
create a manual pr template to copy and paste for new copy translations this pr should be very similar to the pr templated introduced by except that it will be used for new translations
1
154
2,581,441,311
IssuesEvent
2015-02-14 02:02:42
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
closed
Path Path interfaces and util/ implementations into structure/ package
enhancement process structure
With discussions with @spmallette, `Path` is a "data structure" and should be in structure/, not process/.
1.0
Path Path interfaces and util/ implementations into structure/ package - With discussions with @spmallette, `Path` is a "data structure" and should be in structure/, not process/.
process
path path interfaces and util implementations into structure package with discussions with spmallette path is a data structure and should be in structure not process
1
138,761
12,828,642,882
IssuesEvent
2020-07-06 20:58:14
NAVADMC/ADSM
https://api.github.com/repos/NAVADMC/ADSM
closed
Overlay - Interesting note
Documentation enhancement
If you are on a tab other than the scenario description and you hit the overlay, the app jumps back to the Scenario description tab. This was unexpected the first time I did it, but think it happens on purpose, as the overlay is specific to the one screen. Oops, I think the buggy part is that it doesn't let you navigate away if the overlay is on. I think I will add a note in known bugs, just in case someone asks.
1.0
Overlay - Interesting note - If you are on a tab other than the scenario description and you hit the overlay, the app jumps back to the Scenario description tab. This was unexpected the first time I did it, but think it happens on purpose, as the overlay is specific to the one screen. Oops, I think the buggy part is that it doesn't let you navigate away if the overlay is on. I think I will add a note in known bugs, just in case someone asks.
non_process
overlay interesting note if you are on a tab other than the scenario description and you hit the overlay the app jumps back to the scenario description tab this was unexpected the first time i did it but think it happens on purpose as the overlay is specific to the one screen oops i think the buggy part is that it doesn t let you navigate away if the overlay is on i think i will add a note in known bugs just in case someone asks
0
6,064
8,902,724,880
IssuesEvent
2019-01-17 08:32:11
Juris-M/citeproc-js
https://api.github.com/repos/Juris-M/citeproc-js
closed
DOI suffix gets wrapped in link
fix in process
In a CSL style that has a suffix for the `doi` field, the suffix gets wrapped in the link: The CSL looks like this: ```xml <text variable="DOI" prefix="doi:" suffix=", "/> ``` and leads to this HTML output: ```html doi:<a href="https://doi.org/10.1553/medievalworlds_no8_2018s126,">10.1553/medievalworlds_no8_2018s126,</a> ``` (The example is taken from the style chicago-author-date-de.) As one can see, the generated `href` attribute includes the trailing comma, which makes the DOI resolution fail and thus breaks the link. A test with citeproc-hs confirms that this is implementation specific to citeproc-js and can be seen in both Zotero and a server-side application.
1.0
DOI suffix gets wrapped in link - In a CSL style that has a suffix for the `doi` field, the suffix gets wrapped in the link: The CSL looks like this: ```xml <text variable="DOI" prefix="doi:" suffix=", "/> ``` and leads to this HTML output: ```html doi:<a href="https://doi.org/10.1553/medievalworlds_no8_2018s126,">10.1553/medievalworlds_no8_2018s126,</a> ``` (The example is taken from the style chicago-author-date-de.) As one can see, the generated `href` attribute includes the trailing comma, which makes the DOI resolution fail and thus breaks the link. A test with citeproc-hs confirms that this is implementation specific to citeproc-js and can be seen in both Zotero and a server-side application.
process
doi suffix gets wrapped in link in a csl style that has a suffix for the doi field the suffix gets wrapped in the link the csl looks like this xml and leads to this html output html doi a href the example is taken from the style chicago author date de as one can see the generated href attribute includes the trailing comma which makes the doi resolution fail and thus breaks the link a test with citeproc hs confirms that this is implementation specific to citeproc js and can be seen in both zotero and a server side application
1
804,863
29,504,815,762
IssuesEvent
2023-06-03 06:57:21
notofonts/latin-greek-cyrillic
https://api.github.com/repos/notofonts/latin-greek-cyrillic
closed
Combining grapheme joiner (CGJ) with double diacritics
in-evaluation Android Priority-Medium
u͡͏́i (0075 0361 034F 0301 0069) should have the combining acute 0301 on top of the double inverted breve 0361 when the CGF 034F is between the two, whereas ú͡i (0075 0361 0301 0069) should be rendered as the canonically equivalent ú͡i (0075 0301 0361 0069). See http://www.unicode.org/versions/Unicode8.0.0/ch07.pdf#page=43 Another example is t͡͏̇s (0074 0361 034F 0307 0073) which should be used to transliterate the Cyrillic ҵ in ALA-LC romanization in many library catalogues. Note: it is often encoded with the half marks from MARC, for example "Akhȧ︠t︡sarashẇa" in http://catalog.nypl.org/search/o38741987, or without the CGJ, for example "Akhaṫ͡sarashẇa" in https://catalog.lib.uchicago.edu/vufind/Record/2778324.
1.0
Combining grapheme joiner (CGJ) with double diacritics - u͡͏́i (0075 0361 034F 0301 0069) should have the combining acute 0301 on top of the double inverted breve 0361 when the CGF 034F is between the two, whereas ú͡i (0075 0361 0301 0069) should be rendered as the canonically equivalent ú͡i (0075 0301 0361 0069). See http://www.unicode.org/versions/Unicode8.0.0/ch07.pdf#page=43 Another example is t͡͏̇s (0074 0361 034F 0307 0073) which should be used to transliterate the Cyrillic ҵ in ALA-LC romanization in many library catalogues. Note: it is often encoded with the half marks from MARC, for example "Akhȧ︠t︡sarashẇa" in http://catalog.nypl.org/search/o38741987, or without the CGJ, for example "Akhaṫ͡sarashẇa" in https://catalog.lib.uchicago.edu/vufind/Record/2778324.
non_process
combining grapheme joiner cgj with double diacritics u͡͏́i should have the combining acute on top of the double inverted breve when the cgf is between the two whereas ú͡i should be rendered as the canonically equivalent ú͡i see another example is t͡͏̇s which should be used to transliterate the cyrillic ҵ in ala lc romanization in many library catalogues note it is often encoded with the half marks from marc for example akhȧ︠t︡sarashẇa in or without the cgj for example akhaṫ͡sarashẇa in
0
615
3,081,274,416
IssuesEvent
2015-08-22 15:09:53
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Post-processor failed: unexpected EOF
bug crash duplicate post-processor/vsphere
I am building with VMWare vmx post-processor and getting error when sending vmx file to vsphere. Any advice is appreciated. 2015/08/21 22:10:49 ui: ==> vmware-vmx: Running post-processor: vsphere ==> vmware-vmx: Running post-processor: vsphere 2015/08/21 22:10:49 ui: vmware-vmx (vsphere): Uploading output-vmware-vmx/packer-vmware-vmx-{{timestamp}}.vmx to vSphere 2015/08/21 22:10:49 packer-post-processor-vsphere: 2015/08/21 22:10:49 Starting ovftool with parameters: --noSSLVerify=false --acceptAllEulas --name=Mac OSX 10.10.5 --datastore=ESXI HDD 12 --diskMode=thin --network=VM_Network --vmFolder= output-vmware-vmx/packer-vmware-vmx-{{timestamp}}.vmx vi://root:Administr%40tor@10.25.197.12/DevOps/host/Mac Mini Lab vmware-vmx (vsphere): Uploading output-vmware-vmx/packer-vmware-vmx-{{timestamp}}.vmx to vSphere 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime: out of memory: cannot allocate 68719476736-byte block (68721573888 in use) 2015/08/21 22:45:43 packer-post-processor-vsphere: fatal error: out of memory 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime stack: 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.throw(0x67d64a) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/panic.go:491 +0xad 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.largeAlloc_m() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/malloc.c:390 +0xee 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.onM(0xc208012000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/asm_amd64.s:257 +0x68 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.mstart() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/proc.c:818 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 35 [running]: 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.switchtoM() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/asm_amd64.s:198 fp=0xc208093470 sp=0xc208093468 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.mallocgc(0xffffffe00, 0x328920, 0x1, 0x3c) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/malloc.go:199 +0x9f3 fp=0xc208093520 sp=0xc208093470 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.newarray(0x328920, 0xffffffe00, 0xc208021180) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/malloc.go:365 +0xc1 fp=0xc208093558 sp=0xc208093520 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.makeslice(0x318700, 0xffffffe00, 0xffffffe00, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/slice.go:32 +0x15c fp=0xc2080935a0 sp=0xc208093558 2015/08/21 22:45:43 packer-post-processor-vsphere: bytes.makeSlice(0xffffffe00, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bytes/buffer.go:191 +0x6a fp=0xc2080935f0 sp=0xc2080935a0 2015/08/21 22:45:43 packer-post-processor-vsphere: bytes.(*Buffer).ReadFrom(0xc20811c2a0, 0x776fc0, 0xc208104040, 0x7fffffc02, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bytes/buffer.go:163 +0xda fp=0xc208093688 sp=0xc2080935f0 2015/08/21 22:45:43 packer-post-processor-vsphere: io.Copy(0x7716f8, 0xc20811c2a0, 0x776fc0, 0xc208104040, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/io/io.go:358 +0x13d fp=0xc208093740 sp=0xc208093688 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.func·003(0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:221 +0x7d fp=0xc2080937a0 sp=0xc208093740 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.func·004(0xc2080fe1c0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:328 +0x2d fp=0xc2080937d8 sp=0xc2080937a0 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.goexit() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/asm_amd64.s:2232 +0x1 fp=0xc2080937e0 sp=0xc2080937d8 2015/08/21 22:45:43 packer-post-processor-vsphere: created by os/exec.(*Cmd).Start 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:329 +0xb6d 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 1 [select, 34 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Stream).Read(0xc2080ec000, 0xc208119000, 0x1000, 0x1000, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/stream.go:125 +0x3ab 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).fill(0xc20802e600) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:97 +0x1ce 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).ReadByte(0xc20802e600, 0x82, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:196 +0x86 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*ioDecReader).readn1(0xc20803be90, 0x16) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:90 +0x45 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*msgpackDecDriver).initReadNext(0xc20801f420) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/msgpack.go:540 +0x48 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).decode(0xc20802e660, 0x30cca0, 0xc20801f440) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:635 +0x59 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).Decode(0xc20802e660, 0x30cca0, 0xc20801f440, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:630 +0x7f 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*rpcCodec).read(0xc208039cc0, 0x30cca0, 0xc20801f440, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:88 +0xd9 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*goRpcCodec).ReadRequestHeader(0xc208039cc0, 0xc20801f440, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:127 +0x61 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Server).readRequestHeader(0xc208039bc0, 0x771bf8, 0xc208039cc0, 0x0, 0x0, 0xc20801f440, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:575 +0x97 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Server).readRequest(0xc208039bc0, 0x771bf8, 0xc208039cc0, 0xc20807c500, 0xc2080fe280, 0x328820, 0x0, 0x0, 0x0, 0x0, ...) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:542 +0xa2 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Server).ServeCodec(0xc208039bc0, 0x771bf8, 0xc208039cc0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:461 +0x6d 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/rpc.(*Server).Serve(0xc20801f380) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/server.go:141 +0x26f 2015/08/21 22:45:43 packer-post-processor-vsphere: main.main() 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/plugin/post-processor-vsphere/main.go:14 +0xc7 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 5 [syscall, 40 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: os/signal.loop() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/signal/signal_unix.go:21 +0x1f 2015/08/21 22:45:43 packer-post-processor-vsphere: created by os/signal.init·1 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/signal/signal_unix.go:27 +0x35 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 7 [chan receive, 40 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/plugin.func·017() 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/plugin/server.go:92 +0x69 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/mitchellh/packer/packer/plugin.Server 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/plugin/server.go:96 +0xcca 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 8 [IO wait]: 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*pollDesc).Wait(0xc2080801b0, 0x72, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/fd_poll_runtime.go:84 +0x47 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*pollDesc).WaitRead(0xc2080801b0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/fd_poll_runtime.go:89 +0x43 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*netFD).Read(0xc208080150, 0xc20807f000, 0x1000, 0x1000, 0x0, 0x76ff90, 0xc20811e060) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/fd_unix.go:242 +0x40f 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*conn).Read(0xc20802a080, 0xc20807f000, 0x1000, 0x1000, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/net.go:121 +0xdc 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).fill(0xc20802e2a0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:97 +0x1ce 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).Read(0xc20802e2a0, 0xc2080dc004, 0xc, 0xc, 0xc208011608, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:174 +0x26c 2015/08/21 22:45:43 packer-post-processor-vsphere: io.ReadAtLeast(0x7716a8, 0xc20802e2a0, 0xc2080dc004, 0xc, 0xc, 0xc, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/io/io.go:298 +0xf1 2015/08/21 22:45:43 packer-post-processor-vsphere: io.ReadFull(0x7716a8, 0xc20802e2a0, 0xc2080dc004, 0xc, 0xc, 0xc, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/io/io.go:316 +0x6d 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).recvLoop(0xc20806e000, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:381 +0x10d 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).recv(0xc20806e000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:369 +0x28 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/hashicorp/yamux.newSession 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:104 +0x499 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 9 [select]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).send(0xc20806e000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:331 +0x554 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/hashicorp/yamux.newSession 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:105 +0x4b3 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 10 [select]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).keepalive(0xc20806e000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:287 +0x106 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/hashicorp/yamux.newSession 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:107 +0x4de 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 11 [select, 40 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).AcceptStream(0xc20806e000, 0xc200000000, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:191 +0x1a8 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/rpc.(*muxBroker).Run(0xc20801f360) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/mux_broker.go:126 +0x37 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/mitchellh/packer/packer/rpc.NewServer 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/server.go:42 +0x7a 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 77 [syscall, 34 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: syscall.Syscall6(0x7, 0xb0d9, 0xc2080a13dc, 0x0, 0xc208122120, 0x0, 0x0, 0xc2080fc300, 0x3b01, 0xc2080fc320) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/syscall/asm_darwin_amd64.s:44 +0x5 2015/08/21 22:45:43 packer-post-processor-vsphere: syscall.wait4(0xb0d9, 0xc2080a13dc, 0x0, 0xc208122120, 0x90, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/syscall/zsyscall_darwin_amd64.go:32 +0x79 2015/08/21 22:45:43 packer-post-processor-vsphere: syscall.Wait4(0xb0d9, 0xc2080a1424, 0x0, 0xc208122120, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/syscall/syscall_bsd.go:162 +0x60 2015/08/21 22:45:43 packer-post-processor-vsphere: os.(*Process).wait(0xc2080fe4a0, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec_unix.go:22 +0x103 2015/08/21 22:45:43 packer-post-processor-vsphere: os.(*Process).Wait(0xc2080fe4a0, 0xc208066000, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/doc.go:45 +0x3a 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.(*Cmd).Wait(0xc208072140, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:364 +0x23c 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.(*Cmd).Run(0xc208072140, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:246 +0x71 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/post-processor/vsphere.(*PostProcessor).PostProcess(0xc208072280, 0x776f00, 0xc208116100, 0x776f48, 0xc208116120, 0x0, 0x0, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/post-processor/vsphere/post-processor.go:138 +0x1557 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/rpc.(*PostProcessorServer).PostProcess(0xc20801f3a0, 0x0, 0xc2080dc0b0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/post_processor.go:83 +0x30e 2015/08/21 22:45:43 packer-post-processor-vsphere: reflect.Value.call(0x3a1da0, 0x3dde90, 0x13, 0x459370, 0x4, 0xc2080a1f28, 0x3, 0x3, 0x0, 0x0, ...) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/reflect/value.go:419 +0x10e5 2015/08/21 22:45:43 packer-post-processor-vsphere: reflect.Value.Call(0x3a1da0, 0x3dde90, 0x13, 0xc2080a1f28, 0x3, 0x3, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/reflect/value.go:296 +0xbc 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*service).call(0xc208039c00, 0xc208039bc0, 0xc20800aa68, 0xc20807c500, 0xc2080fe280, 0x328820, 0xc2080dc088, 0xca, 0x307900, 0xc2080dc0b0, ...) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:382 +0x1f7 2015/08/21 22:45:43 packer-post-processor-vsphere: created by net/rpc.(*Server).ServeCodec 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:476 +0x44a 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 81 [select, 34 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Stream).Read(0xc2080ec140, 0xc208143000, 0x1000, 0x1000, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/stream.go:125 +0x3ab 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).fill(0xc2080ea540) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:97 +0x1ce 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).ReadByte(0xc2080ea540, 0x14c965, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:196 +0x86 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*ioDecReader).readn1(0xc20803be30, 0x308560) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:90 +0x45 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*msgpackDecDriver).initReadNext(0xc2081160c0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/msgpack.go:540 +0x48 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).decode(0xc2080ea5a0, 0x30cd00, 0xc20811a150) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:635 +0x59 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).Decode(0xc2080ea5a0, 0x30cd00, 0xc20811a150, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:630 +0x7f 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*rpcCodec).read(0xc2080f22c0, 0x30cd00, 0xc20811a150, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:88 +0xd9 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*goRpcCodec).ReadResponseHeader(0xc2080f22c0, 0xc20811a150, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:123 +0x61 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Client).input(0xc2080ea600) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/client.go:109 +0xb6 2015/08/21 22:45:43 packer-post-processor-vsphere: created by net/rpc.NewClientWithCodec 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/client.go:201 +0xd0 2015/08/21 22:45:45 Deleting original artifact for build 'vmware-vmx' 2015/08/21 22:45:45 /usr/local/bin/packer-post-processor-vsphere: plugin process exited 2015/08/21 22:45:45 ui error: Build 'vmware-vmx' errored: 1 error(s) occurred: * Post-processor failed: unexpected EOF 2015/08/21 22:45:45 Builds completed. Waiting on interrupt barrier... 2015/08/21 22:45:45 machine readable: error-count []string{"1"} 2015/08/21 22:45:45 ui error: ==> Some builds didn't complete successfully and had errors: 2015/08/21 22:45:45 machine readable: vmware-vmx,error []string{"1 error(s) occurred:\n\n* Post-processor failed: unexpected EOF"} 2015/08/21 22:45:45 ui error: --> vmware-vmx: 1 error(s) occurred: * Post-processor failed: unexpected EOF 2015/08/21 22:45:45 ui: ==> Builds finished but no artifacts were created. 2015/08/21 22:45:45 waiting for all plugin processes to complete... Build 'vmware-vmx' errored: 1 error(s) occurred: * Post-processor failed: unexpected EOF ==> Some builds didn't complete successfully and had errors: --> vmware-vmx: 1 error(s) occurred: * Post-processor failed: unexpected EOF ==> Builds finished but no artifacts were created. 2015/08/21 22:45:45 /usr/local/bin/packer-builder-vmware-vmx: plugin process exited benjaminbailey-17:packer devops$
1.0
Post-processor failed: unexpected EOF - I am building with VMWare vmx post-processor and getting error when sending vmx file to vsphere. Any advice is appreciated. 2015/08/21 22:10:49 ui: ==> vmware-vmx: Running post-processor: vsphere ==> vmware-vmx: Running post-processor: vsphere 2015/08/21 22:10:49 ui: vmware-vmx (vsphere): Uploading output-vmware-vmx/packer-vmware-vmx-{{timestamp}}.vmx to vSphere 2015/08/21 22:10:49 packer-post-processor-vsphere: 2015/08/21 22:10:49 Starting ovftool with parameters: --noSSLVerify=false --acceptAllEulas --name=Mac OSX 10.10.5 --datastore=ESXI HDD 12 --diskMode=thin --network=VM_Network --vmFolder= output-vmware-vmx/packer-vmware-vmx-{{timestamp}}.vmx vi://root:Administr%40tor@10.25.197.12/DevOps/host/Mac Mini Lab vmware-vmx (vsphere): Uploading output-vmware-vmx/packer-vmware-vmx-{{timestamp}}.vmx to vSphere 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime: out of memory: cannot allocate 68719476736-byte block (68721573888 in use) 2015/08/21 22:45:43 packer-post-processor-vsphere: fatal error: out of memory 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime stack: 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.throw(0x67d64a) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/panic.go:491 +0xad 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.largeAlloc_m() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/malloc.c:390 +0xee 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.onM(0xc208012000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/asm_amd64.s:257 +0x68 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.mstart() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/proc.c:818 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 35 [running]: 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.switchtoM() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/asm_amd64.s:198 fp=0xc208093470 sp=0xc208093468 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.mallocgc(0xffffffe00, 0x328920, 0x1, 0x3c) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/malloc.go:199 +0x9f3 fp=0xc208093520 sp=0xc208093470 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.newarray(0x328920, 0xffffffe00, 0xc208021180) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/malloc.go:365 +0xc1 fp=0xc208093558 sp=0xc208093520 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.makeslice(0x318700, 0xffffffe00, 0xffffffe00, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/slice.go:32 +0x15c fp=0xc2080935a0 sp=0xc208093558 2015/08/21 22:45:43 packer-post-processor-vsphere: bytes.makeSlice(0xffffffe00, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bytes/buffer.go:191 +0x6a fp=0xc2080935f0 sp=0xc2080935a0 2015/08/21 22:45:43 packer-post-processor-vsphere: bytes.(*Buffer).ReadFrom(0xc20811c2a0, 0x776fc0, 0xc208104040, 0x7fffffc02, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bytes/buffer.go:163 +0xda fp=0xc208093688 sp=0xc2080935f0 2015/08/21 22:45:43 packer-post-processor-vsphere: io.Copy(0x7716f8, 0xc20811c2a0, 0x776fc0, 0xc208104040, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/io/io.go:358 +0x13d fp=0xc208093740 sp=0xc208093688 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.func·003(0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:221 +0x7d fp=0xc2080937a0 sp=0xc208093740 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.func·004(0xc2080fe1c0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:328 +0x2d fp=0xc2080937d8 sp=0xc2080937a0 2015/08/21 22:45:43 packer-post-processor-vsphere: runtime.goexit() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/runtime/asm_amd64.s:2232 +0x1 fp=0xc2080937e0 sp=0xc2080937d8 2015/08/21 22:45:43 packer-post-processor-vsphere: created by os/exec.(*Cmd).Start 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:329 +0xb6d 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 1 [select, 34 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Stream).Read(0xc2080ec000, 0xc208119000, 0x1000, 0x1000, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/stream.go:125 +0x3ab 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).fill(0xc20802e600) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:97 +0x1ce 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).ReadByte(0xc20802e600, 0x82, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:196 +0x86 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*ioDecReader).readn1(0xc20803be90, 0x16) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:90 +0x45 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*msgpackDecDriver).initReadNext(0xc20801f420) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/msgpack.go:540 +0x48 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).decode(0xc20802e660, 0x30cca0, 0xc20801f440) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:635 +0x59 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).Decode(0xc20802e660, 0x30cca0, 0xc20801f440, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:630 +0x7f 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*rpcCodec).read(0xc208039cc0, 0x30cca0, 0xc20801f440, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:88 +0xd9 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*goRpcCodec).ReadRequestHeader(0xc208039cc0, 0xc20801f440, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:127 +0x61 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Server).readRequestHeader(0xc208039bc0, 0x771bf8, 0xc208039cc0, 0x0, 0x0, 0xc20801f440, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:575 +0x97 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Server).readRequest(0xc208039bc0, 0x771bf8, 0xc208039cc0, 0xc20807c500, 0xc2080fe280, 0x328820, 0x0, 0x0, 0x0, 0x0, ...) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:542 +0xa2 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Server).ServeCodec(0xc208039bc0, 0x771bf8, 0xc208039cc0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:461 +0x6d 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/rpc.(*Server).Serve(0xc20801f380) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/server.go:141 +0x26f 2015/08/21 22:45:43 packer-post-processor-vsphere: main.main() 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/plugin/post-processor-vsphere/main.go:14 +0xc7 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 5 [syscall, 40 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: os/signal.loop() 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/signal/signal_unix.go:21 +0x1f 2015/08/21 22:45:43 packer-post-processor-vsphere: created by os/signal.init·1 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/signal/signal_unix.go:27 +0x35 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 7 [chan receive, 40 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/plugin.func·017() 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/plugin/server.go:92 +0x69 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/mitchellh/packer/packer/plugin.Server 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/plugin/server.go:96 +0xcca 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 8 [IO wait]: 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*pollDesc).Wait(0xc2080801b0, 0x72, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/fd_poll_runtime.go:84 +0x47 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*pollDesc).WaitRead(0xc2080801b0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/fd_poll_runtime.go:89 +0x43 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*netFD).Read(0xc208080150, 0xc20807f000, 0x1000, 0x1000, 0x0, 0x76ff90, 0xc20811e060) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/fd_unix.go:242 +0x40f 2015/08/21 22:45:43 packer-post-processor-vsphere: net.(*conn).Read(0xc20802a080, 0xc20807f000, 0x1000, 0x1000, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/net.go:121 +0xdc 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).fill(0xc20802e2a0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:97 +0x1ce 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).Read(0xc20802e2a0, 0xc2080dc004, 0xc, 0xc, 0xc208011608, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:174 +0x26c 2015/08/21 22:45:43 packer-post-processor-vsphere: io.ReadAtLeast(0x7716a8, 0xc20802e2a0, 0xc2080dc004, 0xc, 0xc, 0xc, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/io/io.go:298 +0xf1 2015/08/21 22:45:43 packer-post-processor-vsphere: io.ReadFull(0x7716a8, 0xc20802e2a0, 0xc2080dc004, 0xc, 0xc, 0xc, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/io/io.go:316 +0x6d 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).recvLoop(0xc20806e000, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:381 +0x10d 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).recv(0xc20806e000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:369 +0x28 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/hashicorp/yamux.newSession 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:104 +0x499 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 9 [select]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).send(0xc20806e000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:331 +0x554 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/hashicorp/yamux.newSession 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:105 +0x4b3 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 10 [select]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).keepalive(0xc20806e000) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:287 +0x106 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/hashicorp/yamux.newSession 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:107 +0x4de 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 11 [select, 40 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Session).AcceptStream(0xc20806e000, 0xc200000000, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/session.go:191 +0x1a8 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/rpc.(*muxBroker).Run(0xc20801f360) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/mux_broker.go:126 +0x37 2015/08/21 22:45:43 packer-post-processor-vsphere: created by github.com/mitchellh/packer/packer/rpc.NewServer 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/server.go:42 +0x7a 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 77 [syscall, 34 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: syscall.Syscall6(0x7, 0xb0d9, 0xc2080a13dc, 0x0, 0xc208122120, 0x0, 0x0, 0xc2080fc300, 0x3b01, 0xc2080fc320) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/syscall/asm_darwin_amd64.s:44 +0x5 2015/08/21 22:45:43 packer-post-processor-vsphere: syscall.wait4(0xb0d9, 0xc2080a13dc, 0x0, 0xc208122120, 0x90, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/syscall/zsyscall_darwin_amd64.go:32 +0x79 2015/08/21 22:45:43 packer-post-processor-vsphere: syscall.Wait4(0xb0d9, 0xc2080a1424, 0x0, 0xc208122120, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/syscall/syscall_bsd.go:162 +0x60 2015/08/21 22:45:43 packer-post-processor-vsphere: os.(*Process).wait(0xc2080fe4a0, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec_unix.go:22 +0x103 2015/08/21 22:45:43 packer-post-processor-vsphere: os.(*Process).Wait(0xc2080fe4a0, 0xc208066000, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/doc.go:45 +0x3a 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.(*Cmd).Wait(0xc208072140, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:364 +0x23c 2015/08/21 22:45:43 packer-post-processor-vsphere: os/exec.(*Cmd).Run(0xc208072140, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/os/exec/exec.go:246 +0x71 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/post-processor/vsphere.(*PostProcessor).PostProcess(0xc208072280, 0x776f00, 0xc208116100, 0x776f48, 0xc208116120, 0x0, 0x0, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/post-processor/vsphere/post-processor.go:138 +0x1557 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/mitchellh/packer/packer/rpc.(*PostProcessorServer).PostProcess(0xc20801f3a0, 0x0, 0xc2080dc0b0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/mitchellh/packer/packer/rpc/post_processor.go:83 +0x30e 2015/08/21 22:45:43 packer-post-processor-vsphere: reflect.Value.call(0x3a1da0, 0x3dde90, 0x13, 0x459370, 0x4, 0xc2080a1f28, 0x3, 0x3, 0x0, 0x0, ...) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/reflect/value.go:419 +0x10e5 2015/08/21 22:45:43 packer-post-processor-vsphere: reflect.Value.Call(0x3a1da0, 0x3dde90, 0x13, 0xc2080a1f28, 0x3, 0x3, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/reflect/value.go:296 +0xbc 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*service).call(0xc208039c00, 0xc208039bc0, 0xc20800aa68, 0xc20807c500, 0xc2080fe280, 0x328820, 0xc2080dc088, 0xca, 0x307900, 0xc2080dc0b0, ...) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:382 +0x1f7 2015/08/21 22:45:43 packer-post-processor-vsphere: created by net/rpc.(*Server).ServeCodec 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/server.go:476 +0x44a 2015/08/21 22:45:43 packer-post-processor-vsphere: 2015/08/21 22:45:43 packer-post-processor-vsphere: goroutine 81 [select, 34 minutes]: 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/yamux.(*Stream).Read(0xc2080ec140, 0xc208143000, 0x1000, 0x1000, 0x0, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/yamux/stream.go:125 +0x3ab 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).fill(0xc2080ea540) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:97 +0x1ce 2015/08/21 22:45:43 packer-post-processor-vsphere: bufio.(*Reader).ReadByte(0xc2080ea540, 0x14c965, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/bufio/bufio.go:196 +0x86 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*ioDecReader).readn1(0xc20803be30, 0x308560) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:90 +0x45 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*msgpackDecDriver).initReadNext(0xc2081160c0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/msgpack.go:540 +0x48 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).decode(0xc2080ea5a0, 0x30cd00, 0xc20811a150) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:635 +0x59 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*Decoder).Decode(0xc2080ea5a0, 0x30cd00, 0xc20811a150, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/decode.go:630 +0x7f 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*rpcCodec).read(0xc2080f22c0, 0x30cd00, 0xc20811a150, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:88 +0xd9 2015/08/21 22:45:43 packer-post-processor-vsphere: github.com/hashicorp/go-msgpack/codec.(*goRpcCodec).ReadResponseHeader(0xc2080f22c0, 0xc20811a150, 0x0, 0x0) 2015/08/21 22:45:43 packer-post-processor-vsphere: /private/tmp/packer20150813-68003-e1u33m/src/github.com/hashicorp/go-msgpack/codec/rpc.go:123 +0x61 2015/08/21 22:45:43 packer-post-processor-vsphere: net/rpc.(*Client).input(0xc2080ea600) 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/client.go:109 +0xb6 2015/08/21 22:45:43 packer-post-processor-vsphere: created by net/rpc.NewClientWithCodec 2015/08/21 22:45:43 packer-post-processor-vsphere: /usr/local/Cellar/go/1.4.2/libexec/src/net/rpc/client.go:201 +0xd0 2015/08/21 22:45:45 Deleting original artifact for build 'vmware-vmx' 2015/08/21 22:45:45 /usr/local/bin/packer-post-processor-vsphere: plugin process exited 2015/08/21 22:45:45 ui error: Build 'vmware-vmx' errored: 1 error(s) occurred: * Post-processor failed: unexpected EOF 2015/08/21 22:45:45 Builds completed. Waiting on interrupt barrier... 2015/08/21 22:45:45 machine readable: error-count []string{"1"} 2015/08/21 22:45:45 ui error: ==> Some builds didn't complete successfully and had errors: 2015/08/21 22:45:45 machine readable: vmware-vmx,error []string{"1 error(s) occurred:\n\n* Post-processor failed: unexpected EOF"} 2015/08/21 22:45:45 ui error: --> vmware-vmx: 1 error(s) occurred: * Post-processor failed: unexpected EOF 2015/08/21 22:45:45 ui: ==> Builds finished but no artifacts were created. 2015/08/21 22:45:45 waiting for all plugin processes to complete... Build 'vmware-vmx' errored: 1 error(s) occurred: * Post-processor failed: unexpected EOF ==> Some builds didn't complete successfully and had errors: --> vmware-vmx: 1 error(s) occurred: * Post-processor failed: unexpected EOF ==> Builds finished but no artifacts were created. 2015/08/21 22:45:45 /usr/local/bin/packer-builder-vmware-vmx: plugin process exited benjaminbailey-17:packer devops$
process
post processor failed unexpected eof i am building with vmware vmx post processor and getting error when sending vmx file to vsphere any advice is appreciated ui vmware vmx running post processor vsphere vmware vmx running post processor vsphere ui vmware vmx vsphere uploading output vmware vmx packer vmware vmx timestamp vmx to vsphere packer post processor vsphere starting ovftool with parameters nosslverify false acceptalleulas name mac osx datastore esxi hdd diskmode thin network vm network vmfolder output vmware vmx packer vmware vmx timestamp vmx vi root administr devops host mac mini lab vmware vmx vsphere uploading output vmware vmx packer vmware vmx timestamp vmx to vsphere packer post processor vsphere runtime out of memory cannot allocate byte block in use packer post processor vsphere fatal error out of memory packer post processor vsphere packer post processor vsphere runtime stack packer post processor vsphere runtime throw packer post processor vsphere usr local cellar go libexec src runtime panic go packer post processor vsphere runtime largealloc m packer post processor vsphere usr local cellar go libexec src runtime malloc c packer post processor vsphere runtime onm packer post processor vsphere usr local cellar go libexec src runtime asm s packer post processor vsphere runtime mstart packer post processor vsphere usr local cellar go libexec src runtime proc c packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere runtime switchtom packer post processor vsphere usr local cellar go libexec src runtime asm s fp sp packer post processor vsphere runtime mallocgc packer post processor vsphere usr local cellar go libexec src runtime malloc go fp sp packer post processor vsphere runtime newarray packer post processor vsphere usr local cellar go libexec src runtime malloc go fp sp packer post processor vsphere runtime makeslice packer post processor vsphere usr local cellar go libexec src runtime slice go fp sp packer post processor vsphere bytes makeslice packer post processor vsphere usr local cellar go libexec src bytes buffer go fp sp packer post processor vsphere bytes buffer readfrom packer post processor vsphere usr local cellar go libexec src bytes buffer go fp sp packer post processor vsphere io copy packer post processor vsphere usr local cellar go libexec src io io go fp sp packer post processor vsphere os exec func· packer post processor vsphere usr local cellar go libexec src os exec exec go fp sp packer post processor vsphere os exec func· packer post processor vsphere usr local cellar go libexec src os exec exec go fp sp packer post processor vsphere runtime goexit packer post processor vsphere usr local cellar go libexec src runtime asm s fp sp packer post processor vsphere created by os exec cmd start packer post processor vsphere usr local cellar go libexec src os exec exec go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere github com hashicorp yamux stream read packer post processor vsphere private tmp src github com hashicorp yamux stream go packer post processor vsphere bufio reader fill packer post processor vsphere usr local cellar go libexec src bufio bufio go packer post processor vsphere bufio reader readbyte packer post processor vsphere usr local cellar go libexec src bufio bufio go packer post processor vsphere github com hashicorp go msgpack codec iodecreader packer post processor vsphere private tmp src github com hashicorp go msgpack codec decode go packer post processor vsphere github com hashicorp go msgpack codec msgpackdecdriver initreadnext packer post processor vsphere private tmp src github com hashicorp go msgpack codec msgpack go packer post processor vsphere github com hashicorp go msgpack codec decoder decode packer post processor vsphere private tmp src github com hashicorp go msgpack codec decode go packer post processor vsphere github com hashicorp go msgpack codec decoder decode packer post processor vsphere private tmp src github com hashicorp go msgpack codec decode go packer post processor vsphere github com hashicorp go msgpack codec rpccodec read packer post processor vsphere private tmp src github com hashicorp go msgpack codec rpc go packer post processor vsphere github com hashicorp go msgpack codec gorpccodec readrequestheader packer post processor vsphere private tmp src github com hashicorp go msgpack codec rpc go packer post processor vsphere net rpc server readrequestheader packer post processor vsphere usr local cellar go libexec src net rpc server go packer post processor vsphere net rpc server readrequest packer post processor vsphere usr local cellar go libexec src net rpc server go packer post processor vsphere net rpc server servecodec packer post processor vsphere usr local cellar go libexec src net rpc server go packer post processor vsphere github com mitchellh packer packer rpc server serve packer post processor vsphere private tmp src github com mitchellh packer packer rpc server go packer post processor vsphere main main packer post processor vsphere private tmp src github com mitchellh packer plugin post processor vsphere main go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere os signal loop packer post processor vsphere usr local cellar go libexec src os signal signal unix go packer post processor vsphere created by os signal init· packer post processor vsphere usr local cellar go libexec src os signal signal unix go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere github com mitchellh packer packer plugin func· packer post processor vsphere private tmp src github com mitchellh packer packer plugin server go packer post processor vsphere created by github com mitchellh packer packer plugin server packer post processor vsphere private tmp src github com mitchellh packer packer plugin server go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere net polldesc wait packer post processor vsphere usr local cellar go libexec src net fd poll runtime go packer post processor vsphere net polldesc waitread packer post processor vsphere usr local cellar go libexec src net fd poll runtime go packer post processor vsphere net netfd read packer post processor vsphere usr local cellar go libexec src net fd unix go packer post processor vsphere net conn read packer post processor vsphere usr local cellar go libexec src net net go packer post processor vsphere bufio reader fill packer post processor vsphere usr local cellar go libexec src bufio bufio go packer post processor vsphere bufio reader read packer post processor vsphere usr local cellar go libexec src bufio bufio go packer post processor vsphere io readatleast packer post processor vsphere usr local cellar go libexec src io io go packer post processor vsphere io readfull packer post processor vsphere usr local cellar go libexec src io io go packer post processor vsphere github com hashicorp yamux session recvloop packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere github com hashicorp yamux session recv packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere created by github com hashicorp yamux newsession packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere github com hashicorp yamux session send packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere created by github com hashicorp yamux newsession packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere github com hashicorp yamux session keepalive packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere created by github com hashicorp yamux newsession packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere github com hashicorp yamux session acceptstream packer post processor vsphere private tmp src github com hashicorp yamux session go packer post processor vsphere github com mitchellh packer packer rpc muxbroker run packer post processor vsphere private tmp src github com mitchellh packer packer rpc mux broker go packer post processor vsphere created by github com mitchellh packer packer rpc newserver packer post processor vsphere private tmp src github com mitchellh packer packer rpc server go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere syscall packer post processor vsphere usr local cellar go libexec src syscall asm darwin s packer post processor vsphere syscall packer post processor vsphere usr local cellar go libexec src syscall zsyscall darwin go packer post processor vsphere syscall packer post processor vsphere usr local cellar go libexec src syscall syscall bsd go packer post processor vsphere os process wait packer post processor vsphere usr local cellar go libexec src os exec unix go packer post processor vsphere os process wait packer post processor vsphere usr local cellar go libexec src os doc go packer post processor vsphere os exec cmd wait packer post processor vsphere usr local cellar go libexec src os exec exec go packer post processor vsphere os exec cmd run packer post processor vsphere usr local cellar go libexec src os exec exec go packer post processor vsphere github com mitchellh packer post processor vsphere postprocessor postprocess packer post processor vsphere private tmp src github com mitchellh packer post processor vsphere post processor go packer post processor vsphere github com mitchellh packer packer rpc postprocessorserver postprocess packer post processor vsphere private tmp src github com mitchellh packer packer rpc post processor go packer post processor vsphere reflect value call packer post processor vsphere usr local cellar go libexec src reflect value go packer post processor vsphere reflect value call packer post processor vsphere usr local cellar go libexec src reflect value go packer post processor vsphere net rpc service call packer post processor vsphere usr local cellar go libexec src net rpc server go packer post processor vsphere created by net rpc server servecodec packer post processor vsphere usr local cellar go libexec src net rpc server go packer post processor vsphere packer post processor vsphere goroutine packer post processor vsphere github com hashicorp yamux stream read packer post processor vsphere private tmp src github com hashicorp yamux stream go packer post processor vsphere bufio reader fill packer post processor vsphere usr local cellar go libexec src bufio bufio go packer post processor vsphere bufio reader readbyte packer post processor vsphere usr local cellar go libexec src bufio bufio go packer post processor vsphere github com hashicorp go msgpack codec iodecreader packer post processor vsphere private tmp src github com hashicorp go msgpack codec decode go packer post processor vsphere github com hashicorp go msgpack codec msgpackdecdriver initreadnext packer post processor vsphere private tmp src github com hashicorp go msgpack codec msgpack go packer post processor vsphere github com hashicorp go msgpack codec decoder decode packer post processor vsphere private tmp src github com hashicorp go msgpack codec decode go packer post processor vsphere github com hashicorp go msgpack codec decoder decode packer post processor vsphere private tmp src github com hashicorp go msgpack codec decode go packer post processor vsphere github com hashicorp go msgpack codec rpccodec read packer post processor vsphere private tmp src github com hashicorp go msgpack codec rpc go packer post processor vsphere github com hashicorp go msgpack codec gorpccodec readresponseheader packer post processor vsphere private tmp src github com hashicorp go msgpack codec rpc go packer post processor vsphere net rpc client input packer post processor vsphere usr local cellar go libexec src net rpc client go packer post processor vsphere created by net rpc newclientwithcodec packer post processor vsphere usr local cellar go libexec src net rpc client go deleting original artifact for build vmware vmx usr local bin packer post processor vsphere plugin process exited ui error build vmware vmx errored error s occurred post processor failed unexpected eof builds completed waiting on interrupt barrier machine readable error count string ui error some builds didn t complete successfully and had errors machine readable vmware vmx error string error s occurred n n post processor failed unexpected eof ui error vmware vmx error s occurred post processor failed unexpected eof ui builds finished but no artifacts were created waiting for all plugin processes to complete build vmware vmx errored error s occurred post processor failed unexpected eof some builds didn t complete successfully and had errors vmware vmx error s occurred post processor failed unexpected eof builds finished but no artifacts were created usr local bin packer builder vmware vmx plugin process exited benjaminbailey packer devops
1
13,060
15,394,566,878
IssuesEvent
2021-03-03 18:04:23
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Obsolete cell cycle arrest in response to nitrogen starvation & children
cell cycle and DNA processes obsoletion
Those terms have only been used by Pombase. Some annotations can be moved to 'GO:0051728 cell cycle switching, mitotic to meiotic cell cycle' * GO:0071849 G1 cell cycle arrest in response to nitrogen starvation * GO:0071851 mitotic G1 cell cycle arrest in response to nitrogen starvation * GO:1903694 positive regulation of mitotic G1 cell cycle arrest in response to nitrogen starvation * GO:1903693 regulation of mitotic G1 cell cycle arrest in response to nitrogen starvation
1.0
Obsolete cell cycle arrest in response to nitrogen starvation & children - Those terms have only been used by Pombase. Some annotations can be moved to 'GO:0051728 cell cycle switching, mitotic to meiotic cell cycle' * GO:0071849 G1 cell cycle arrest in response to nitrogen starvation * GO:0071851 mitotic G1 cell cycle arrest in response to nitrogen starvation * GO:1903694 positive regulation of mitotic G1 cell cycle arrest in response to nitrogen starvation * GO:1903693 regulation of mitotic G1 cell cycle arrest in response to nitrogen starvation
process
obsolete cell cycle arrest in response to nitrogen starvation children those terms have only been used by pombase some annotations can be moved to go cell cycle switching mitotic to meiotic cell cycle go cell cycle arrest in response to nitrogen starvation go mitotic cell cycle arrest in response to nitrogen starvation go positive regulation of mitotic cell cycle arrest in response to nitrogen starvation go regulation of mitotic cell cycle arrest in response to nitrogen starvation
1
4,793
7,688,045,335
IssuesEvent
2018-05-17 08:14:22
openvstorage/framework
https://api.github.com/repos/openvstorage/framework
closed
Auto-refresh dynamics
process_wontfix
### Feature description This feature is a responsive to big lists having longer-to-calculate dynamics. Retrieving a big list which serializes the dynamics of the object can take so long that the cached dynamics of that object have already expired. Fetching that same list again would thus result in re-calculating the dynamics. There should be separate process which keeps on refreshing the dynamics to avoid this issue. It might be mandatory to track which dynamics are more 'hot' than others to get some priority into this.
1.0
Auto-refresh dynamics - ### Feature description This feature is a responsive to big lists having longer-to-calculate dynamics. Retrieving a big list which serializes the dynamics of the object can take so long that the cached dynamics of that object have already expired. Fetching that same list again would thus result in re-calculating the dynamics. There should be separate process which keeps on refreshing the dynamics to avoid this issue. It might be mandatory to track which dynamics are more 'hot' than others to get some priority into this.
process
auto refresh dynamics feature description this feature is a responsive to big lists having longer to calculate dynamics retrieving a big list which serializes the dynamics of the object can take so long that the cached dynamics of that object have already expired fetching that same list again would thus result in re calculating the dynamics there should be separate process which keeps on refreshing the dynamics to avoid this issue it might be mandatory to track which dynamics are more hot than others to get some priority into this
1
6,291
9,300,925,177
IssuesEvent
2019-03-23 17:37:24
juanmait/seedjs
https://api.github.com/repos/juanmait/seedjs
opened
seed-process: add typings folder example with some envs
enhancement seed-process
- **seed-process**: add `typings/` folder example with some envs like in [seed-lib-node]. [seed-lib-node]: https://github.com/juanmait/seedjs/tree/master/seed-lib-node/typings
1.0
seed-process: add typings folder example with some envs - - **seed-process**: add `typings/` folder example with some envs like in [seed-lib-node]. [seed-lib-node]: https://github.com/juanmait/seedjs/tree/master/seed-lib-node/typings
process
seed process add typings folder example with some envs seed process add typings folder example with some envs like in
1
260,460
27,784,413,493
IssuesEvent
2023-03-17 01:06:05
AlexRogalskiy/javascript-tools
https://api.github.com/repos/AlexRogalskiy/javascript-tools
opened
CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz
Mend: dependency security vulnerability
## CVE-2022-3517 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/minimatch/package.json,/node_modules/npm/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - npm-7.0.10.tgz (Root Library) - npm-6.14.11.tgz - glob-7.1.6.tgz - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. <p>Publish Date: 2022-10-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3517>CVE-2022-3517</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-17</p> <p>Fix Resolution: minimatch - 3.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-3517 (High) detected in minimatch-3.0.4.tgz - ## CVE-2022-3517 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>minimatch-3.0.4.tgz</b></p></summary> <p>a glob matcher in javascript</p> <p>Library home page: <a href="https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz">https://registry.npmjs.org/minimatch/-/minimatch-3.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/minimatch/package.json,/node_modules/npm/node_modules/minimatch/package.json</p> <p> Dependency Hierarchy: - npm-7.0.10.tgz (Root Library) - npm-6.14.11.tgz - glob-7.1.6.tgz - :x: **minimatch-3.0.4.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in the minimatch package. This flaw allows a Regular Expression Denial of Service (ReDoS) when calling the braceExpand function with specific arguments, resulting in a Denial of Service. <p>Publish Date: 2022-10-17 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3517>CVE-2022-3517</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-17</p> <p>Fix Resolution: minimatch - 3.0.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in minimatch tgz cve high severity vulnerability vulnerable library minimatch tgz a glob matcher in javascript library home page a href path to dependency file package json path to vulnerable library node modules minimatch package json node modules npm node modules minimatch package json dependency hierarchy npm tgz root library npm tgz glob tgz x minimatch tgz vulnerable library found in base branch master vulnerability details a vulnerability was found in the minimatch package this flaw allows a regular expression denial of service redos when calling the braceexpand function with specific arguments resulting in a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution minimatch step up your open source security game with mend
0
333,273
29,519,810,419
IssuesEvent
2023-06-04 23:11:59
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix raw_ops.test_tensorflow_AddN
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix raw_ops.test_tensorflow_AddN - | | | |---|---| |tensorflow|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
non_process
fix raw ops test tensorflow addn tensorflow img src torch img src numpy img src jax img src paddle img src
0
73,430
15,254,519,931
IssuesEvent
2021-02-20 12:21:52
AlexRogalskiy/javascript-tools
https://api.github.com/repos/AlexRogalskiy/javascript-tools
opened
CVE-2020-28500 (Medium) detected in lodash-4.17.20.tgz
security vulnerability
## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p> <p>Path to dependency file: javascript-tools/package.json</p> <p>Path to vulnerable library: javascript-tools/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - github-7.2.0.tgz (Root Library) - :x: **lodash-4.17.20.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/javascript-tools/commit/6fc4e0f2da6b8c8aa0988f7137daa5d2b005c039">6fc4e0f2da6b8c8aa0988f7137daa5d2b005c039</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Steps to reproduce (provided by reporter Liyuan Chen): var lo = require('lodash'); function build_blank (n) { var ret = "1" for (var i = 0; i < n; i++) { ret += " " } return ret + "1"; } var s = build_blank(50000) var time0 = Date.now(); lo.trim(s) var time_cost0 = Date.now() - time0; console.log("time_cost0: " + time_cost0) var time1 = Date.now(); lo.toNumber(s) var time_cost1 = Date.now() - time1; console.log("time_cost1: " + time_cost1) var time2 = Date.now(); lo.trimEnd(s) var time_cost2 = Date.now() - time2; console.log("time_cost2: " + time_cost2) <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28500 (Medium) detected in lodash-4.17.20.tgz - ## CVE-2020-28500 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.20.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.20.tgz</a></p> <p>Path to dependency file: javascript-tools/package.json</p> <p>Path to vulnerable library: javascript-tools/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - github-7.2.0.tgz (Root Library) - :x: **lodash-4.17.20.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/javascript-tools/commit/6fc4e0f2da6b8c8aa0988f7137daa5d2b005c039">6fc4e0f2da6b8c8aa0988f7137daa5d2b005c039</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> All versions of package lodash; all versions of package org.fujion.webjars:lodash are vulnerable to Regular Expression Denial of Service (ReDoS) via the toNumber, trim and trimEnd functions. Steps to reproduce (provided by reporter Liyuan Chen): var lo = require('lodash'); function build_blank (n) { var ret = "1" for (var i = 0; i < n; i++) { ret += " " } return ret + "1"; } var s = build_blank(50000) var time0 = Date.now(); lo.trim(s) var time_cost0 = Date.now() - time0; console.log("time_cost0: " + time_cost0) var time1 = Date.now(); lo.toNumber(s) var time_cost1 = Date.now() - time1; console.log("time_cost1: " + time_cost1) var time2 = Date.now(); lo.trimEnd(s) var time_cost2 = Date.now() - time2; console.log("time_cost2: " + time_cost2) <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28500>CVE-2020-28500</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file javascript tools package json path to vulnerable library javascript tools node modules lodash package json dependency hierarchy github tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package lodash all versions of package org fujion webjars lodash are vulnerable to regular expression denial of service redos via the tonumber trim and trimend functions steps to reproduce provided by reporter liyuan chen var lo require lodash function build blank n var ret for var i i n i ret return ret var s build blank var date now lo trim s var time date now console log time time var date now lo tonumber s var time date now console log time time var date now lo trimend s var time date now console log time time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href step up your open source security game with whitesource
0
16,552
21,568,599,194
IssuesEvent
2022-05-02 04:17:57
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Fish Bowl 2
suggested title in process
The classic follow up to the groundbreaking original Fish Bowl motion picture. Title: Fish Bowl 2 Type (film/tv show): Film Film or show in which it appears: Ed, Edd n Eddy Is the parent film/show streaming anywhere? Yes About when in the parent film/show does it appear? Appears briefly in several episodes, the first being S1E1b - Nagged to Ed Actual footage of the film/show can be seen (yes/no)? Yes ![Fish-Bowl-2](https://user-images.githubusercontent.com/75587301/129507292-8cb7b4df-2b7c-4277-8d1a-2a22250c5eb5.gif)
1.0
Fish Bowl 2 - The classic follow up to the groundbreaking original Fish Bowl motion picture. Title: Fish Bowl 2 Type (film/tv show): Film Film or show in which it appears: Ed, Edd n Eddy Is the parent film/show streaming anywhere? Yes About when in the parent film/show does it appear? Appears briefly in several episodes, the first being S1E1b - Nagged to Ed Actual footage of the film/show can be seen (yes/no)? Yes ![Fish-Bowl-2](https://user-images.githubusercontent.com/75587301/129507292-8cb7b4df-2b7c-4277-8d1a-2a22250c5eb5.gif)
process
fish bowl the classic follow up to the groundbreaking original fish bowl motion picture title fish bowl type film tv show film film or show in which it appears ed edd n eddy is the parent film show streaming anywhere yes about when in the parent film show does it appear appears briefly in several episodes the first being nagged to ed actual footage of the film show can be seen yes no yes
1
16,621
21,678,126,864
IssuesEvent
2022-05-09 01:23:14
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add The Jerry Langford Show
suggested title in process
Please add as much of the following info as you can: Title: The Jerry Langford Show Type (film/tv show): TV show Film or show in which it appears: King Of Comedy Is the parent film/show streaming anywhere? About when in the parent film/show does it appear? Actual footage of the film/show can be seen (yes/no)? Yes
1.0
Add The Jerry Langford Show - Please add as much of the following info as you can: Title: The Jerry Langford Show Type (film/tv show): TV show Film or show in which it appears: King Of Comedy Is the parent film/show streaming anywhere? About when in the parent film/show does it appear? Actual footage of the film/show can be seen (yes/no)? Yes
process
add the jerry langford show please add as much of the following info as you can title the jerry langford show type film tv show tv show film or show in which it appears king of comedy is the parent film show streaming anywhere about when in the parent film show does it appear actual footage of the film show can be seen yes no yes
1
549,866
16,101,308,206
IssuesEvent
2021-04-27 09:38:16
googleapis/python-kms
https://api.github.com/repos/googleapis/python-kms
opened
Synthesis failed for python-kms
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate python-kms. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the python-kms API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` /bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:77:1 DEBUG: Rule 'com_google_protoc_java_resource_names_plugin' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "4b714b35ee04ba90f560ee60e64c7357428efcb6b0f3a298f343f8ec2c6d4a5d" DEBUG: Call stack for the definition of repository 'com_google_protoc_java_resource_names_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:234:1 DEBUG: Rule 'protoc_docs_plugin' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "33b387245455775e0de45869c7355cc5a9e98b396a6fc43b02812a63b75fee20" DEBUG: Call stack for the definition of repository 'protoc_docs_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:258:1 DEBUG: Rule 'rules_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "48f7e716f4098b85296ad93f5a133baf712968c13fbc2fdf3a6136158fe86eac" DEBUG: Call stack for the definition of repository 'rules_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:42:1 DEBUG: Rule 'gapic_generator_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "fe995def6873fcbdc2a8764ef4bce96eb971a9d1950fe9db9be442f3c64fb3b6" DEBUG: Call stack for the definition of repository 'gapic_generator_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:278:1 DEBUG: Rule 'com_googleapis_gapic_generator_go' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "c0d0efba86429cee5e52baf838165b0ed7cafae1748d025abec109d25e006628" DEBUG: Call stack for the definition of repository 'com_googleapis_gapic_generator_go' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:300:1 DEBUG: Rule 'gapic_generator_php' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "3dffc5c34a5f35666843df04b42d6ce1c545b992f9c093a777ec40833b548d86" DEBUG: Call stack for the definition of repository 'gapic_generator_php' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:364:1 DEBUG: Rule 'gapic_generator_csharp' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "4db430cfb9293e4521ec8e8138f8095faf035d8e752cf332d227710d749939eb" DEBUG: Call stack for the definition of repository 'gapic_generator_csharp' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:386:1 DEBUG: Rule 'gapic_generator_ruby' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "a14ec475388542f2ea70d16d75579065758acc4b99fdd6d59463d54e1a9e4499" DEBUG: Call stack for the definition of repository 'gapic_generator_ruby' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:400:1 DEBUG: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/rules_python/python/pip.bzl:61:5: DEPRECATED: the pip_repositories rule has been replaced with pip_install, please see rules_python 0.1 release notes DEBUG: Rule 'bazel_skylib' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "1dde365491125a3db70731e25658dfdd3bc5dbdfd11b840b3e987ecf043c7ca0" DEBUG: Call stack for the definition of repository 'bazel_skylib' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:35:1 Analyzing: target //google/cloud/kms/v1:kms-v1-py (1 packages loaded, 0 targets configured) ERROR: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/upb/bazel/upb_proto_library.bzl:257:29: aspect() got unexpected keyword argument 'incompatible_use_toolchain_transition' ERROR: Analysis of target '//google/cloud/kms/v1:kms-v1-py' failed; build aborted: error loading package '@com_github_grpc_grpc//': in /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/com_github_grpc_grpc/bazel/grpc_build_system.bzl: Extension file 'bazel/upb_proto_library.bzl' has errors INFO: Elapsed time: 0.202s INFO: 0 processes. FAILED: Build did NOT complete successfully (2 packages loaded, 3 targets configured) FAILED: Build did NOT complete successfully (2 packages loaded, 3 targets configured) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/python-kms/synth.py", line 37, in <module> include_protos=True, File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 52, in py_library return self._generate_code(service, version, "python", False, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 204, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/kms/v1:kms-v1-py']' returned non-zero exit status 1. 2021-04-27 02:38:14,209 autosynth [ERROR] > Synthesis failed 2021-04-27 02:38:14,210 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at b50d22c chore(revert): revert preventing normalization (#100) 2021-04-27 02:38:14,217 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2021-04-27 02:38:14,223 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/b27a8a99-fe7b-4b47-9eb5-bc1969bf8fbc/targets/github%2Fsynthtool;config=default/tests;query=python-kms;failed=false).
1.0
Synthesis failed for python-kms - Hello! Autosynth couldn't regenerate python-kms. :broken_heart: Please investigate and fix this issue within 5 business days. While it remains broken, this library cannot be updated with changes to the python-kms API, and the library grows stale. See https://github.com/googleapis/synthtool/blob/master/autosynth/TroubleShooting.md for trouble shooting tips. Here's the output from running `synth.py`: ``` /bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:77:1 DEBUG: Rule 'com_google_protoc_java_resource_names_plugin' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "4b714b35ee04ba90f560ee60e64c7357428efcb6b0f3a298f343f8ec2c6d4a5d" DEBUG: Call stack for the definition of repository 'com_google_protoc_java_resource_names_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:234:1 DEBUG: Rule 'protoc_docs_plugin' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "33b387245455775e0de45869c7355cc5a9e98b396a6fc43b02812a63b75fee20" DEBUG: Call stack for the definition of repository 'protoc_docs_plugin' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:258:1 DEBUG: Rule 'rules_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "48f7e716f4098b85296ad93f5a133baf712968c13fbc2fdf3a6136158fe86eac" DEBUG: Call stack for the definition of repository 'rules_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:42:1 DEBUG: Rule 'gapic_generator_python' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "fe995def6873fcbdc2a8764ef4bce96eb971a9d1950fe9db9be442f3c64fb3b6" DEBUG: Call stack for the definition of repository 'gapic_generator_python' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:278:1 DEBUG: Rule 'com_googleapis_gapic_generator_go' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "c0d0efba86429cee5e52baf838165b0ed7cafae1748d025abec109d25e006628" DEBUG: Call stack for the definition of repository 'com_googleapis_gapic_generator_go' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:300:1 DEBUG: Rule 'gapic_generator_php' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "3dffc5c34a5f35666843df04b42d6ce1c545b992f9c093a777ec40833b548d86" DEBUG: Call stack for the definition of repository 'gapic_generator_php' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:364:1 DEBUG: Rule 'gapic_generator_csharp' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "4db430cfb9293e4521ec8e8138f8095faf035d8e752cf332d227710d749939eb" DEBUG: Call stack for the definition of repository 'gapic_generator_csharp' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:386:1 DEBUG: Rule 'gapic_generator_ruby' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "a14ec475388542f2ea70d16d75579065758acc4b99fdd6d59463d54e1a9e4499" DEBUG: Call stack for the definition of repository 'gapic_generator_ruby' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:400:1 DEBUG: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/rules_python/python/pip.bzl:61:5: DEPRECATED: the pip_repositories rule has been replaced with pip_install, please see rules_python 0.1 release notes DEBUG: Rule 'bazel_skylib' indicated that a canonical reproducible form can be obtained by modifying arguments sha256 = "1dde365491125a3db70731e25658dfdd3bc5dbdfd11b840b3e987ecf043c7ca0" DEBUG: Call stack for the definition of repository 'bazel_skylib' which is a http_archive (rule definition at /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/bazel_tools/tools/build_defs/repo/http.bzl:296:16): - <builtin> - /home/kbuilder/.cache/synthtool/googleapis/WORKSPACE:35:1 Analyzing: target //google/cloud/kms/v1:kms-v1-py (1 packages loaded, 0 targets configured) ERROR: /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/upb/bazel/upb_proto_library.bzl:257:29: aspect() got unexpected keyword argument 'incompatible_use_toolchain_transition' ERROR: Analysis of target '//google/cloud/kms/v1:kms-v1-py' failed; build aborted: error loading package '@com_github_grpc_grpc//': in /home/kbuilder/.cache/bazel/_bazel_kbuilder/a732f932c2cbeb7e37e1543f189a2a73/external/com_github_grpc_grpc/bazel/grpc_build_system.bzl: Extension file 'bazel/upb_proto_library.bzl' has errors INFO: Elapsed time: 0.202s INFO: 0 processes. FAILED: Build did NOT complete successfully (2 packages loaded, 3 targets configured) FAILED: Build did NOT complete successfully (2 packages loaded, 3 targets configured) Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module> main() File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 219, in _call_with_frames_removed File "/home/kbuilder/.cache/synthtool/python-kms/synth.py", line 37, in <module> include_protos=True, File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 52, in py_library return self._generate_code(service, version, "python", False, **kwargs) File "/tmpfs/src/github/synthtool/synthtool/gcp/gapic_bazel.py", line 204, in _generate_code shell.run(bazel_run_args) File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 39, in run raise exc File "/tmpfs/src/github/synthtool/synthtool/shell.py", line 33, in run encoding="utf-8", File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 438, in run output=stdout, stderr=stderr) subprocess.CalledProcessError: Command '['bazel', '--max_idle_secs=240', 'build', '//google/cloud/kms/v1:kms-v1-py']' returned non-zero exit status 1. 2021-04-27 02:38:14,209 autosynth [ERROR] > Synthesis failed 2021-04-27 02:38:14,210 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at b50d22c chore(revert): revert preventing normalization (#100) 2021-04-27 02:38:14,217 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2021-04-27 02:38:14,223 autosynth [DEBUG] > Running: git clean -fdx Removing __pycache__/ Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 356, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 191, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 336, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 68, in synthesize_loop has_changes = toolbox.synthesize_version_in_new_branch(synthesizer, youngest) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 259, in synthesize_version_in_new_branch synthesizer.synthesize(synth_log_path, self.environ) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 120, in synthesize synth_proc.check_returncode() # Raise an exception. File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode self.stderr) subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'synth.metadata', 'synth.py', '--']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/b27a8a99-fe7b-4b47-9eb5-bc1969bf8fbc/targets/github%2Fsynthtool;config=default/tests;query=python-kms;failed=false).
non_process
synthesis failed for python kms hello autosynth couldn t regenerate python kms broken heart please investigate and fix this issue within business days while it remains broken this library cannot be updated with changes to the python kms api and the library grows stale see for trouble shooting tips here s the output from running synth py bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule com google protoc java resource names plugin indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository com google protoc java resource names plugin which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule protoc docs plugin indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository protoc docs plugin which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule rules python indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository rules python which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator python indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator python which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule com googleapis gapic generator go indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository com googleapis gapic generator go which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator php indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator php which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator csharp indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator csharp which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug rule gapic generator ruby indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository gapic generator ruby which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace debug home kbuilder cache bazel bazel kbuilder external rules python python pip bzl deprecated the pip repositories rule has been replaced with pip install please see rules python release notes debug rule bazel skylib indicated that a canonical reproducible form can be obtained by modifying arguments debug call stack for the definition of repository bazel skylib which is a http archive rule definition at home kbuilder cache bazel bazel kbuilder external bazel tools tools build defs repo http bzl home kbuilder cache synthtool googleapis workspace analyzing target google cloud kms kms py packages loaded targets configured error home kbuilder cache bazel bazel kbuilder external upb bazel upb proto library bzl aspect got unexpected keyword argument incompatible use toolchain transition error analysis of target google cloud kms kms py failed build aborted error loading package com github grpc grpc in home kbuilder cache bazel bazel kbuilder external com github grpc grpc bazel grpc build system bzl extension file bazel upb proto library bzl has errors info elapsed time info processes failed build did not complete successfully packages loaded targets configured failed build did not complete successfully packages loaded targets configured traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file home kbuilder cache synthtool python kms synth py line in include protos true file tmpfs src github synthtool synthtool gcp gapic bazel py line in py library return self generate code service version python false kwargs file tmpfs src github synthtool synthtool gcp gapic bazel py line in generate code shell run bazel run args file tmpfs src github synthtool synthtool shell py line in run raise exc file tmpfs src github synthtool synthtool shell py line in run encoding utf file home kbuilder pyenv versions lib subprocess py line in run output stdout stderr stderr subprocess calledprocesserror command returned non zero exit status autosynth synthesis failed autosynth running git reset hard head head is now at chore revert revert preventing normalization autosynth running git checkout autosynth switched to branch autosynth autosynth running git clean fdx removing pycache traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop has changes toolbox synthesize version in new branch synthesizer youngest file tmpfs src github synthtool autosynth synth toolbox py line in synthesize version in new branch synthesizer synthesize synth log path self environ file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
132,967
10,775,593,198
IssuesEvent
2019-11-03 15:20:26
LD4P/sinopia_editor
https://api.github.com/repos/LD4P/sinopia_editor
closed
automated test to assert license/copyright statement at top of all files we create
BFF future tests
Is there a way to do this as an automated check?
1.0
automated test to assert license/copyright statement at top of all files we create - Is there a way to do this as an automated check?
non_process
automated test to assert license copyright statement at top of all files we create is there a way to do this as an automated check
0
498,311
14,405,527,980
IssuesEvent
2020-12-03 18:50:37
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
opened
`cy.request()` timed out waiting `30000ms` for img in docs in 5.0 (Firefox only)
internal-priority
[https://circleci.com/gh/cypress-io/cypress/421794](https://circleci.com/gh/cypress-io/cypress/421794) This test finishes in 23 seconds normally in Chrome. When turning on firefoxGcInterval during open mode - tests take 53 secs in Firefox and pass. In 4.12.1, this test in Firefox ran 35 secs. CypressError: `cy.request()` timed out waiting `30000ms` for a response from your server. The request we sent was: Method: GET URL: http://localhost:2222/img/examples/frontend-masters-kent-c-dodds-javascript-testing.jpg No response was received within the timeout. https://on.**********/request request/<@http://localhost:2222/__cypress/runner/cypress_runner.js:158610:19 tryCatcher@http://localhost:2222/__cypress/runner/cypress_runner.js:9852:23 catchFilter/<@http://localhost:2222/__cypress/runner/cypress_runner.js:4974:41 tryCatcher@http://localhost:2222/__cypress/runner/cypress_runner.js:9852:23!Screen Shot 2020-08-14 at 11.03.49 AM.png|width=1429,height=638!
1.0
`cy.request()` timed out waiting `30000ms` for img in docs in 5.0 (Firefox only) - [https://circleci.com/gh/cypress-io/cypress/421794](https://circleci.com/gh/cypress-io/cypress/421794) This test finishes in 23 seconds normally in Chrome. When turning on firefoxGcInterval during open mode - tests take 53 secs in Firefox and pass. In 4.12.1, this test in Firefox ran 35 secs. CypressError: `cy.request()` timed out waiting `30000ms` for a response from your server. The request we sent was: Method: GET URL: http://localhost:2222/img/examples/frontend-masters-kent-c-dodds-javascript-testing.jpg No response was received within the timeout. https://on.**********/request request/<@http://localhost:2222/__cypress/runner/cypress_runner.js:158610:19 tryCatcher@http://localhost:2222/__cypress/runner/cypress_runner.js:9852:23 catchFilter/<@http://localhost:2222/__cypress/runner/cypress_runner.js:4974:41 tryCatcher@http://localhost:2222/__cypress/runner/cypress_runner.js:9852:23!Screen Shot 2020-08-14 at 11.03.49 AM.png|width=1429,height=638!
non_process
cy request timed out waiting for img in docs in firefox only this test finishes in seconds normally in chrome when turning on firefoxgcinterval during open mode tests take secs in firefox and pass in this test in firefox ran secs cypresserror cy request timed out waiting for a response from your server the request we sent was method get url no response was received within the timeout request trycatcher catchfilter trycatcher shot at am png width height
0
368,126
25,777,458,455
IssuesEvent
2022-12-09 13:12:50
bounswe/bounswe2022group2
https://api.github.com/repos/bounswe/bounswe2022group2
closed
Milestone 2: Fill User Interface Part for Mobile
priority-high type-documentation status-needreview milestone
### Issue Description As it is described in the customer milestone 2 description pdf file, we need to create a title named "User Interface / User Experience". The details of this part are given as follows in the description: - For each user interface design, provide links to the code in the project repository - Screenshots of all implemented Web and Mobile interfaces Therefore, as the mobile team, we need to fill this section with the screenshots, and code links of the screens/designs we implemented. Each team member will fill this section with his/her own implemented features. ### Step Details Steps that will be performed: - [x] Bahrican added the designs/screens he implemented @bahricanyesil - [x] Egemen added the designs/screens he implemented @egemenatikk - [x] Onur added the designs/screens he implemented @onurkomurcu - [x] Ezgi added the designs/screens she implemented @ezgy ### Final Actions PRs will be opened by each member for this issue. After all team members added his/her sections or the deadline has passed, the issue will be closed. ### Deadline of the Issue 09.12.2022 - 16:00 ### Reviewer Egemen Atik ### Deadline for the Review 09.12.2022 - 16:30
1.0
Milestone 2: Fill User Interface Part for Mobile - ### Issue Description As it is described in the customer milestone 2 description pdf file, we need to create a title named "User Interface / User Experience". The details of this part are given as follows in the description: - For each user interface design, provide links to the code in the project repository - Screenshots of all implemented Web and Mobile interfaces Therefore, as the mobile team, we need to fill this section with the screenshots, and code links of the screens/designs we implemented. Each team member will fill this section with his/her own implemented features. ### Step Details Steps that will be performed: - [x] Bahrican added the designs/screens he implemented @bahricanyesil - [x] Egemen added the designs/screens he implemented @egemenatikk - [x] Onur added the designs/screens he implemented @onurkomurcu - [x] Ezgi added the designs/screens she implemented @ezgy ### Final Actions PRs will be opened by each member for this issue. After all team members added his/her sections or the deadline has passed, the issue will be closed. ### Deadline of the Issue 09.12.2022 - 16:00 ### Reviewer Egemen Atik ### Deadline for the Review 09.12.2022 - 16:30
non_process
milestone fill user interface part for mobile issue description as it is described in the customer milestone description pdf file we need to create a title named user interface user experience the details of this part are given as follows in the description for each user interface design provide links to the code in the project repository screenshots of all implemented web and mobile interfaces therefore as the mobile team we need to fill this section with the screenshots and code links of the screens designs we implemented each team member will fill this section with his her own implemented features step details steps that will be performed bahrican added the designs screens he implemented bahricanyesil egemen added the designs screens he implemented egemenatikk onur added the designs screens he implemented onurkomurcu ezgi added the designs screens she implemented ezgy final actions prs will be opened by each member for this issue after all team members added his her sections or the deadline has passed the issue will be closed deadline of the issue reviewer egemen atik deadline for the review
0
11,654
14,516,116,877
IssuesEvent
2020-12-13 14:51:50
luc-github/Repetier-Firmware-4-Davinci
https://api.github.com/repos/luc-github/Repetier-Firmware-4-Davinci
closed
2.0a Bed temperature wrong
0.92.10 Waiting to be processed enhancement
Hello Since I use on my DaVinci 2.0A FW0.92.10 I have a temperature difference of about 25 ° C at my print-table. The display will show too much on the display and also in the Repetier-Host 25 ° C. I have checked with an infrared thermometer and with several temperature sensors. How can I correct this? Everything else works very well! Thank you! Greeting Hannes P.S. I use Google translator because my english is bad.
1.0
2.0a Bed temperature wrong - Hello Since I use on my DaVinci 2.0A FW0.92.10 I have a temperature difference of about 25 ° C at my print-table. The display will show too much on the display and also in the Repetier-Host 25 ° C. I have checked with an infrared thermometer and with several temperature sensors. How can I correct this? Everything else works very well! Thank you! Greeting Hannes P.S. I use Google translator because my english is bad.
process
bed temperature wrong hello since i use on my davinci i have a temperature difference of about ° c at my print table the display will show too much on the display and also in the repetier host ° c i have checked with an infrared thermometer and with several temperature sensors how can i correct this everything else works very well thank you greeting hannes p s i use google translator because my english is bad
1
5,032
2,770,930,993
IssuesEvent
2015-05-01 18:02:16
Polymer/designer
https://api.github.com/repos/Polymer/designer
closed
Need to be able to zoom out or pan around the canvas area
designer1 enhancement wontfix
It would be great if I could zoom out from the canvas a bit to fit more elements on screen, or at least have the ability to pan around. I have an app which is ~960px across and that doesn't fit well in the current confines
1.0
Need to be able to zoom out or pan around the canvas area - It would be great if I could zoom out from the canvas a bit to fit more elements on screen, or at least have the ability to pan around. I have an app which is ~960px across and that doesn't fit well in the current confines
non_process
need to be able to zoom out or pan around the canvas area it would be great if i could zoom out from the canvas a bit to fit more elements on screen or at least have the ability to pan around i have an app which is across and that doesn t fit well in the current confines
0
646,564
21,052,882,556
IssuesEvent
2022-03-31 22:20:14
SSWConsulting/SSW.SophieBot
https://api.github.com/repos/SSWConsulting/SSW.SophieBot
closed
✨ Query - Which client is Calum working for?
feature Priority: medium
<!-- These comments automatically delete --> <!-- **Tip:** Delete parts that are not relevant --> <!-- Next to Cc:, @ mention users who should be in the loop --> Cc: @JeanThirion @AttackOnMorty <!-- add intended user next to **Hi** --> ### Pain <!-- Explain the pain you are experiencing --> From @JeanThirion's feedback, > "Who is XXX booked for" We need to support querying clients ### Tasks <!--Add GitHub tasks--> - [x] Query - - "Which client is Calum working for?" - "Who is Calum booked for?" - "Who is Calum working for?" Thanks!
1.0
✨ Query - Which client is Calum working for? - <!-- These comments automatically delete --> <!-- **Tip:** Delete parts that are not relevant --> <!-- Next to Cc:, @ mention users who should be in the loop --> Cc: @JeanThirion @AttackOnMorty <!-- add intended user next to **Hi** --> ### Pain <!-- Explain the pain you are experiencing --> From @JeanThirion's feedback, > "Who is XXX booked for" We need to support querying clients ### Tasks <!--Add GitHub tasks--> - [x] Query - - "Which client is Calum working for?" - "Who is Calum booked for?" - "Who is Calum working for?" Thanks!
non_process
✨ query which client is calum working for cc jeanthirion attackonmorty pain from jeanthirion s feedback who is xxx booked for we need to support querying clients tasks query which client is calum working for who is calum booked for who is calum working for thanks
0
15,421
19,606,035,480
IssuesEvent
2022-01-06 09:35:48
plazi/community
https://api.github.com/repos/plazi/community
opened
to be processed: https://link.springer.com/content/pdf/10.1007/s12225-021-09988-7.pdf
process request
this is one more from the CAS new species press release https://link.springer.com/content/pdf/10.1007/s12225-021-09988-7.pdf thanks for processing holotype, gbif
1.0
to be processed: https://link.springer.com/content/pdf/10.1007/s12225-021-09988-7.pdf - this is one more from the CAS new species press release https://link.springer.com/content/pdf/10.1007/s12225-021-09988-7.pdf thanks for processing holotype, gbif
process
to be processed this is one more from the cas new species press release thanks for processing holotype gbif
1
12,708
15,081,217,008
IssuesEvent
2021-02-05 12:53:03
emacs-ess/ESS
https://api.github.com/repos/emacs-ess/ESS
closed
C-level Debugging of R now quickly gets to "no process"
bug process
``` C-u M-x R <Enter> -d gdb <Enter> ``` open `gdb` in the `*R*` buffer (and you must typically switch there yourself). There, you type `r` at the gdb prompt `(gdb)` so indeed, R starts running seemingly regularly. Now, e.g., you press `C-c C-c` with which you go into the debugger and then e.g., ``` C-c C-c Program received signal SIGINT, Interrupt. 0x00007ffff7160f7a in select () from /lib64/libc.so.6 Missing separate debuginfos, use: dnf debuginfo-install bzip2-libs-1.0.8-2.fc32.x86_64 libgcc-10.2.1-9.fc32.x86_64 libgfortran-10.2.1-9.fc32.x86_64 libgomp-10.2.1-9.fc32.x86_64 libicu-65.1-2.fc32.x86_64 libstdc++-10.2.1-9.fc32.x86_64 ncurses-libs-6.1-15.20191109.fc32.x86_64 pcre2-10.36-1.fc32.x86_64 readline-8.0-4.fc32.x86_64 sssd-client-2.4.0-1.fc32.x86_64 xz-libs-5.2.5-1.fc32.x86_64 (gdb) br do_names Breakpoint 1 at 0x45cfb0: file ../../../R/src/main/attrib.c, line 1016. (gdb) c Continuing. > ``` with 'c' ('continue') I go back to the R prompt after having set a (C) breakpoint at the `do_names()` C function which is called from R's names() function. This is all as it has worked for many years with ESS and R. But since relatively recently (I discovered ~ a week ago; but I had n't tried to run R in the C level debugger for quite a few weeks), very quickly the R session seems to "lose the process": The emacs status line then suddenly changes from > (iESS [R] run Eldoc) to > (iESS [R] no process Eldoc) and I also get a message in the minibuffer > Current buffer has no process and from then on, I cannot run any command anymore in the `iESS` buffer. It may be that this only happens when I press `[Tab]` for R completion of function names, etc.
1.0
C-level Debugging of R now quickly gets to "no process" - ``` C-u M-x R <Enter> -d gdb <Enter> ``` open `gdb` in the `*R*` buffer (and you must typically switch there yourself). There, you type `r` at the gdb prompt `(gdb)` so indeed, R starts running seemingly regularly. Now, e.g., you press `C-c C-c` with which you go into the debugger and then e.g., ``` C-c C-c Program received signal SIGINT, Interrupt. 0x00007ffff7160f7a in select () from /lib64/libc.so.6 Missing separate debuginfos, use: dnf debuginfo-install bzip2-libs-1.0.8-2.fc32.x86_64 libgcc-10.2.1-9.fc32.x86_64 libgfortran-10.2.1-9.fc32.x86_64 libgomp-10.2.1-9.fc32.x86_64 libicu-65.1-2.fc32.x86_64 libstdc++-10.2.1-9.fc32.x86_64 ncurses-libs-6.1-15.20191109.fc32.x86_64 pcre2-10.36-1.fc32.x86_64 readline-8.0-4.fc32.x86_64 sssd-client-2.4.0-1.fc32.x86_64 xz-libs-5.2.5-1.fc32.x86_64 (gdb) br do_names Breakpoint 1 at 0x45cfb0: file ../../../R/src/main/attrib.c, line 1016. (gdb) c Continuing. > ``` with 'c' ('continue') I go back to the R prompt after having set a (C) breakpoint at the `do_names()` C function which is called from R's names() function. This is all as it has worked for many years with ESS and R. But since relatively recently (I discovered ~ a week ago; but I had n't tried to run R in the C level debugger for quite a few weeks), very quickly the R session seems to "lose the process": The emacs status line then suddenly changes from > (iESS [R] run Eldoc) to > (iESS [R] no process Eldoc) and I also get a message in the minibuffer > Current buffer has no process and from then on, I cannot run any command anymore in the `iESS` buffer. It may be that this only happens when I press `[Tab]` for R completion of function names, etc.
process
c level debugging of r now quickly gets to no process c u m x r d gdb open gdb in the r buffer and you must typically switch there yourself there you type r at the gdb prompt gdb so indeed r starts running seemingly regularly now e g you press c c c c with which you go into the debugger and then e g c c c c program received signal sigint interrupt in select from libc so missing separate debuginfos use dnf debuginfo install libs libgcc libgfortran libgomp libicu libstdc ncurses libs readline sssd client xz libs gdb br do names breakpoint at file r src main attrib c line gdb c continuing with c continue i go back to the r prompt after having set a c breakpoint at the do names c function which is called from r s names function this is all as it has worked for many years with ess and r but since relatively recently i discovered a week ago but i had n t tried to run r in the c level debugger for quite a few weeks very quickly the r session seems to lose the process the emacs status line then suddenly changes from iess run eldoc to iess no process eldoc and i also get a message in the minibuffer current buffer has no process and from then on i cannot run any command anymore in the iess buffer it may be that this only happens when i press for r completion of function names etc
1
13,048
15,387,907,208
IssuesEvent
2021-03-03 10:06:42
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Error: Error in migration engine. Reason: [C:\Users\runneradmin\.cargo\git\checkouts\quaint-9f01e008b9a89c14\7f02e62\src\connector\postgres\error.rs:67:35] column on null constraint violation error
bug/1-repro-available kind/bug process/candidate status/needs-action team/migrations tech/engines topic: migrate
<!-- If required, please update the title to be clear and descriptive --> Command: `prisma migrate up --experimental` Version: `2.3.0` Binary Version: `e11114fa1ea826f9e7b4fa1ced34e78892fe8e0e` Report: https://prisma-errors.netlify.app/report/9757 OS: `x64 win32 10.0.18363` JS Stacktrace: ``` Error: Error in migration engine. Reason: [C:\Users\runneradmin\.cargo\git\checkouts\quaint-9f01e008b9a89c14\7f02e62\src\connector\postgres\error.rs:67:35] column on null constraint violation error Please create an issue in the migrate repo with your `schema.prisma` and the prisma command you tried to use 🙏: https://github.com/prisma/migrate/issues/new at ChildProcess.<anonymous> (C:\Users\Lenovo\train\untitled-allbet\node_modules\@prisma\cli\build\index.js:2:2134122) at ChildProcess.emit (events.js:315:20) at Process.ChildProcess._handle.onexit (internal/child_process.js:275:12) ``` Rust Stacktrace: ``` [C:\Users\runneradmin\.cargo\git\checkouts\quaint-9f01e008b9a89c14\7f02e62\src\connector\postgres\error.rs:67:35] column on null constraint violation error ```
1.0
Error: Error in migration engine. Reason: [C:\Users\runneradmin\.cargo\git\checkouts\quaint-9f01e008b9a89c14\7f02e62\src\connector\postgres\error.rs:67:35] column on null constraint violation error - <!-- If required, please update the title to be clear and descriptive --> Command: `prisma migrate up --experimental` Version: `2.3.0` Binary Version: `e11114fa1ea826f9e7b4fa1ced34e78892fe8e0e` Report: https://prisma-errors.netlify.app/report/9757 OS: `x64 win32 10.0.18363` JS Stacktrace: ``` Error: Error in migration engine. Reason: [C:\Users\runneradmin\.cargo\git\checkouts\quaint-9f01e008b9a89c14\7f02e62\src\connector\postgres\error.rs:67:35] column on null constraint violation error Please create an issue in the migrate repo with your `schema.prisma` and the prisma command you tried to use 🙏: https://github.com/prisma/migrate/issues/new at ChildProcess.<anonymous> (C:\Users\Lenovo\train\untitled-allbet\node_modules\@prisma\cli\build\index.js:2:2134122) at ChildProcess.emit (events.js:315:20) at Process.ChildProcess._handle.onexit (internal/child_process.js:275:12) ``` Rust Stacktrace: ``` [C:\Users\runneradmin\.cargo\git\checkouts\quaint-9f01e008b9a89c14\7f02e62\src\connector\postgres\error.rs:67:35] column on null constraint violation error ```
process
error error in migration engine reason column on null constraint violation error command prisma migrate up experimental version binary version report os js stacktrace error error in migration engine reason column on null constraint violation error please create an issue in the migrate repo with your schema prisma and the prisma command you tried to use 🙏 at childprocess c users lenovo train untitled allbet node modules prisma cli build index js at childprocess emit events js at process childprocess handle onexit internal child process js rust stacktrace column on null constraint violation error
1
119,355
15,504,477,903
IssuesEvent
2021-03-11 14:20:44
Feminine-Divine/periodo
https://api.github.com/repos/Feminine-Divine/periodo
closed
Design: Create a Logo
Level 3 coding design good first issue gssoc21
## Description Create the Logo for the Project that needs to be put in the corner of the navbar
1.0
Design: Create a Logo - ## Description Create the Logo for the Project that needs to be put in the corner of the navbar
non_process
design create a logo description create the logo for the project that needs to be put in the corner of the navbar
0
6,836
3,475,547,145
IssuesEvent
2015-12-25 19:40:24
TehNut/ResourcefulCrops
https://api.github.com/repos/TehNut/ResourcefulCrops
closed
[1.7.10] Lighting glitch
bug code complete
Something weird is happening on the border of chunks. It depends on the plants growing just next to the glitched ones. ![2015-11-28_13 47 24](https://cloud.githubusercontent.com/assets/8314356/11451941/a5b80a3a-95d6-11e5-8fd7-1b6b0485a6f4.png) My mods directory: [pic](http://i.imgur.com/Kn09LGM.png) Forge 10.13.4.1558-1.7.10 .. also Iron Seeds? not Iron Plant/Crop?
1.0
[1.7.10] Lighting glitch - Something weird is happening on the border of chunks. It depends on the plants growing just next to the glitched ones. ![2015-11-28_13 47 24](https://cloud.githubusercontent.com/assets/8314356/11451941/a5b80a3a-95d6-11e5-8fd7-1b6b0485a6f4.png) My mods directory: [pic](http://i.imgur.com/Kn09LGM.png) Forge 10.13.4.1558-1.7.10 .. also Iron Seeds? not Iron Plant/Crop?
non_process
lighting glitch something weird is happening on the border of chunks it depends on the plants growing just next to the glitched ones my mods directory forge also iron seeds not iron plant crop
0
34,087
16,430,618,129
IssuesEvent
2021-05-20 00:42:09
crate-ci/typos
https://api.github.com/repos/crate-ci/typos
closed
Speed up varcom support
Performance
When there are no hits, its only about a 10% cost. When there are hits, it can go up to 50% For "code" ``` check_file/FoundFiles/code time: [23.589 us 23.753 us 23.933 us] thrpt: [12.114 MiB/s 12.205 MiB/s 12.290 MiB/s] check_file/Identifiers/code time: [26.211 us 26.350 us 26.498 us] thrpt: [10.941 MiB/s 11.003 MiB/s 11.061 MiB/s] check_file/Words/code time: [28.795 us 28.913 us 29.046 us] thrpt: [9.9814 MiB/s 10.027 MiB/s 10.068 MiB/s] check_file/Typos/code time: [32.651 us 32.788 us 32.934 us] thrpt: [8.8029 MiB/s 8.8421 MiB/s 8.8794 MiB/s] and with varcon check_file/Typos/code time: [35.860 us 36.021 us 36.187 us] thrpt: [8.0117 MiB/s 8.0486 MiB/s 8.0846 MiB/s] change: time: [+8.6784% +9.7190% +10.748%] (p = 0.00 < 0.05) thrpt: [-9.7049% -8.8581% -7.9854%] Performance has regressed. ``` For "corpus" (high token count compared to non-tokens, lots of corrections) ``` check_file/FoundFiles/corpus time: [53.392 us 53.820 us 54.261 us] thrpt: [10.464 GiB/s 10.550 GiB/s 10.635 GiB/s] check_file/Identifiers/corpus time: [2.5148 ms 2.5232 ms 2.5327 ms] thrpt: [229.57 MiB/s 230.44 MiB/s 231.21 MiB/s] check_file/Words/corpus time: [6.5589 ms 6.5755 ms 6.5940 ms] thrpt: [88.177 MiB/s 88.425 MiB/s 88.649 MiB/s] check_file/Typos/corpus time: [17.806 ms 17.900 ms 18.008 ms] thrpt: [32.288 MiB/s 32.482 MiB/s 32.654 MiB/s] and with varcon check_file/Typos/corpus time: [26.966 ms 27.215 ms 27.521 ms] thrpt: [21.127 MiB/s 21.365 MiB/s 21.562 MiB/s] change: time: [+50.409% +52.035% +53.975%] (p = 0.00 < 0.05) thrpt: [-35.054% -34.226% -33.515%] Performance has regressed. ```
True
Speed up varcom support - When there are no hits, its only about a 10% cost. When there are hits, it can go up to 50% For "code" ``` check_file/FoundFiles/code time: [23.589 us 23.753 us 23.933 us] thrpt: [12.114 MiB/s 12.205 MiB/s 12.290 MiB/s] check_file/Identifiers/code time: [26.211 us 26.350 us 26.498 us] thrpt: [10.941 MiB/s 11.003 MiB/s 11.061 MiB/s] check_file/Words/code time: [28.795 us 28.913 us 29.046 us] thrpt: [9.9814 MiB/s 10.027 MiB/s 10.068 MiB/s] check_file/Typos/code time: [32.651 us 32.788 us 32.934 us] thrpt: [8.8029 MiB/s 8.8421 MiB/s 8.8794 MiB/s] and with varcon check_file/Typos/code time: [35.860 us 36.021 us 36.187 us] thrpt: [8.0117 MiB/s 8.0486 MiB/s 8.0846 MiB/s] change: time: [+8.6784% +9.7190% +10.748%] (p = 0.00 < 0.05) thrpt: [-9.7049% -8.8581% -7.9854%] Performance has regressed. ``` For "corpus" (high token count compared to non-tokens, lots of corrections) ``` check_file/FoundFiles/corpus time: [53.392 us 53.820 us 54.261 us] thrpt: [10.464 GiB/s 10.550 GiB/s 10.635 GiB/s] check_file/Identifiers/corpus time: [2.5148 ms 2.5232 ms 2.5327 ms] thrpt: [229.57 MiB/s 230.44 MiB/s 231.21 MiB/s] check_file/Words/corpus time: [6.5589 ms 6.5755 ms 6.5940 ms] thrpt: [88.177 MiB/s 88.425 MiB/s 88.649 MiB/s] check_file/Typos/corpus time: [17.806 ms 17.900 ms 18.008 ms] thrpt: [32.288 MiB/s 32.482 MiB/s 32.654 MiB/s] and with varcon check_file/Typos/corpus time: [26.966 ms 27.215 ms 27.521 ms] thrpt: [21.127 MiB/s 21.365 MiB/s 21.562 MiB/s] change: time: [+50.409% +52.035% +53.975%] (p = 0.00 < 0.05) thrpt: [-35.054% -34.226% -33.515%] Performance has regressed. ```
non_process
speed up varcom support when there are no hits its only about a cost when there are hits it can go up to for code check file foundfiles code time thrpt check file identifiers code time thrpt check file words code time thrpt check file typos code time thrpt and with varcon check file typos code time thrpt change time p thrpt performance has regressed for corpus high token count compared to non tokens lots of corrections check file foundfiles corpus time thrpt check file identifiers corpus time thrpt check file words corpus time thrpt check file typos corpus time thrpt and with varcon check file typos corpus time thrpt change time p thrpt performance has regressed
0
728,562
25,084,625,184
IssuesEvent
2022-11-07 22:26:52
bounswe/bounswe2022group6
https://api.github.com/repos/bounswe/bounswe2022group6
closed
Researching Location API Implementation
Priority: Medium State: In Progress Type: Development Backend
External tools and frameworks for implementing Location API should be researched for requirements [1.1.1.2.5](https://github.com/bounswe/bounswe2022group6/wiki/Requirements#1112-adding-information-to-an-account), [1.1.1.3.7](https://github.com/bounswe/bounswe2022group6/wiki/Requirements#1113-editing-the-information-in-an-account), and [1.1.1.4.7](https://github.com/bounswe/bounswe2022group6/wiki/Requirements#1114-removing-information-from-an-account). Also, location feature for posts should be included after relevant requirements have been added. * Location information should be in this precision: Country/(State)/City/(District). Information in parentheses are optional for implementation. <b>Deadline:</b> 08.11.2022
1.0
Researching Location API Implementation - External tools and frameworks for implementing Location API should be researched for requirements [1.1.1.2.5](https://github.com/bounswe/bounswe2022group6/wiki/Requirements#1112-adding-information-to-an-account), [1.1.1.3.7](https://github.com/bounswe/bounswe2022group6/wiki/Requirements#1113-editing-the-information-in-an-account), and [1.1.1.4.7](https://github.com/bounswe/bounswe2022group6/wiki/Requirements#1114-removing-information-from-an-account). Also, location feature for posts should be included after relevant requirements have been added. * Location information should be in this precision: Country/(State)/City/(District). Information in parentheses are optional for implementation. <b>Deadline:</b> 08.11.2022
non_process
researching location api implementation external tools and frameworks for implementing location api should be researched for requirements and also location feature for posts should be included after relevant requirements have been added location information should be in this precision country state city district information in parentheses are optional for implementation deadline
0
9,114
12,194,734,312
IssuesEvent
2020-04-29 16:14:39
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
Figure out automated CNI for multi-node clusters
area/cni co/multinode kind/process priority/important-soon
Currently multinode clusters need a CNI applied manually to make `minikube service` work. We should decide what CNI will be the canonical CNI choice to apply automatically for multinode and actually apply it on start or add node.
1.0
Figure out automated CNI for multi-node clusters - Currently multinode clusters need a CNI applied manually to make `minikube service` work. We should decide what CNI will be the canonical CNI choice to apply automatically for multinode and actually apply it on start or add node.
process
figure out automated cni for multi node clusters currently multinode clusters need a cni applied manually to make minikube service work we should decide what cni will be the canonical cni choice to apply automatically for multinode and actually apply it on start or add node
1
215,390
24,164,959,893
IssuesEvent
2022-09-22 14:24:12
SmartBear/ready-aws-plugin
https://api.github.com/repos/SmartBear/ready-aws-plugin
closed
CVE-2017-12626 (High) detected in poi-3.10-FINAL.jar - autoclosed
security vulnerability
## CVE-2017-12626 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>poi-3.10-FINAL.jar</b></p></summary> <p>Apache POI - Java API To Access Microsoft Format Files</p> <p>Library home page: <a href="http://poi.apache.org/">http://poi.apache.org/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/poi/poi/3.10-FINAL/poi-3.10-FINAL.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.3.0.jar (Root Library) - :x: **poi-3.10-FINAL.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/ready-aws-plugin/commit/3df791878e4133498f61805bd442d5e371c981f7">3df791878e4133498f61805bd442d5e371c981f7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache POI in versions prior to release 3.17 are vulnerable to Denial of Service Attacks: 1) Infinite Loops while parsing crafted WMF, EMF, MSG and macros (POI bugs 61338 and 61294), and 2) Out of Memory Exceptions while parsing crafted DOC, PPT and XLS (POI bugs 52372 and 61295). <p>Publish Date: 2018-01-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-12626>CVE-2017-12626</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/453d9af5dbabaccd9afb58d27279a9dbfe8e35f4e5ea1645ddd6960b@%3Cdev.poi.apache.org%3E">https://lists.apache.org/thread.html/453d9af5dbabaccd9afb58d27279a9dbfe8e35f4e5ea1645ddd6960b@%3Cdev.poi.apache.org%3E</a></p> <p>Release Date: 2018-01-29</p> <p>Fix Resolution: org.apache.poi:poi-scratchpad:3.17,org.apache.poi:poi:3.17</p> </p> </details> <p></p>
True
CVE-2017-12626 (High) detected in poi-3.10-FINAL.jar - autoclosed - ## CVE-2017-12626 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>poi-3.10-FINAL.jar</b></p></summary> <p>Apache POI - Java API To Access Microsoft Format Files</p> <p>Library home page: <a href="http://poi.apache.org/">http://poi.apache.org/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/poi/poi/3.10-FINAL/poi-3.10-FINAL.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.3.0.jar (Root Library) - :x: **poi-3.10-FINAL.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/ready-aws-plugin/commit/3df791878e4133498f61805bd442d5e371c981f7">3df791878e4133498f61805bd442d5e371c981f7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache POI in versions prior to release 3.17 are vulnerable to Denial of Service Attacks: 1) Infinite Loops while parsing crafted WMF, EMF, MSG and macros (POI bugs 61338 and 61294), and 2) Out of Memory Exceptions while parsing crafted DOC, PPT and XLS (POI bugs 52372 and 61295). <p>Publish Date: 2018-01-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-12626>CVE-2017-12626</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://lists.apache.org/thread.html/453d9af5dbabaccd9afb58d27279a9dbfe8e35f4e5ea1645ddd6960b@%3Cdev.poi.apache.org%3E">https://lists.apache.org/thread.html/453d9af5dbabaccd9afb58d27279a9dbfe8e35f4e5ea1645ddd6960b@%3Cdev.poi.apache.org%3E</a></p> <p>Release Date: 2018-01-29</p> <p>Fix Resolution: org.apache.poi:poi-scratchpad:3.17,org.apache.poi:poi:3.17</p> </p> </details> <p></p>
non_process
cve high detected in poi final jar autoclosed cve high severity vulnerability vulnerable library poi final jar apache poi java api to access microsoft format files library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org apache poi poi final poi final jar dependency hierarchy ready api soapui pro jar root library x poi final jar vulnerable library found in head commit a href found in base branch master vulnerability details apache poi in versions prior to release are vulnerable to denial of service attacks infinite loops while parsing crafted wmf emf msg and macros poi bugs and and out of memory exceptions while parsing crafted doc ppt and xls poi bugs and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache poi poi scratchpad org apache poi poi
0
289,867
21,794,213,540
IssuesEvent
2022-05-15 11:33:30
Pradumnasaraf/open-source-with-pradumna
https://api.github.com/repos/Pradumnasaraf/open-source-with-pradumna
opened
[DOCS] Add process/guide - "How to raise an issue"
documentation
### Description #### Hey! Contributor, #### We are adding a guide/process for every activity (like Creating a PR or raising an issue) and making learning easy for new contributors. ### Chnages/Action required - Add a complete process of **"How to raise an issue"**, steps will contain a text, screenshot, Video (it can be a screen recording to show the steps), and GIF (if needed) - Path of the file in which **steps** need to be added - [`open-source-with-pradumna/pages/How-to/guide/rasing-issue.md`](https://github.com/Pradumnasaraf/open-source-with-pradumna/blob/main/pages/How-to/guide/rasing-issue.md) --- If you have any suggestions feel free to Open an [Issue](https://github.com/Pradumnasaraf/open-source-with-pradumna/issues) ### Screenshots #### This addition of the documentation will get added and hyperlinked in [`How-to/README.md`](https://github.com/Pradumnasaraf/open-source-with-pradumna/tree/main/pages/How-to) and also get hosted on the website https://Opensource.pradumnasaraf.co ![Screenshot from 2022-05-15 14-11-21](https://user-images.githubusercontent.com/51878265/168464628-905b0a25-627b-4560-b8e4-bd09940d8c7b.png)
1.0
[DOCS] Add process/guide - "How to raise an issue" - ### Description #### Hey! Contributor, #### We are adding a guide/process for every activity (like Creating a PR or raising an issue) and making learning easy for new contributors. ### Chnages/Action required - Add a complete process of **"How to raise an issue"**, steps will contain a text, screenshot, Video (it can be a screen recording to show the steps), and GIF (if needed) - Path of the file in which **steps** need to be added - [`open-source-with-pradumna/pages/How-to/guide/rasing-issue.md`](https://github.com/Pradumnasaraf/open-source-with-pradumna/blob/main/pages/How-to/guide/rasing-issue.md) --- If you have any suggestions feel free to Open an [Issue](https://github.com/Pradumnasaraf/open-source-with-pradumna/issues) ### Screenshots #### This addition of the documentation will get added and hyperlinked in [`How-to/README.md`](https://github.com/Pradumnasaraf/open-source-with-pradumna/tree/main/pages/How-to) and also get hosted on the website https://Opensource.pradumnasaraf.co ![Screenshot from 2022-05-15 14-11-21](https://user-images.githubusercontent.com/51878265/168464628-905b0a25-627b-4560-b8e4-bd09940d8c7b.png)
non_process
add process guide how to raise an issue description hey contributor we are adding a guide process for every activity like creating a pr or raising an issue and making learning easy for new contributors chnages action required add a complete process of how to raise an issue steps will contain a text screenshot video it can be a screen recording to show the steps and gif if needed path of the file in which steps need to be added if you have any suggestions feel free to open an screenshots this addition of the documentation will get added and hyperlinked in and also get hosted on the website
0
16,128
6,107,323,302
IssuesEvent
2017-06-21 07:47:20
opencv/opencv
https://api.github.com/repos/opencv/opencv
closed
NEON doesn't get enabled on ARM
bug category: build/install
<!-- If you have a question rather than reporting a bug please go to http://answers.opencv.org where you get much faster responses. If you need further assistance please read [How To Contribute](https://github.com/opencv/opencv/wiki/How_to_contribute). This is a template helping you to create an issue which can be processed as quickly as possible. This is the bug reporting section for the OpenCV library. --> ##### System information (version) <!-- Example - OpenCV => 3.1 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2015 --> - OpenCV => master ( b21b6944447c95e364433c924dd560a16b91bffb ) - Operating System / Platform => Linux ARM32bit / Linux ARM64bit - Compiler => GCC 4.8.4, 5.4.0 ##### Detailed description * CPU dispatch option is introduced, but it keeps blocking to activate the NEON on ARM platform * The main problem comes from obsolete options ```ENABLE_NEON``` [here](https://github.com/tomoaki0705/opencv/blob/b21b6944447c95e364433c924dd560a16b91bffb/CMakeLists.txt#L305-L306 ) ```cmake OCV_OPTION(ENABLE_NEON "Enable NEON instructions" "${NEON}" IF CMAKE_COMPILER_IS_GNUCXX AND (ARM OR AARCH64 OR IOS) ) OCV_OPTION(ENABLE_VFPV3 "Enable VFPv3-D32 instructions" OFF IF CMAKE_COMPILER_IS_GNUCXX AND (ARM OR AARCH64 OR IOS) ) ``` * and [here](https://github.com/tomoaki0705/opencv/blob/b21b6944447c95e364433c924dd560a16b91bffb/cmake/OpenCVCompilerOptimizations.cmake#L79-L80) ```cmake ocv_optimization_process_obsolete_option(ENABLE_VFPV3 VFPV3 OFF) ocv_optimization_process_obsolete_option(ENABLE_NEON NEON OFF) ``` * This macro ```ocv_optimization_process_obsolete_option``` will check the obsolete option and pass it to the "new" dispatch option, but I think [there is a bug in this macro ](https://github.com/tomoaki0705/opencv/blob/b21b6944447c95e364433c924dd560a16b91bffb/cmake/OpenCVCompilerOptimizations.cmake#L57 ) which ends up always switching flags to OFF ```cmake if(${legacy_flag}) # !!!! this will not be evaluated as ON/OFF !!!! if(NOT ";${CPU_BASELINE_REQUIRE};" MATCHES ";${OPT};") set(CPU_BASELINE_REQUIRE "${CPU_BASELINE_REQUIRE};${OPT}" CACHE STRING "${HELP_CPU_BASELINE_REQUIRE}" FORCE) endif() else() if(NOT ";${CPU_BASELINE_DISABLE};" MATCHES ";${OPT};") set(CPU_BASELINE_DISABLE "${CPU_BASELINE_DISABLE};${OPT}" CACHE STRING "${HELP_CPU_BASELINE_DISABLE}" FORCE) endif() endif() ``` * as a conclusion, the ```NEON``` flag and ```FP16``` flags always become false because those flag always go to ```CPU_BASELINE_DISABLE``` option. * I'll push a fix for this issue. ##### Steps to reproduce * on ARM 32bit (Jetson TK1) ``` cmake -DCPU_BASELINE=NEON .. ``` * on ARM 64bit (Jetson TX1) ``` cmake .. ```
1.0
NEON doesn't get enabled on ARM - <!-- If you have a question rather than reporting a bug please go to http://answers.opencv.org where you get much faster responses. If you need further assistance please read [How To Contribute](https://github.com/opencv/opencv/wiki/How_to_contribute). This is a template helping you to create an issue which can be processed as quickly as possible. This is the bug reporting section for the OpenCV library. --> ##### System information (version) <!-- Example - OpenCV => 3.1 - Operating System / Platform => Windows 64 Bit - Compiler => Visual Studio 2015 --> - OpenCV => master ( b21b6944447c95e364433c924dd560a16b91bffb ) - Operating System / Platform => Linux ARM32bit / Linux ARM64bit - Compiler => GCC 4.8.4, 5.4.0 ##### Detailed description * CPU dispatch option is introduced, but it keeps blocking to activate the NEON on ARM platform * The main problem comes from obsolete options ```ENABLE_NEON``` [here](https://github.com/tomoaki0705/opencv/blob/b21b6944447c95e364433c924dd560a16b91bffb/CMakeLists.txt#L305-L306 ) ```cmake OCV_OPTION(ENABLE_NEON "Enable NEON instructions" "${NEON}" IF CMAKE_COMPILER_IS_GNUCXX AND (ARM OR AARCH64 OR IOS) ) OCV_OPTION(ENABLE_VFPV3 "Enable VFPv3-D32 instructions" OFF IF CMAKE_COMPILER_IS_GNUCXX AND (ARM OR AARCH64 OR IOS) ) ``` * and [here](https://github.com/tomoaki0705/opencv/blob/b21b6944447c95e364433c924dd560a16b91bffb/cmake/OpenCVCompilerOptimizations.cmake#L79-L80) ```cmake ocv_optimization_process_obsolete_option(ENABLE_VFPV3 VFPV3 OFF) ocv_optimization_process_obsolete_option(ENABLE_NEON NEON OFF) ``` * This macro ```ocv_optimization_process_obsolete_option``` will check the obsolete option and pass it to the "new" dispatch option, but I think [there is a bug in this macro ](https://github.com/tomoaki0705/opencv/blob/b21b6944447c95e364433c924dd560a16b91bffb/cmake/OpenCVCompilerOptimizations.cmake#L57 ) which ends up always switching flags to OFF ```cmake if(${legacy_flag}) # !!!! this will not be evaluated as ON/OFF !!!! if(NOT ";${CPU_BASELINE_REQUIRE};" MATCHES ";${OPT};") set(CPU_BASELINE_REQUIRE "${CPU_BASELINE_REQUIRE};${OPT}" CACHE STRING "${HELP_CPU_BASELINE_REQUIRE}" FORCE) endif() else() if(NOT ";${CPU_BASELINE_DISABLE};" MATCHES ";${OPT};") set(CPU_BASELINE_DISABLE "${CPU_BASELINE_DISABLE};${OPT}" CACHE STRING "${HELP_CPU_BASELINE_DISABLE}" FORCE) endif() endif() ``` * as a conclusion, the ```NEON``` flag and ```FP16``` flags always become false because those flag always go to ```CPU_BASELINE_DISABLE``` option. * I'll push a fix for this issue. ##### Steps to reproduce * on ARM 32bit (Jetson TK1) ``` cmake -DCPU_BASELINE=NEON .. ``` * on ARM 64bit (Jetson TX1) ``` cmake .. ```
non_process
neon doesn t get enabled on arm if you have a question rather than reporting a bug please go to where you get much faster responses if you need further assistance please read this is a template helping you to create an issue which can be processed as quickly as possible this is the bug reporting section for the opencv library system information version example opencv operating system platform windows bit compiler visual studio opencv master operating system platform linux linux compiler gcc detailed description cpu dispatch option is introduced but it keeps blocking to activate the neon on arm platform the main problem comes from obsolete options enable neon cmake ocv option enable neon enable neon instructions neon if cmake compiler is gnucxx and arm or or ios ocv option enable enable instructions off if cmake compiler is gnucxx and arm or or ios and cmake ocv optimization process obsolete option enable off ocv optimization process obsolete option enable neon neon off this macro ocv optimization process obsolete option will check the obsolete option and pass it to the new dispatch option but i think which ends up always switching flags to off cmake if legacy flag this will not be evaluated as on off if not cpu baseline require matches opt set cpu baseline require cpu baseline require opt cache string help cpu baseline require force endif else if not cpu baseline disable matches opt set cpu baseline disable cpu baseline disable opt cache string help cpu baseline disable force endif endif as a conclusion the neon flag and flags always become false because those flag always go to cpu baseline disable option i ll push a fix for this issue steps to reproduce on arm jetson cmake dcpu baseline neon on arm jetson cmake
0
115,418
4,668,705,958
IssuesEvent
2016-10-06 04:17:32
GovLab/noi3
https://api.github.com/repos/GovLab/noi3
closed
back to my profile
Priority 1
If i click on a user from search results or match me and i'm looking at their profile, there's no way to go back to my profile from there
1.0
back to my profile - If i click on a user from search results or match me and i'm looking at their profile, there's no way to go back to my profile from there
non_process
back to my profile if i click on a user from search results or match me and i m looking at their profile there s no way to go back to my profile from there
0
968
3,422,993,353
IssuesEvent
2015-12-09 02:34:02
MaretEngineering/MROV
https://api.github.com/repos/MaretEngineering/MROV
closed
Resizable Processing Window
enhancement Processing wontfix
Minor issue, but we should probably make the window resizable or at least have terms relative to one another so you only need to change the size in one place. An example: ```Java text("^", 614, 300); ```
1.0
Resizable Processing Window - Minor issue, but we should probably make the window resizable or at least have terms relative to one another so you only need to change the size in one place. An example: ```Java text("^", 614, 300); ```
process
resizable processing window minor issue but we should probably make the window resizable or at least have terms relative to one another so you only need to change the size in one place an example java text
1
862
2,551,902,215
IssuesEvent
2015-02-02 13:44:47
kogarashisan/LiquidLava
https://api.github.com/repos/kogarashisan/LiquidLava
opened
Ability to filter class trees
Category: Documentation P2 T4 Type: Feature
Take functionality from FilteredTree example. Should display "nothing found" when there are no records. Should automatically expand folders with only one item inside them.
1.0
Ability to filter class trees - Take functionality from FilteredTree example. Should display "nothing found" when there are no records. Should automatically expand folders with only one item inside them.
non_process
ability to filter class trees take functionality from filteredtree example should display nothing found when there are no records should automatically expand folders with only one item inside them
0
58,107
14,238,456,902
IssuesEvent
2020-11-18 18:39:22
MyEnroll/Prototypes
https://api.github.com/repos/MyEnroll/Prototypes
closed
WS-2019-0332 (Medium) detected in handlebars-4.0.5.min.js - autoclosed
security vulnerability
## WS-2019-0332 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.min.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js</a></p> <p>Path to dependency file: Prototypes/GetEmail/index.html</p> <p>Path to vulnerable library: Prototypes/GetEmail/index.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.0.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MyEnroll/Prototypes/commit/ca966e3e2dc52111bcbb4d16c9bbc0b48dec9f80">ca966e3e2dc52111bcbb4d16c9bbc0b48dec9f80</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331. <p>Publish Date: 2019-11-17 <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7>WS-2019-0332</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p> <p>Release Date: 2019-12-05</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"handlebars.js","packageVersion":"4.0.5","isTransitiveDependency":false,"dependencyTree":"handlebars.js:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.5.3"}],"vulnerabilityIdentifier":"WS-2019-0332","vulnerabilityDetails":"Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331.","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
True
WS-2019-0332 (Medium) detected in handlebars-4.0.5.min.js - autoclosed - ## WS-2019-0332 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.min.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js</a></p> <p>Path to dependency file: Prototypes/GetEmail/index.html</p> <p>Path to vulnerable library: Prototypes/GetEmail/index.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.0.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MyEnroll/Prototypes/commit/ca966e3e2dc52111bcbb4d16c9bbc0b48dec9f80">ca966e3e2dc52111bcbb4d16c9bbc0b48dec9f80</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331. <p>Publish Date: 2019-11-17 <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7>WS-2019-0332</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p> <p>Release Date: 2019-12-05</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"handlebars.js","packageVersion":"4.0.5","isTransitiveDependency":false,"dependencyTree":"handlebars.js:4.0.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"handlebars - 4.5.3"}],"vulnerabilityIdentifier":"WS-2019-0332","vulnerabilityDetails":"Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331.","vulnerabilityUrl":"https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
non_process
ws medium detected in handlebars min js autoclosed ws medium severity vulnerability vulnerable library handlebars min js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file prototypes getemail index html path to vulnerable library prototypes getemail index html dependency hierarchy x handlebars min js vulnerable library found in head commit a href found in base branch master vulnerability details arbitrary code execution vulnerability found in handlebars before lookup helper fails to validate templates attack may submit templates that execute arbitrary javascript in the system it is due to an incomplete fix for a ws publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution handlebars isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails arbitrary code execution vulnerability found in handlebars before lookup helper fails to validate templates attack may submit templates that execute arbitrary javascript in the system it is due to an incomplete fix for a ws vulnerabilityurl
0
3,293
2,754,458,354
IssuesEvent
2015-04-25 17:56:26
mquandalle/meteor-jade
https://api.github.com/repos/mquandalle/meteor-jade
closed
Dead link to Components Inside Tags in README.md
bug documentation
The webpage `https://github.com/mquandalle/meteor-jade/#components-inside-tags` doesn't exist (or at least doesn't seem to for me), but its linked to in `https://github.com/mquandalle/meteor-jade/#anonymous-helper`. Or am I missing something here?
1.0
Dead link to Components Inside Tags in README.md - The webpage `https://github.com/mquandalle/meteor-jade/#components-inside-tags` doesn't exist (or at least doesn't seem to for me), but its linked to in `https://github.com/mquandalle/meteor-jade/#anonymous-helper`. Or am I missing something here?
non_process
dead link to components inside tags in readme md the webpage doesn t exist or at least doesn t seem to for me but its linked to in or am i missing something here
0
3,137
6,189,746,766
IssuesEvent
2017-07-04 13:42:18
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
closed
[sct_register_multimodal] restrict deformation optional
card:WORK_IN_PROCESS enhancement priority:MEDIUM sct_register_multimodal
Please make `restrict deformation` an optional argument. Atef is using Ants for the Brain, sct_register_multimodal is very useful (notably because AntsRegister provides no feedbacks on user), but I had to remove `Line 639: --restrict-deformation 1x1x0 `
1.0
[sct_register_multimodal] restrict deformation optional - Please make `restrict deformation` an optional argument. Atef is using Ants for the Brain, sct_register_multimodal is very useful (notably because AntsRegister provides no feedbacks on user), but I had to remove `Line 639: --restrict-deformation 1x1x0 `
process
restrict deformation optional please make restrict deformation an optional argument atef is using ants for the brain sct register multimodal is very useful notably because antsregister provides no feedbacks on user but i had to remove line restrict deformation
1
6,603
9,684,275,419
IssuesEvent
2019-05-23 13:26:43
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
1.14 Release Checklist
area: Process
Add items as required -------- - [x] Major enhancements: - [x] Pre RC1 Steps - [x] Verify that all external components and external dependencies are up to date, for example - mbedTLS - tinycrypt - FatFS - [x] Check known vulnerabilities, and fix exploitable vulnerabilities or verify them as unexploitable. See https://github.com/zephyrproject-rtos/zephyr/wiki/Release-procedure for tagging details -------- - [x] Release Notes WIP - [x] Create draft doc/release-notes-1.14.rst with feature summary from https://www.zephyrproject.org/developers/#releases-overview - [x] Update doc/release-notes-1.14.rst details sections (relative to 1.13.0 release) as important issues are addressed (subsystem owners) -------- - [x] Release data - [x] Update version in VERSION - [x] Run sanitycheck with --all --release and checkin resulting file scripts/sanity_chk/sanity_last_release.csv -------- - [x] Finalize Release Notes - [x] Add list of GitHub issues (spell check issue list and fix in GitHub as needed) - [x] Review doc/release-notes/release-notes-1.14.rst overview summary and details sections (see Sign-off below) - [x] Update doc/release-notes/index.rst (add reference to 1.14 notes) - [x] Update doc/index.rst (add reference to doc/1.14.0) - [x] Update doc/conf.py (add version to pick list) - [x] Update doc/LICENSING.rst for new components that do not have a license - [x] Create release and add notes on https://github.com/zephyrproject-rtos/zephyr/releases - [x] Final tagged release as zephyr-v1.14.0 (with release notes) - [x] Update https://www.zephyrproject.org/developers/#releases-overview - [x] Update https://www.zephyrproject.org/developers/#downloads (and git checkout version example) - [x] Update doc generation (in future should be automatic based on tag) ping @nashif - [x] Email devel@lists.zephyrproject.org & announce@lists.zephyrproject.org lists about release -------- - [x] Final Release Notes Sign-off - Subsystems - [x] Networking - @jukkar - [x] Bluetooth - @jhedberg , @carlescufi - Architectures - [x] ARC - @ruuddw - [x] ARM - @galak, @MaureenHelm - [x] X86 - @nashif, @andrewboie - [x] Xtensa, RISCV32, NIOS-II - @nashif, @andrewboie - [x] Security - @agross-linaro -------- - [x] After Release - [x] Update patchlevel in Makefile (PATCHLEVEL = 99) - [x] Marketing Blog for announcing release (https://www.zephyrproject.org/blog) - [x] Create branch for 1.14.x releases, update github protections for branch - [x] Create checklist issue for next release - [x] Add a comment to this issue linking to the next release checklist issue - [x] Update https://www.zephyrproject.org with latest stats from Github [@dbkinder] - [x] Move all PRs/Issues in the 1.14 milestone to the next milestone - [x] Close v1.14.0 milestone - [x] Create v1.14.1 milestone
1.0
1.14 Release Checklist - Add items as required -------- - [x] Major enhancements: - [x] Pre RC1 Steps - [x] Verify that all external components and external dependencies are up to date, for example - mbedTLS - tinycrypt - FatFS - [x] Check known vulnerabilities, and fix exploitable vulnerabilities or verify them as unexploitable. See https://github.com/zephyrproject-rtos/zephyr/wiki/Release-procedure for tagging details -------- - [x] Release Notes WIP - [x] Create draft doc/release-notes-1.14.rst with feature summary from https://www.zephyrproject.org/developers/#releases-overview - [x] Update doc/release-notes-1.14.rst details sections (relative to 1.13.0 release) as important issues are addressed (subsystem owners) -------- - [x] Release data - [x] Update version in VERSION - [x] Run sanitycheck with --all --release and checkin resulting file scripts/sanity_chk/sanity_last_release.csv -------- - [x] Finalize Release Notes - [x] Add list of GitHub issues (spell check issue list and fix in GitHub as needed) - [x] Review doc/release-notes/release-notes-1.14.rst overview summary and details sections (see Sign-off below) - [x] Update doc/release-notes/index.rst (add reference to 1.14 notes) - [x] Update doc/index.rst (add reference to doc/1.14.0) - [x] Update doc/conf.py (add version to pick list) - [x] Update doc/LICENSING.rst for new components that do not have a license - [x] Create release and add notes on https://github.com/zephyrproject-rtos/zephyr/releases - [x] Final tagged release as zephyr-v1.14.0 (with release notes) - [x] Update https://www.zephyrproject.org/developers/#releases-overview - [x] Update https://www.zephyrproject.org/developers/#downloads (and git checkout version example) - [x] Update doc generation (in future should be automatic based on tag) ping @nashif - [x] Email devel@lists.zephyrproject.org & announce@lists.zephyrproject.org lists about release -------- - [x] Final Release Notes Sign-off - Subsystems - [x] Networking - @jukkar - [x] Bluetooth - @jhedberg , @carlescufi - Architectures - [x] ARC - @ruuddw - [x] ARM - @galak, @MaureenHelm - [x] X86 - @nashif, @andrewboie - [x] Xtensa, RISCV32, NIOS-II - @nashif, @andrewboie - [x] Security - @agross-linaro -------- - [x] After Release - [x] Update patchlevel in Makefile (PATCHLEVEL = 99) - [x] Marketing Blog for announcing release (https://www.zephyrproject.org/blog) - [x] Create branch for 1.14.x releases, update github protections for branch - [x] Create checklist issue for next release - [x] Add a comment to this issue linking to the next release checklist issue - [x] Update https://www.zephyrproject.org with latest stats from Github [@dbkinder] - [x] Move all PRs/Issues in the 1.14 milestone to the next milestone - [x] Close v1.14.0 milestone - [x] Create v1.14.1 milestone
process
release checklist add items as required major enhancements pre steps verify that all external components and external dependencies are up to date for example mbedtls tinycrypt fatfs check known vulnerabilities and fix exploitable vulnerabilities or verify them as unexploitable see for tagging details release notes wip create draft doc release notes rst with feature summary from update doc release notes rst details sections relative to release as important issues are addressed subsystem owners release data update version in version run sanitycheck with all release and checkin resulting file scripts sanity chk sanity last release csv finalize release notes add list of github issues spell check issue list and fix in github as needed review doc release notes release notes rst overview summary and details sections see sign off below update doc release notes index rst add reference to notes update doc index rst add reference to doc update doc conf py add version to pick list update doc licensing rst for new components that do not have a license create release and add notes on final tagged release as zephyr with release notes update update and git checkout version example update doc generation in future should be automatic based on tag ping nashif email devel lists zephyrproject org announce lists zephyrproject org lists about release final release notes sign off subsystems networking jukkar bluetooth jhedberg carlescufi architectures arc ruuddw arm galak maureenhelm nashif andrewboie xtensa nios ii nashif andrewboie security agross linaro after release update patchlevel in makefile patchlevel marketing blog for announcing release create branch for x releases update github protections for branch create checklist issue for next release add a comment to this issue linking to the next release checklist issue update with latest stats from github move all prs issues in the milestone to the next milestone close milestone create milestone
1
8,733
11,863,424,329
IssuesEvent
2020-03-25 19:41:54
hashgraph/hedera-mirror-node
https://api.github.com/repos/hashgraph/hedera-mirror-node
opened
Push docker image on commit to master
P2 enhancement process
**Problem** In order to be able to continuously deploy the helm chart to a development environment, we need to package our app in a docker image on every commit so that it can available to pull from a remote machine. **Solution** - Update CircleCI to push docker images with a tag of `master` to GCR if branch is master - Update CircleCI to cleanup unused images so we don't incur increasing costs. Something like below (inspired by [this](https://gist.github.com/ahmetb/7ce6d741bd5baa194a3fac6b1fec8bb7) gist): ``` gcloud container images delete -q --force-delete-tags $(gcloud container images list-tags ${IMAGE} --limit=1000 --sort-by=TIMESTAMP --filter="NOT tags:*" --format='get(digest)') ```
1.0
Push docker image on commit to master - **Problem** In order to be able to continuously deploy the helm chart to a development environment, we need to package our app in a docker image on every commit so that it can available to pull from a remote machine. **Solution** - Update CircleCI to push docker images with a tag of `master` to GCR if branch is master - Update CircleCI to cleanup unused images so we don't incur increasing costs. Something like below (inspired by [this](https://gist.github.com/ahmetb/7ce6d741bd5baa194a3fac6b1fec8bb7) gist): ``` gcloud container images delete -q --force-delete-tags $(gcloud container images list-tags ${IMAGE} --limit=1000 --sort-by=TIMESTAMP --filter="NOT tags:*" --format='get(digest)') ```
process
push docker image on commit to master problem in order to be able to continuously deploy the helm chart to a development environment we need to package our app in a docker image on every commit so that it can available to pull from a remote machine solution update circleci to push docker images with a tag of master to gcr if branch is master update circleci to cleanup unused images so we don t incur increasing costs something like below inspired by gist gcloud container images delete q force delete tags gcloud container images list tags image limit sort by timestamp filter not tags format get digest
1
270,470
23,510,939,015
IssuesEvent
2022-08-18 16:27:56
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission failed
C-test-failure O-robot O-roachtest branch-master T-kv T-kv-replication
roachtest.kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6124150?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6124150?buildTab=artifacts#/kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission) on master @ [b173a16715e71e94115820374da1eb350b3b459d](https://github.com/cockroachdb/cockroach/commits/b173a16715e71e94115820374da1eb350b3b459d): ``` | 1298.0s 0 0.0 532.5 0.0 0.0 0.0 0.0 write | 1299.0s 0 0.0 532.1 0.0 0.0 0.0 0.0 write | 1300.0s 0 0.0 531.6 0.0 0.0 0.0 0.0 write | _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms) | 1301.0s 0 0.0 531.2 0.0 0.0 0.0 0.0 write | 1302.0s 0 0.0 530.8 0.0 0.0 0.0 0.0 write | 1303.0s 0 1.0 530.4 103079.2 103079.2 103079.2 103079.2 write | 1304.0s 0 0.0 530.0 0.0 0.0 0.0 0.0 write | 1305.0s 0 0.0 529.6 0.0 0.0 0.0 0.0 write | 1306.0s 0 0.0 529.2 0.0 0.0 0.0 0.0 write | 1307.0s 0 0.0 528.8 0.0 0.0 0.0 0.0 write Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 4. Command with error: | `````` | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --splits=1000 --duration=30m0s --read-percent=0 --min-block-bytes=65536 --max-block-bytes=65536 {pgurl:1-3} | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError monitor.go:127,kv.go:164,kv.go:305,test_runner.go:896: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:164 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func3 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:305 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:896 Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func2 | main/pkg/cmd/roachtest/monitor.go:171 Wraps: (4) monitor task failed Wraps: (5) attached stack trace -- stack trace: | main.init | main/pkg/cmd/roachtest/monitor.go:80 | runtime.doInit | GOROOT/src/runtime/proc.go:6240 | runtime.main | GOROOT/src/runtime/proc.go:233 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1571 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=32</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/test-eng <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-18679
2.0
roachtest: kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission failed - roachtest.kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6124150?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/6124150?buildTab=artifacts#/kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission) on master @ [b173a16715e71e94115820374da1eb350b3b459d](https://github.com/cockroachdb/cockroach/commits/b173a16715e71e94115820374da1eb350b3b459d): ``` | 1298.0s 0 0.0 532.5 0.0 0.0 0.0 0.0 write | 1299.0s 0 0.0 532.1 0.0 0.0 0.0 0.0 write | 1300.0s 0 0.0 531.6 0.0 0.0 0.0 0.0 write | _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms) | 1301.0s 0 0.0 531.2 0.0 0.0 0.0 0.0 write | 1302.0s 0 0.0 530.8 0.0 0.0 0.0 0.0 write | 1303.0s 0 1.0 530.4 103079.2 103079.2 103079.2 103079.2 write | 1304.0s 0 0.0 530.0 0.0 0.0 0.0 0.0 write | 1305.0s 0 0.0 529.6 0.0 0.0 0.0 0.0 write | 1306.0s 0 0.0 529.2 0.0 0.0 0.0 0.0 write | 1307.0s 0 0.0 528.8 0.0 0.0 0.0 0.0 write Wraps: (4) COMMAND_PROBLEM Wraps: (5) Node 4. Command with error: | `````` | ./workload run kv --init --histograms=perf/stats.json --concurrency=192 --splits=1000 --duration=30m0s --read-percent=0 --min-block-bytes=65536 --max-block-bytes=65536 {pgurl:1-3} | `````` Wraps: (6) exit status 1 Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *cluster.WithCommandDetails (4) errors.Cmd (5) *hintdetail.withDetail (6) *exec.ExitError monitor.go:127,kv.go:164,kv.go:305,test_runner.go:896: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func2 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:164 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func3 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:305 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:896 Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func2 | main/pkg/cmd/roachtest/monitor.go:171 Wraps: (4) monitor task failed Wraps: (5) attached stack trace -- stack trace: | main.init | main/pkg/cmd/roachtest/monitor.go:80 | runtime.doInit | GOROOT/src/runtime/proc.go:6240 | runtime.main | GOROOT/src/runtime/proc.go:233 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1571 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *withstack.withStack (6) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=32</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/test-eng <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv0/enc=false/nodes=3/cpu=32/size=64kb/no-admission.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-18679
non_process
roachtest enc false nodes cpu size no admission failed roachtest enc false nodes cpu size no admission with on master write write write elapsed errors ops sec inst ops sec cum ms ms ms pmax ms write write write write write write write wraps command problem wraps node command with error workload run kv init histograms perf stats json concurrency splits duration read percent min block bytes max block bytes pgurl wraps exit status error types withstack withstack errutil withprefix cluster withcommanddetails errors cmd hintdetail withdetail exec exiterror monitor go kv go kv go test runner go monitor failure monitor task failed t fatal was called attached stack trace stack trace main monitorimpl waite main pkg cmd roachtest monitor go main monitorimpl wait main pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests registerkv github com cockroachdb cockroach pkg cmd roachtest tests kv go github com cockroachdb cockroach pkg cmd roachtest tests registerkv github com cockroachdb cockroach pkg cmd roachtest tests kv go main testrunner runtest main pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace stack trace main monitorimpl wait main pkg cmd roachtest monitor go wraps monitor task failed wraps attached stack trace stack trace main init main pkg cmd roachtest monitor go runtime doinit goroot src runtime proc go runtime main goroot src runtime proc go runtime goexit goroot src runtime asm s wraps t fatal was called error types withstack withstack errutil withprefix withstack withstack errutil withprefix withstack withstack errutil leaferror parameters roachtest cloud gce roachtest cpu roachtest ssd help see see cc cockroachdb test eng jira issue crdb
0
193,264
6,883,196,662
IssuesEvent
2017-11-21 08:30:07
HoneycuttInc/Thorncastle
https://api.github.com/repos/HoneycuttInc/Thorncastle
closed
General: Verify Jerry's edits to the lab guide
Medium Priority
I edited the lab guide in IDL studio to fix a number of bugs. Please compare the version of the markdown in the LODS platform to the version in GitHub to verify my changes.
1.0
General: Verify Jerry's edits to the lab guide - I edited the lab guide in IDL studio to fix a number of bugs. Please compare the version of the markdown in the LODS platform to the version in GitHub to verify my changes.
non_process
general verify jerry s edits to the lab guide i edited the lab guide in idl studio to fix a number of bugs please compare the version of the markdown in the lods platform to the version in github to verify my changes
0
15,722
19,897,253,202
IssuesEvent
2022-01-25 01:25:48
varabyte/kobweb
https://api.github.com/repos/varabyte/kobweb
closed
Separate kobweb binary version from library version
process
Originally, we were furiously changing both the kobweb binary and library at the same time so often it was easier to keep them together. However, more and more, the binary isn't changing that much while the library changes often, and there's no reason in the future, for example, a 0.9.0 kobweb binary couldn't instantiate a project using Kobweb version 0.7.8, or vice versa. At some point, we should add a separate publishing path for the CLI binary, perhaps with a totally different release tag. This doesn't have to happen _today_ but probably should happen when we hit 1.0
1.0
Separate kobweb binary version from library version - Originally, we were furiously changing both the kobweb binary and library at the same time so often it was easier to keep them together. However, more and more, the binary isn't changing that much while the library changes often, and there's no reason in the future, for example, a 0.9.0 kobweb binary couldn't instantiate a project using Kobweb version 0.7.8, or vice versa. At some point, we should add a separate publishing path for the CLI binary, perhaps with a totally different release tag. This doesn't have to happen _today_ but probably should happen when we hit 1.0
process
separate kobweb binary version from library version originally we were furiously changing both the kobweb binary and library at the same time so often it was easier to keep them together however more and more the binary isn t changing that much while the library changes often and there s no reason in the future for example a kobweb binary couldn t instantiate a project using kobweb version or vice versa at some point we should add a separate publishing path for the cli binary perhaps with a totally different release tag this doesn t have to happen today but probably should happen when we hit
1
1,153
5,173,901,313
IssuesEvent
2017-01-18 17:11:01
char-lie/mfm
https://api.github.com/repos/char-lie/mfm
closed
Make Face request synchronous
in progress ☠ major ⚒ enhancement ⛲ architecture
Fitters are hard to read and support. Need to implement `synchronous` `Face` image getter, e. g. ```python new_face_image = self.get_face(face) ``` Possible solutions: - ~AsyncIO may help this, but code will be incompatible with Python <3.5;~ - ~Another solution is to use [events](https://docs.python.org/3/library/threading.html#event-objects);~ - Implement synchronous `Face` drawing for fitting only, because it blocks keyboard/mouse input.
1.0
Make Face request synchronous - Fitters are hard to read and support. Need to implement `synchronous` `Face` image getter, e. g. ```python new_face_image = self.get_face(face) ``` Possible solutions: - ~AsyncIO may help this, but code will be incompatible with Python <3.5;~ - ~Another solution is to use [events](https://docs.python.org/3/library/threading.html#event-objects);~ - Implement synchronous `Face` drawing for fitting only, because it blocks keyboard/mouse input.
non_process
make face request synchronous fitters are hard to read and support need to implement synchronous face image getter e g python new face image self get face face possible solutions asyncio may help this but code will be incompatible with python another solution is to use implement synchronous face drawing for fitting only because it blocks keyboard mouse input
0
2,540
5,300,418,922
IssuesEvent
2017-02-10 04:46:32
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Make docker-import postprocessor show docker stdout/stderr output
bug post-processor/docker
My import failed, and `packer` emitted this last part of the log: ``` ==> docker: Running post-processor: docker-import docker (docker-import): Importing image: Container docker (docker-import): Repository: docker.clarin.eu/alpine__python_3:0.99 Build 'docker' errored: 1 error(s) occurred: * Post-processor failed: Error importing container: exit status 1 ==> Some builds didn't complete successfully and had errors: --> docker: 1 error(s) occurred: * Post-processor failed: Error importing container: exit status 1 ==> Builds finished but no artifacts were created. ``` In the previous, no clue is given by the docker-import postprocessor/packer as to why the import failed. After I issued the import command (the command line invoked by packer based on the source code), `docker` gave me the following helpful error message: ``` repository name component must match "[a-z0-9]+(?:[._-][a-z0-9]+)*" ``` My feature request is to ensure that all output on stdout/stderr emitted by Docker will be emitted by Packer as well.
1.0
Make docker-import postprocessor show docker stdout/stderr output - My import failed, and `packer` emitted this last part of the log: ``` ==> docker: Running post-processor: docker-import docker (docker-import): Importing image: Container docker (docker-import): Repository: docker.clarin.eu/alpine__python_3:0.99 Build 'docker' errored: 1 error(s) occurred: * Post-processor failed: Error importing container: exit status 1 ==> Some builds didn't complete successfully and had errors: --> docker: 1 error(s) occurred: * Post-processor failed: Error importing container: exit status 1 ==> Builds finished but no artifacts were created. ``` In the previous, no clue is given by the docker-import postprocessor/packer as to why the import failed. After I issued the import command (the command line invoked by packer based on the source code), `docker` gave me the following helpful error message: ``` repository name component must match "[a-z0-9]+(?:[._-][a-z0-9]+)*" ``` My feature request is to ensure that all output on stdout/stderr emitted by Docker will be emitted by Packer as well.
process
make docker import postprocessor show docker stdout stderr output my import failed and packer emitted this last part of the log docker running post processor docker import docker docker import importing image container docker docker import repository docker clarin eu alpine python build docker errored error s occurred post processor failed error importing container exit status some builds didn t complete successfully and had errors docker error s occurred post processor failed error importing container exit status builds finished but no artifacts were created in the previous no clue is given by the docker import postprocessor packer as to why the import failed after i issued the import command the command line invoked by packer based on the source code docker gave me the following helpful error message repository name component must match my feature request is to ensure that all output on stdout stderr emitted by docker will be emitted by packer as well
1
16,759
21,927,601,211
IssuesEvent
2022-05-23 06:42:05
RobertCraigie/prisma-client-py
https://api.github.com/repos/RobertCraigie/prisma-client-py
opened
Prisma CLI and Client use different base paths for relative SQLite files
bug/2-confirmed kind/bug process/candidate topic: client level/intermediate priority/high
<!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> With this schema as `prisma/schema.prisma`: ```prisma datasource db { provider = "sqlite" url = "file:dev.db" } model User { id String @id @default(cuid()) name String } ``` Running `prisma db push` creates a database file at `prisma/dev.db`. Now running the client, e.g. ```python import asyncio from prisma import Prisma from prisma.models import User async def bar() -> None: p = Prisma(auto_register=True) await p.connect() await User.prisma().create(data={'name': 'Robert'}) asyncio.run(bar()) ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> The CLI and the Client should use the same database path.
1.0
Prisma CLI and Client use different base paths for relative SQLite files - <!-- Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output. See https://prisma-client-py.readthedocs.io/en/stable/reference/logging/ for how to enable additional logging output. --> ## Bug description <!-- A clear and concise description of what the bug is. --> With this schema as `prisma/schema.prisma`: ```prisma datasource db { provider = "sqlite" url = "file:dev.db" } model User { id String @id @default(cuid()) name String } ``` Running `prisma db push` creates a database file at `prisma/dev.db`. Now running the client, e.g. ```python import asyncio from prisma import Prisma from prisma.models import User async def bar() -> None: p = Prisma(auto_register=True) await p.connect() await User.prisma().create(data={'name': 'Robert'}) asyncio.run(bar()) ``` ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> The CLI and the Client should use the same database path.
process
prisma cli and client use different base paths for relative sqlite files thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description with this schema as prisma schema prisma prisma datasource db provider sqlite url file dev db model user id string id default cuid name string running prisma db push creates a database file at prisma dev db now running the client e g python import asyncio from prisma import prisma from prisma models import user async def bar none p prisma auto register true await p connect await user prisma create data name robert asyncio run bar expected behavior the cli and the client should use the same database path
1
166,025
20,711,380,181
IssuesEvent
2022-03-12 01:14:32
snowflakedb/snowflake-jdbc
https://api.github.com/repos/snowflakedb/snowflake-jdbc
closed
SNOW-558878: CVE-2020-9548 (High) detected in jackson-databind-2.9.8.jar - autoclosed
security vulnerability
## CVE-2020-9548 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml</p> <p>Path to vulnerable library: /sitory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-jdbc/commit/8560bcca9d395d1ee02123536c2e958f7d386fe0">8560bcca9d395d1ee02123536c2e958f7d386fe0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9548>CVE-2020-9548</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.6,2.9.10.4</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.6,2.9.10.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-9548","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9548","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
SNOW-558878: CVE-2020-9548 (High) detected in jackson-databind-2.9.8.jar - autoclosed - ## CVE-2020-9548 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml</p> <p>Path to vulnerable library: /sitory/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowflakedb/snowflake-jdbc/commit/8560bcca9d395d1ee02123536c2e958f7d386fe0">8560bcca9d395d1ee02123536c2e958f7d386fe0</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9548>CVE-2020-9548</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9548</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.6,2.9.10.4</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tmp/ws-ua_20220312003410_KQQKCS/archiveExtraction_BELGGC/FUIDAN/20220312003410/snowflake-jdbc_depth_0/dependencies/arrow-vector-0.15.1/META-INF/maven/org.apache.arrow/arrow-vector/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.7.9.7,2.8.11.6,2.9.10.4","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-9548","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPConfig (aka anteros-core).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9548","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_process
snow cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws ua kqqkcs archiveextraction belggc fuidan snowflake jdbc depth dependencies arrow vector meta inf maven org apache arrow arrow vector pom xml path to vulnerable library sitory com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpconfig aka anteros core publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpconfig aka anteros core vulnerabilityurl
0
9,870
12,882,439,503
IssuesEvent
2020-07-12 16:55:01
OI-wiki/OI-wiki
https://api.github.com/repos/OI-wiki/OI-wiki
opened
规范 MathJax 公式
Discussion / 需要讨论 Format fix needed / 格式需修正 Need Processing / 需要处理 Other Bug / 内容无关的 Bug Work in Progress / 施工中 help wanted / 需要帮助
重构 TeX 导出工具期间发现了不少问题。这些问题导致导出 TeX PDF 时出现版面错误、符号错误等现象。包括但不限于以下: 页面 string/bm.md: > 首先考虑 $delta_1$ 不起作用的情况,也就是发现失配字符在 $pat$ 上重现的位置在已经匹配完的 $m$ 个字符中,这种情况的概率 $\textit{probdelta_1_worthless}$ 为 * 存在滥用 MathJax 公式的现象,甚至单个变量名也要用 $ 套起来; * 命名不规范(TeX 不允许同一字符后连续接两个下标); 页面 geometry/distance.md > $(y_1 - y_2 \lt 0)\rightarrow |x_1-x_2|+|y_1-y_2|=x_1 - y_1 - (x_2 - y_2)$ * \lt 命令与 TeX 不兼容(需要 \newcommand 之类的 dirty hack); 页面 math/permutation_group.md > 对于两个置换 $f=\pmatrix{a_1,a_2,\dots,a_n\\a_{p_1},a_{p_2},\dots,a_{p_n}}$ 和 $g=\pmatrix{a_{p_1},a_{p_2},\dots,a_{p_n}\\a_{q_1},a_{q_2},\dots,a_{q_n}}$ * \pmatrix 是过时写法 (old form),应使用 \begin{pmatrix} ... \end{pmatrix} 代替。 ## 可能需要做的修复 检查目前 OI Wiki 的所有页面,规范公式书写格式,尽可能保证公式同时被 MathJax 和 TeX 接受。 (接受的标准为 TeX 编译器不报 error 和 warning)
1.0
规范 MathJax 公式 - 重构 TeX 导出工具期间发现了不少问题。这些问题导致导出 TeX PDF 时出现版面错误、符号错误等现象。包括但不限于以下: 页面 string/bm.md: > 首先考虑 $delta_1$ 不起作用的情况,也就是发现失配字符在 $pat$ 上重现的位置在已经匹配完的 $m$ 个字符中,这种情况的概率 $\textit{probdelta_1_worthless}$ 为 * 存在滥用 MathJax 公式的现象,甚至单个变量名也要用 $ 套起来; * 命名不规范(TeX 不允许同一字符后连续接两个下标); 页面 geometry/distance.md > $(y_1 - y_2 \lt 0)\rightarrow |x_1-x_2|+|y_1-y_2|=x_1 - y_1 - (x_2 - y_2)$ * \lt 命令与 TeX 不兼容(需要 \newcommand 之类的 dirty hack); 页面 math/permutation_group.md > 对于两个置换 $f=\pmatrix{a_1,a_2,\dots,a_n\\a_{p_1},a_{p_2},\dots,a_{p_n}}$ 和 $g=\pmatrix{a_{p_1},a_{p_2},\dots,a_{p_n}\\a_{q_1},a_{q_2},\dots,a_{q_n}}$ * \pmatrix 是过时写法 (old form),应使用 \begin{pmatrix} ... \end{pmatrix} 代替。 ## 可能需要做的修复 检查目前 OI Wiki 的所有页面,规范公式书写格式,尽可能保证公式同时被 MathJax 和 TeX 接受。 (接受的标准为 TeX 编译器不报 error 和 warning)
process
规范 mathjax 公式 重构 tex 导出工具期间发现了不少问题。这些问题导致导出 tex pdf 时出现版面错误、符号错误等现象。包括但不限于以下: 页面 string bm md: 首先考虑 delta 不起作用的情况,也就是发现失配字符在 pat 上重现的位置在已经匹配完的 m 个字符中,这种情况的概率 textit probdelta worthless 为 存在滥用 mathjax 公式的现象,甚至单个变量名也要用 套起来; 命名不规范(tex 不允许同一字符后连续接两个下标); 页面 geometry distance md y y lt rightarrow x x y y x y x y lt 命令与 tex 不兼容(需要 newcommand 之类的 dirty hack); 页面 math permutation group md 对于两个置换 f pmatrix a a dots a n a p a p dots a p n 和 g pmatrix a p a p dots a p n a q a q dots a q n pmatrix 是过时写法 old form ,应使用 begin pmatrix end pmatrix 代替。 可能需要做的修复 检查目前 oi wiki 的所有页面,规范公式书写格式,尽可能保证公式同时被 mathjax 和 tex 接受。 (接受的标准为 tex 编译器不报 error 和 warning)
1
6,284
9,284,240,541
IssuesEvent
2019-03-21 00:36:16
googleapis/sample-tester
https://api.github.com/repos/googleapis/sample-tester
opened
Test `MultiVisitor`
Type: Process
- [ ] in the usual case - [ ] when some of the subordinate visitors return None (see #19)
1.0
Test `MultiVisitor` - - [ ] in the usual case - [ ] when some of the subordinate visitors return None (see #19)
process
test multivisitor in the usual case when some of the subordinate visitors return none see
1
146,672
13,187,670,500
IssuesEvent
2020-08-13 04:10:46
py-suruga/pycon-jp-2020-tutorial
https://api.github.com/repos/py-suruga/pycon-jp-2020-tutorial
closed
チュートリアルで実際に手を動かしてもらう部分を決める
documentation
実装で作業してもらうときは、答えがあることを前提 --- slackbotはステップ3~4ぐらいが望ましい。天気botは最初から実装がかなり難しいので、触る程度になるかも。(bs4とxmlの説明ぐらいにとどめたい) 1. ハードコードした挨拶bot->botが使う共通インターフェイスを実装して、モジュール化を行う。テストがしやすい環境を整える 2. 挨拶botの世界挨拶の対応: モジュール内に用意したマッピングした挨拶のセット-> 自作のモジュールを使ってみる 3. connpassbot オンラインイベントの検索の実装 -> 外部のAPIを使ってjsonのパースまで。requestsでスクレイピング 4. (説明のみ):天気botの概要: bs4でxmlのパース, 負荷を掛けないための工夫 --- pytestで作業する項目は、挨拶botの世界言語の対応、connpassで予め取得したjsonを使ってのチェック。parametarizeとfixtureの便利さを体験する 1. 挨拶botの各国の回答: parametarize 2. connpassのオンライン回答ができているかの: fixtureでjsonを共通でロードしてテスト 3. (optional):天気botの地域ごとの対応: --- sphinxは構造化されたドキュメントの書き方。最初の使い方をメイン。pythonとの連携としてautodocでできるリファレンスの説明 1. sphinx-quickstartでプロジェクトを作る 2.ある程度できているドキュメントのビルドをして、ファイルを一つ追加して再ビルド 3. autodocを使ってリファレンス生成
1.0
チュートリアルで実際に手を動かしてもらう部分を決める - 実装で作業してもらうときは、答えがあることを前提 --- slackbotはステップ3~4ぐらいが望ましい。天気botは最初から実装がかなり難しいので、触る程度になるかも。(bs4とxmlの説明ぐらいにとどめたい) 1. ハードコードした挨拶bot->botが使う共通インターフェイスを実装して、モジュール化を行う。テストがしやすい環境を整える 2. 挨拶botの世界挨拶の対応: モジュール内に用意したマッピングした挨拶のセット-> 自作のモジュールを使ってみる 3. connpassbot オンラインイベントの検索の実装 -> 外部のAPIを使ってjsonのパースまで。requestsでスクレイピング 4. (説明のみ):天気botの概要: bs4でxmlのパース, 負荷を掛けないための工夫 --- pytestで作業する項目は、挨拶botの世界言語の対応、connpassで予め取得したjsonを使ってのチェック。parametarizeとfixtureの便利さを体験する 1. 挨拶botの各国の回答: parametarize 2. connpassのオンライン回答ができているかの: fixtureでjsonを共通でロードしてテスト 3. (optional):天気botの地域ごとの対応: --- sphinxは構造化されたドキュメントの書き方。最初の使い方をメイン。pythonとの連携としてautodocでできるリファレンスの説明 1. sphinx-quickstartでプロジェクトを作る 2.ある程度できているドキュメントのビルドをして、ファイルを一つ追加して再ビルド 3. autodocを使ってリファレンス生成
non_process
チュートリアルで実際に手を動かしてもらう部分を決める 実装で作業してもらうときは、答えがあることを前提 。天気botは最初から実装がかなり難しいので、触る程度になるかも。 ) ハードコードした挨拶bot botが使う共通インターフェイスを実装して、モジュール化を行う。テストがしやすい環境を整える 挨拶botの世界挨拶の対応 モジュール内に用意したマッピングした挨拶のセット 自作のモジュールを使ってみる connpassbot オンラインイベントの検索の実装 外部のapiを使ってjsonのパースまで。requestsでスクレイピング (説明のみ) 天気botの概要 負荷を掛けないための工夫 pytestで作業する項目は、挨拶botの世界言語の対応、connpassで予め取得したjsonを使ってのチェック。parametarizeとfixtureの便利さを体験する 挨拶botの各国の回答 parametarize connpassのオンライン回答ができているかの fixtureでjsonを共通でロードしてテスト (optional) 天気botの地域ごとの対応 sphinxは構造化されたドキュメントの書き方。最初の使い方をメイン。pythonとの連携としてautodocでできるリファレンスの説明 sphinx quickstartでプロジェクトを作る ある程度できているドキュメントのビルドをして、ファイルを一つ追加して再ビルド autodocを使ってリファレンス生成
0
10,632
13,441,582,116
IssuesEvent
2020-09-08 04:35:54
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
PDF multi-group table publishing fails
bug preprocess priority/medium
## Expected Behavior Each `tgroup` must be rendered correctly. ![image](https://user-images.githubusercontent.com/12742806/92199064-aba73580-eea8-11ea-8981-eefab96cd6bb.png) ## Actual Behavior When the first `tgroup` has `cols` greater than then next one, the publishing fails. However, if you switch the two it publishes fine. ## Possible Solution <!-- Optional suggestions on how to fix the issue, or implement the changes. --> <!-- If you know how to fix the issue, please create a pull request instead. --> ## Steps to Reproduce <!-- Test case, Gist, set of files or steps required to reproduce the issue. --> Sample Topic: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_id"> <title></title> <body> <table frame="all" rowsep="1" colsep="1" id="table_id_1"> <title/> <tgroup cols="2"> <colspec colname="c1" colnum="1" colwidth="1*"/> <colspec colname="c2" colnum="2" colwidth="1*"/> <tbody> <row> <entry>A</entry> <entry>B</entry> </row> </tbody> </tgroup> <tgroup cols="3"> <colspec colname="c1" colnum="1" colwidth="1*"/> <colspec colname="c2" colnum="2" colwidth="1*"/> <colspec colname="c3" colnum="3" colwidth="1*"/> <tbody> <row> <entry>A</entry> <entry>B</entry> <entry>C</entry> </row> </tbody> </tgroup> </table> </body> </topic> ``` [sample.zip](https://github.com/dita-ot/dita-ot/files/5172678/sample.zip) 1. Running it in with pdf format should give you the error. <!-- Create a Gist via <https://gist.github.com/> to upload your test files. --> <!-- Link to the Gist from the issue or attach a .zip archive of your files. --> ## Copy of the error message, log file or stack trace Error: java.lang.IndexOutOfBoundsException: Index: 2, Size: 2 ## Environment <!-- Include relevant details about the environment you experienced this in. --> * DITA-OT version: DITA-OT 3.5.2 * Operating system and version: _(Windows)_ * How did you run DITA-OT? _(`dita` command)_ * Transformation type: _(PDF)_ <!-- Before submitting, check the Preview tab above to verify the XML markup appears correctly and remember you can edit the description later to add information. -->
1.0
PDF multi-group table publishing fails - ## Expected Behavior Each `tgroup` must be rendered correctly. ![image](https://user-images.githubusercontent.com/12742806/92199064-aba73580-eea8-11ea-8981-eefab96cd6bb.png) ## Actual Behavior When the first `tgroup` has `cols` greater than then next one, the publishing fails. However, if you switch the two it publishes fine. ## Possible Solution <!-- Optional suggestions on how to fix the issue, or implement the changes. --> <!-- If you know how to fix the issue, please create a pull request instead. --> ## Steps to Reproduce <!-- Test case, Gist, set of files or steps required to reproduce the issue. --> Sample Topic: ``` <?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE topic PUBLIC "-//OASIS//DTD DITA Topic//EN" "topic.dtd"> <topic id="topic_id"> <title></title> <body> <table frame="all" rowsep="1" colsep="1" id="table_id_1"> <title/> <tgroup cols="2"> <colspec colname="c1" colnum="1" colwidth="1*"/> <colspec colname="c2" colnum="2" colwidth="1*"/> <tbody> <row> <entry>A</entry> <entry>B</entry> </row> </tbody> </tgroup> <tgroup cols="3"> <colspec colname="c1" colnum="1" colwidth="1*"/> <colspec colname="c2" colnum="2" colwidth="1*"/> <colspec colname="c3" colnum="3" colwidth="1*"/> <tbody> <row> <entry>A</entry> <entry>B</entry> <entry>C</entry> </row> </tbody> </tgroup> </table> </body> </topic> ``` [sample.zip](https://github.com/dita-ot/dita-ot/files/5172678/sample.zip) 1. Running it in with pdf format should give you the error. <!-- Create a Gist via <https://gist.github.com/> to upload your test files. --> <!-- Link to the Gist from the issue or attach a .zip archive of your files. --> ## Copy of the error message, log file or stack trace Error: java.lang.IndexOutOfBoundsException: Index: 2, Size: 2 ## Environment <!-- Include relevant details about the environment you experienced this in. --> * DITA-OT version: DITA-OT 3.5.2 * Operating system and version: _(Windows)_ * How did you run DITA-OT? _(`dita` command)_ * Transformation type: _(PDF)_ <!-- Before submitting, check the Preview tab above to verify the XML markup appears correctly and remember you can edit the description later to add information. -->
process
pdf multi group table publishing fails expected behavior each tgroup must be rendered correctly actual behavior when the first tgroup has cols greater than then next one the publishing fails however if you switch the two it publishes fine possible solution steps to reproduce sample topic a b a b c running it in with pdf format should give you the error copy of the error message log file or stack trace error java lang indexoutofboundsexception index size environment dita ot version dita ot operating system and version windows how did you run dita ot dita command transformation type pdf before submitting check the preview tab above to verify the xml markup appears correctly and remember you can edit the description later to add information
1
762,309
26,714,435,688
IssuesEvent
2023-01-28 09:59:28
azerothcore/azerothcore-wotlk
https://api.github.com/repos/azerothcore/azerothcore-wotlk
closed
`Master of Elements` is ignored when having `Burnout`
Confirmed Priority-High Class - Mage
### Current Behaviour The [Master of Elements](https://wotlkdb.com/?spell=29076) talent is ignored when the player also have the talent [Burnout](https://wotlkdb.com/?spell=44472). ![image](https://user-images.githubusercontent.com/22718174/209576425-90159774-e178-4a84-a927-0b65b358de6a.png) ### Expected Blizzlike Behaviour If I have exactly 1000 mana while having both talents maxxed and I crit with a spell that costs 100 mana, I should end up with 925 mana (30% of the base mana cost has been regained from `Master of Elements` and a 5% mana loss from `Burnout`). ### Source ![image](https://user-images.githubusercontent.com/22718174/209576554-99ab303d-700c-418b-8b4f-a2f3ac69039c.png) ### Steps to reproduce the problem https://user-images.githubusercontent.com/22718174/209577657-c31f519c-6706-4eef-8293-ca336b81f01f.mp4 **Setup for guarantee critical hit.** `.cheat cooldown` Get 10 stacks of [Combustion](https://wotlkdb.com/?spell=11129) for 100% crit. **Experimentations** I have exactly 14987 max mana. `Fireball` costs 503 mana. Critting with 3/3 `Master of Elements` and 0/5 `Burnout` should refund ~150.9 mana. Knowing the following, I did properly end up at 14634 mana (150 mana has been refunded). I took 1/5 `Burnout` and waited for my mana to be at max (drank a potion in the video). If I crit with `Fireball`, my mana should lower by 1% of the mana cost from my `Fireball`: ~5.03. Having both 3/3 `Master of Elements` and 1/5 `Burnout` should set me at ~14629.87 mana. Instead, I've ended at 14479 mana. That is a loss of 508 mana (-503 mana from `Fireball` and -5 from `Burnout`. I did not regain mana from `Master of Elements`). I tried with 5/5 `Burnout`. The results are similar. If I crit with `Fireball`, my mana should lower by 5% of the mana cost from my `Fireball`: ~25.15. Having both 3/3 `Master of Elements` and 5/5 `Burnout` should set me at ~14609.75 mana. Instead, I've ended at 14459 mana. That is a loss of 528 mana (-503 mana from `Fireball` and -25 from `Burnout`. I did not regain mana from `Master of Elements`). ### Extra Notes I found it strange that whenever I critted, I was loosing mana. When I crit with both talents, I did notice that my current mana did change then the next frame, changed again (could not see how much it was). Could be related to the QAston bug report: https://github.com/azerothcore/azerothcore-wotlk/issues/13282 ### AC rev. hash/commit AzerothCore rev. 721711d70f31+ 2022-12-02 08:59:47 +0000 (master branch) (Win64, RelWithDebInfo, Static) Using SSL version: OpenSSL 1.1.1m 14 Dec 2021 (library: OpenSSL 1.1.1m 14 Dec 2021) Using Boost version: 1.78.0 Using MySQL version: 80027 Using CMake version: 3.22.1 Compiled on: Windows 10.0.19044 Worldserver listening connections on port 8085 Realmlist (Realm Id: 1) configured in port 8085 VMAPs status: Enabled. LineOfSight: 1, getHeight: 1, indoorCheck: 1 MMAPs status: Enabled maps directory located in ./maps. Total size: 291014951 bytes vmaps directory located in ./vmaps. Total size: 658130721 bytes mmaps directory located in ./mmaps. Total size: 2320900676 bytes Using enUS DBC Locale as default. All available DBC locales: enUS Using World DB: ACDB 335.8-dev LoginDatabase queue size: 0 CharacterDatabase queue size: 0 WorldDatabase queue size: 0 No modules enabled ### Operating system Windows 10 x64 ### Custom changes or Modules _No response_
1.0
`Master of Elements` is ignored when having `Burnout` - ### Current Behaviour The [Master of Elements](https://wotlkdb.com/?spell=29076) talent is ignored when the player also have the talent [Burnout](https://wotlkdb.com/?spell=44472). ![image](https://user-images.githubusercontent.com/22718174/209576425-90159774-e178-4a84-a927-0b65b358de6a.png) ### Expected Blizzlike Behaviour If I have exactly 1000 mana while having both talents maxxed and I crit with a spell that costs 100 mana, I should end up with 925 mana (30% of the base mana cost has been regained from `Master of Elements` and a 5% mana loss from `Burnout`). ### Source ![image](https://user-images.githubusercontent.com/22718174/209576554-99ab303d-700c-418b-8b4f-a2f3ac69039c.png) ### Steps to reproduce the problem https://user-images.githubusercontent.com/22718174/209577657-c31f519c-6706-4eef-8293-ca336b81f01f.mp4 **Setup for guarantee critical hit.** `.cheat cooldown` Get 10 stacks of [Combustion](https://wotlkdb.com/?spell=11129) for 100% crit. **Experimentations** I have exactly 14987 max mana. `Fireball` costs 503 mana. Critting with 3/3 `Master of Elements` and 0/5 `Burnout` should refund ~150.9 mana. Knowing the following, I did properly end up at 14634 mana (150 mana has been refunded). I took 1/5 `Burnout` and waited for my mana to be at max (drank a potion in the video). If I crit with `Fireball`, my mana should lower by 1% of the mana cost from my `Fireball`: ~5.03. Having both 3/3 `Master of Elements` and 1/5 `Burnout` should set me at ~14629.87 mana. Instead, I've ended at 14479 mana. That is a loss of 508 mana (-503 mana from `Fireball` and -5 from `Burnout`. I did not regain mana from `Master of Elements`). I tried with 5/5 `Burnout`. The results are similar. If I crit with `Fireball`, my mana should lower by 5% of the mana cost from my `Fireball`: ~25.15. Having both 3/3 `Master of Elements` and 5/5 `Burnout` should set me at ~14609.75 mana. Instead, I've ended at 14459 mana. That is a loss of 528 mana (-503 mana from `Fireball` and -25 from `Burnout`. I did not regain mana from `Master of Elements`). ### Extra Notes I found it strange that whenever I critted, I was loosing mana. When I crit with both talents, I did notice that my current mana did change then the next frame, changed again (could not see how much it was). Could be related to the QAston bug report: https://github.com/azerothcore/azerothcore-wotlk/issues/13282 ### AC rev. hash/commit AzerothCore rev. 721711d70f31+ 2022-12-02 08:59:47 +0000 (master branch) (Win64, RelWithDebInfo, Static) Using SSL version: OpenSSL 1.1.1m 14 Dec 2021 (library: OpenSSL 1.1.1m 14 Dec 2021) Using Boost version: 1.78.0 Using MySQL version: 80027 Using CMake version: 3.22.1 Compiled on: Windows 10.0.19044 Worldserver listening connections on port 8085 Realmlist (Realm Id: 1) configured in port 8085 VMAPs status: Enabled. LineOfSight: 1, getHeight: 1, indoorCheck: 1 MMAPs status: Enabled maps directory located in ./maps. Total size: 291014951 bytes vmaps directory located in ./vmaps. Total size: 658130721 bytes mmaps directory located in ./mmaps. Total size: 2320900676 bytes Using enUS DBC Locale as default. All available DBC locales: enUS Using World DB: ACDB 335.8-dev LoginDatabase queue size: 0 CharacterDatabase queue size: 0 WorldDatabase queue size: 0 No modules enabled ### Operating system Windows 10 x64 ### Custom changes or Modules _No response_
non_process
master of elements is ignored when having burnout current behaviour the talent is ignored when the player also have the talent expected blizzlike behaviour if i have exactly mana while having both talents maxxed and i crit with a spell that costs mana i should end up with mana of the base mana cost has been regained from master of elements and a mana loss from burnout source steps to reproduce the problem setup for guarantee critical hit cheat cooldown get stacks of for crit experimentations i have exactly max mana fireball costs mana critting with master of elements and burnout should refund mana knowing the following i did properly end up at mana mana has been refunded i took burnout and waited for my mana to be at max drank a potion in the video if i crit with fireball my mana should lower by of the mana cost from my fireball having both master of elements and burnout should set me at mana instead i ve ended at mana that is a loss of mana mana from fireball and from burnout i did not regain mana from master of elements i tried with burnout the results are similar if i crit with fireball my mana should lower by of the mana cost from my fireball having both master of elements and burnout should set me at mana instead i ve ended at mana that is a loss of mana mana from fireball and from burnout i did not regain mana from master of elements extra notes i found it strange that whenever i critted i was loosing mana when i crit with both talents i did notice that my current mana did change then the next frame changed again could not see how much it was could be related to the qaston bug report ac rev hash commit azerothcore rev master branch relwithdebinfo static using ssl version openssl dec library openssl dec using boost version using mysql version using cmake version compiled on windows worldserver listening connections on port realmlist realm id configured in port vmaps status enabled lineofsight getheight indoorcheck mmaps status enabled maps directory located in maps total size bytes vmaps directory located in vmaps total size bytes mmaps directory located in mmaps total size bytes using enus dbc locale as default all available dbc locales enus using world db acdb dev logindatabase queue size characterdatabase queue size worlddatabase queue size no modules enabled operating system windows custom changes or modules no response
0
41,776
16,903,963,901
IssuesEvent
2021-06-24 03:38:16
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
azurerm_dashboard causes HTTP 400 when calling portal.DashboardsClient#CreateOrUpdate
bug service/dashboards
<!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> Terraform v0.12.29 + provider.azurerm v2.47.0 Your version of Terraform is out of date! The latest version is 0.14.6. You can update by downloading from https://www.terraform.io/downloads.html ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * `azurerm_dashboard ` ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_dashboard" "my-board" { name = "my-cool-dashboard" resource_group_name = azurerm_resource_group.my-group.name location = azurerm_resource_group.my-group.location tags = { source = "terraform" } dashboard_properties = templatefile("dash.tpl", { md_content = "__My data2__" }) } ``` dash.tpl can be found [here](https://github.com/booyaa/tf_azurerm_dashboard_bug_10616/blob/main/dash.tpl) ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html). ---> https://gist.github.com/booyaa/65c9222909236663bb282b80eb5da326 ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behaviour <!--- What should have happened? ---> Dashboard should been create ### Actual Behaviour <!--- What actually happened? ---> ```sh terraform apply "dev.plan" azurerm_dashboard.my-board: Creating... Error: creating/updating Dashboard "my-cool-dashboard" (Resource Group "rg-dev"): portal.DashboardsClient#CreateOrUpdate: Failure responding to request: StatusCode=400 -- Original Error: autorest/azure: error response cannot be parsed: "{\"error\":{\"code\":37,\"message\":\"The request content was invalid and could not be deserialized: 'Required property 'lenses' not found in JSON. Path 'properties', line 1, position 39.'.\"}}" error: json: cannot unmarshal number into Go struct field RequestError.error of type string on main.tf line 30, in resource "azurerm_dashboard" "my-board": 30: resource "azurerm_dashboard" "my-board" { ``` ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` An example has been provided: [link](https://github.com/booyaa/tf_azurerm_dashboard_bug_10616) ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in a Azure China/Germany/Government? ---> The example repo is pretty much based on the example given in the documentation: [link](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/dashboard) ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Such as vendor documentation? ---> * #0000
1.0
azurerm_dashboard causes HTTP 400 when calling portal.DashboardsClient#CreateOrUpdate - <!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> Terraform v0.12.29 + provider.azurerm v2.47.0 Your version of Terraform is out of date! The latest version is 0.14.6. You can update by downloading from https://www.terraform.io/downloads.html ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * `azurerm_dashboard ` ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "azurerm_dashboard" "my-board" { name = "my-cool-dashboard" resource_group_name = azurerm_resource_group.my-group.name location = azurerm_resource_group.my-group.location tags = { source = "terraform" } dashboard_properties = templatefile("dash.tpl", { md_content = "__My data2__" }) } ``` dash.tpl can be found [here](https://github.com/booyaa/tf_azurerm_dashboard_bug_10616/blob/main/dash.tpl) ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html). ---> https://gist.github.com/booyaa/65c9222909236663bb282b80eb5da326 ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behaviour <!--- What should have happened? ---> Dashboard should been create ### Actual Behaviour <!--- What actually happened? ---> ```sh terraform apply "dev.plan" azurerm_dashboard.my-board: Creating... Error: creating/updating Dashboard "my-cool-dashboard" (Resource Group "rg-dev"): portal.DashboardsClient#CreateOrUpdate: Failure responding to request: StatusCode=400 -- Original Error: autorest/azure: error response cannot be parsed: "{\"error\":{\"code\":37,\"message\":\"The request content was invalid and could not be deserialized: 'Required property 'lenses' not found in JSON. Path 'properties', line 1, position 39.'.\"}}" error: json: cannot unmarshal number into Go struct field RequestError.error of type string on main.tf line 30, in resource "azurerm_dashboard" "my-board": 30: resource "azurerm_dashboard" "my-board" { ``` ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` An example has been provided: [link](https://github.com/booyaa/tf_azurerm_dashboard_bug_10616) ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in a Azure China/Germany/Government? ---> The example repo is pretty much based on the example given in the documentation: [link](https://registry.terraform.io/providers/hashicorp/azurerm/latest/docs/resources/dashboard) ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Such as vendor documentation? ---> * #0000
non_process
azurerm dashboard causes http when calling portal dashboardsclient createorupdate please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version terraform provider azurerm your version of terraform is out of date the latest version is you can update by downloading from affected resource s azurerm dashboard terraform configuration files hcl resource azurerm dashboard my board name my cool dashboard resource group name azurerm resource group my group name location azurerm resource group my group location tags source terraform dashboard properties templatefile dash tpl md content my dash tpl can be found debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the panic output expected behaviour dashboard should been create actual behaviour sh terraform apply dev plan azurerm dashboard my board creating error creating updating dashboard my cool dashboard resource group rg dev portal dashboardsclient createorupdate failure responding to request statuscode original error autorest azure error response cannot be parsed error code message the request content was invalid and could not be deserialized required property lenses not found in json path properties line position error json cannot unmarshal number into go struct field requesterror error of type string on main tf line in resource azurerm dashboard my board resource azurerm dashboard my board steps to reproduce terraform apply an example has been provided important factoids the example repo is pretty much based on the example given in the documentation references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here such as vendor documentation
0
8,215
11,405,557,393
IssuesEvent
2020-01-31 12:23:54
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
opened
[Introspection] Investigate and improve introspection speed
kind/improvement process/candidate topic: introspection
Right now Introspection sometimes fails for big schemas (See https://github.com/prisma/prisma2/issues/1493). We should investigate and improve the behavior of Introspection by e.g. parallelizing queries or similar.
1.0
[Introspection] Investigate and improve introspection speed - Right now Introspection sometimes fails for big schemas (See https://github.com/prisma/prisma2/issues/1493). We should investigate and improve the behavior of Introspection by e.g. parallelizing queries or similar.
process
investigate and improve introspection speed right now introspection sometimes fails for big schemas see we should investigate and improve the behavior of introspection by e g parallelizing queries or similar
1
19,183
25,303,570,552
IssuesEvent
2022-11-17 12:35:44
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
opened
Change term - footprintRadiusSpatialFit
Term - change Process - ready for public comment Class - Location non-normative
## Term change * Submitter: John Wieczorek * Efficacy Justification (why is this change necessary?): Clarification * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Current Term definition: https://dwc.tdwg.org/list/#dwc_footprintSpatialFit "The ratio of the area of the footprint (footprintWKT) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given footprint does not completely contain the original representation. The footprintSpatialFit is undefined (and should be left empty) if the original representation is a point without uncertainty and the given georeference is not that same point (without uncertainty). If both the original and the given georeference are the same point, the footprintSpatialFit is 1." Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): footprintSpatialFit * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Location * Definition of the term (normative): "The ratio of the area of the footprint (footprintWKT) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given footprint does not completely contain the original representation. The footprintSpatialFit is undefined (and should be left empty) if the original representation is **any geometry without area (e.g., a point or polyline) and**~~a point~~ without uncertainty and the given georeference is not that same **geometry**~~point~~ (without uncertainty). If both the original and the given georeference are the same point, the footprintSpatialFit is 1." * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): * Refines (identifier of the broader term this term refines; normative): * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/footprintRadiusSpatialFit-2020-08-20** * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/Gathering/FootprintSpatialFit This definition change only clarifies what to do in non-area representations other than a point. It does not change the meaning or the usage.
1.0
Change term - footprintRadiusSpatialFit - ## Term change * Submitter: John Wieczorek * Efficacy Justification (why is this change necessary?): Clarification * Demand Justification (if the change is semantic in nature, name at least two organizations that independently need this term): * Stability Justification (what concerns are there that this might affect existing implementations?): None * Implications for dwciri: namespace (does this change affect a dwciri term version)?: None Current Term definition: https://dwc.tdwg.org/list/#dwc_footprintSpatialFit "The ratio of the area of the footprint (footprintWKT) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given footprint does not completely contain the original representation. The footprintSpatialFit is undefined (and should be left empty) if the original representation is a point without uncertainty and the given georeference is not that same point (without uncertainty). If both the original and the given georeference are the same point, the footprintSpatialFit is 1." Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Term name (in lowerCamelCase for properties, UpperCamelCase for classes): footprintSpatialFit * Organized in Class (e.g., Occurrence, Event, Location, Taxon): Location * Definition of the term (normative): "The ratio of the area of the footprint (footprintWKT) to the area of the true (original, or most specific) spatial representation of the Location. Legal values are 0, greater than or equal to 1, or undefined. A value of 1 is an exact match or 100% overlap. A value of 0 should be used if the given footprint does not completely contain the original representation. The footprintSpatialFit is undefined (and should be left empty) if the original representation is **any geometry without area (e.g., a point or polyline) and**~~a point~~ without uncertainty and the given georeference is not that same **geometry**~~point~~ (without uncertainty). If both the original and the given georeference are the same point, the footprintSpatialFit is 1." * Usage comments (recommendations regarding content, etc., not normative): * Examples (not normative): * Refines (identifier of the broader term this term refines; normative): * Replaces (identifier of the existing term that would be deprecated and replaced by this term; normative): **http://rs.tdwg.org/dwc/terms/version/footprintRadiusSpatialFit-2020-08-20** * ABCD 2.06 (XPATH of the equivalent term in ABCD or EFG; not normative): DataSets/DataSet/Units/Unit/Gathering/FootprintSpatialFit This definition change only clarifies what to do in non-area representations other than a point. It does not change the meaning or the usage.
process
change term footprintradiusspatialfit term change submitter john wieczorek efficacy justification why is this change necessary clarification demand justification if the change is semantic in nature name at least two organizations that independently need this term stability justification what concerns are there that this might affect existing implementations none implications for dwciri namespace does this change affect a dwciri term version none current term definition the ratio of the area of the footprint footprintwkt to the area of the true original or most specific spatial representation of the location legal values are greater than or equal to or undefined a value of is an exact match or overlap a value of should be used if the given footprint does not completely contain the original representation the footprintspatialfit is undefined and should be left empty if the original representation is a point without uncertainty and the given georeference is not that same point without uncertainty if both the original and the given georeference are the same point the footprintspatialfit is proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough term name in lowercamelcase for properties uppercamelcase for classes footprintspatialfit organized in class e g occurrence event location taxon location definition of the term normative the ratio of the area of the footprint footprintwkt to the area of the true original or most specific spatial representation of the location legal values are greater than or equal to or undefined a value of is an exact match or overlap a value of should be used if the given footprint does not completely contain the original representation the footprintspatialfit is undefined and should be left empty if the original representation is any geometry without area e g a point or polyline and a point without uncertainty and the given georeference is not that same geometry point without uncertainty if both the original and the given georeference are the same point the footprintspatialfit is usage comments recommendations regarding content etc not normative examples not normative refines identifier of the broader term this term refines normative replaces identifier of the existing term that would be deprecated and replaced by this term normative abcd xpath of the equivalent term in abcd or efg not normative datasets dataset units unit gathering footprintspatialfit this definition change only clarifies what to do in non area representations other than a point it does not change the meaning or the usage
1
12,080
14,740,034,547
IssuesEvent
2021-01-07 08:24:00
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
SA Hosted - SA Billing - Late Fee Account List
anc-process anp-important ant-bug has attachment
In GitLab by @kdjstudios on Oct 3, 2018, 10:54 [SA_Hosted.xlsx](/uploads/3c6eed7ad00ab9874cc4a1321432968e/SA_Hosted.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-27286/conversation
1.0
SA Hosted - SA Billing - Late Fee Account List - In GitLab by @kdjstudios on Oct 3, 2018, 10:54 [SA_Hosted.xlsx](/uploads/3c6eed7ad00ab9874cc4a1321432968e/SA_Hosted.xlsx) HD: http://www.servicedesk.answernet.com/profiles/ticket/2018-10-03-27286/conversation
process
sa hosted sa billing late fee account list in gitlab by kdjstudios on oct uploads sa hosted xlsx hd
1
18,167
24,207,760,676
IssuesEvent
2022-09-25 13:28:43
sebastianbergmann/phpunit
https://api.github.com/repos/sebastianbergmann/phpunit
closed
À tout de suite: fast self-test suite
feature/test-runner feature/data-provider feature/test-dependencies feature/process-isolation type/tests
As a _developer of PHPUnit_ I want _a fast integration test of all basic use cases_ so that _experimenting with the new logging is easier and more fun_ ### Inspiration ### Related to the [redesign of the result event and logging system](#3390) and the updates for [test execution order logging](#3380) and [TestDox](#3439). Working on the test execution ordering I wished I had a quick ~test suite~ collection of tests that shows me the general state of PHPUnit with a focus on the developer experience. The current test collection is too unwieldy to run on my old laptop continuously. ### What functionality to test? ### - are all test result status and messages presented to the user; example: #3436 - do exceptions and errors bubble to the surface; example: #3364 - running order of test collection within current architecture; #3147 and all its related issues like #3093, #3156 - buffered logging of async test events to decouple execution order and result printer; example: #3380 - output format of the loggers included with the basic package ### Implementation ### - create a small test suite with the basics as a part of the main test collection, but which can also be reused for individual end-to-end tests - reuse existing tests where possible - review existing end-to-end tests: remove duplicates, obsolete regression tests - make it easier to test loggers via unit tests ### List of scenarios ### #### Basics #### - [x] all test result types with standard messages - [x] all test result types with verbose messages #### Fixtures #### - [x] TestCase class with a failing `setUpBeforeClass()` - [x] TestCase class with a failing `tearDownAfterClass()` - [x] TestCase class with a `setUp()` that fails always - [ ] TestCase class with a `tearDown()` that fails #### Dataproviders #### - [ ] Test with `@dataprovider` and a provider that returns a simple array - [ ] Test with `@dataprovider` and a provider that returns named rows - [ ] Test with an incompatible `@dataprovider` - [ ] Test with a `@dataprovider` that returns no data - [ ] Test with a Generator `@dataprovider` - [ ] Test with a Generator `@dataprovider` with a duplicate key - [ ] Test with a malformed `@dataprovider` row - [ ] skipped dataprovider; see `DataProviderSkippedTest` #### Test execution order #### - [ ] Test that `@depends` on a previous test (original: no output) - [ ] Test that `@depends` on a 'later' test (original: skipped, when no-depends) - [ ] Test that `@depends` on a Test in another class - [ ] unbuffered logging: reordering affects both run order and output order - [ ] buffered logging: reordering affects only run order and result, but not output order - [ ] no difference between buffered and unbuffered logging when in original order - [ ] output random configuration to make random runs are reproducible, see #3515 #### Test collection structure #### - [ ] TestSuite with only nested TestSuites - [ ] TestSuite with only TestCases - [ ] TestSuite with a mix of suites and tests - [ ] Test class without any tests - [ ] PHPT end-to-end - [ ] skipped PHPT end-to-end via a triggered `--SKIPIF--` section - [ ] Test skipped by `@requires` #### Isolation #### - [ ] Test class with `@runClassInSeparateProcess`, all events reach user - [ ] Test with `@runInSeparateProcess`, all events reach user #### Included printers and logging formats #### - [x] `--debug` - [ ] the classic dot-o-matic printer - [ ] TestDox (with all features) - [ ] TeamCity - [ ] JUnit - [ ] JUnit v5 see #3162 - [ ] the formats above with _and_ without colorization - [ ] the formats above with both quiet _and_ verbose
1.0
À tout de suite: fast self-test suite - As a _developer of PHPUnit_ I want _a fast integration test of all basic use cases_ so that _experimenting with the new logging is easier and more fun_ ### Inspiration ### Related to the [redesign of the result event and logging system](#3390) and the updates for [test execution order logging](#3380) and [TestDox](#3439). Working on the test execution ordering I wished I had a quick ~test suite~ collection of tests that shows me the general state of PHPUnit with a focus on the developer experience. The current test collection is too unwieldy to run on my old laptop continuously. ### What functionality to test? ### - are all test result status and messages presented to the user; example: #3436 - do exceptions and errors bubble to the surface; example: #3364 - running order of test collection within current architecture; #3147 and all its related issues like #3093, #3156 - buffered logging of async test events to decouple execution order and result printer; example: #3380 - output format of the loggers included with the basic package ### Implementation ### - create a small test suite with the basics as a part of the main test collection, but which can also be reused for individual end-to-end tests - reuse existing tests where possible - review existing end-to-end tests: remove duplicates, obsolete regression tests - make it easier to test loggers via unit tests ### List of scenarios ### #### Basics #### - [x] all test result types with standard messages - [x] all test result types with verbose messages #### Fixtures #### - [x] TestCase class with a failing `setUpBeforeClass()` - [x] TestCase class with a failing `tearDownAfterClass()` - [x] TestCase class with a `setUp()` that fails always - [ ] TestCase class with a `tearDown()` that fails #### Dataproviders #### - [ ] Test with `@dataprovider` and a provider that returns a simple array - [ ] Test with `@dataprovider` and a provider that returns named rows - [ ] Test with an incompatible `@dataprovider` - [ ] Test with a `@dataprovider` that returns no data - [ ] Test with a Generator `@dataprovider` - [ ] Test with a Generator `@dataprovider` with a duplicate key - [ ] Test with a malformed `@dataprovider` row - [ ] skipped dataprovider; see `DataProviderSkippedTest` #### Test execution order #### - [ ] Test that `@depends` on a previous test (original: no output) - [ ] Test that `@depends` on a 'later' test (original: skipped, when no-depends) - [ ] Test that `@depends` on a Test in another class - [ ] unbuffered logging: reordering affects both run order and output order - [ ] buffered logging: reordering affects only run order and result, but not output order - [ ] no difference between buffered and unbuffered logging when in original order - [ ] output random configuration to make random runs are reproducible, see #3515 #### Test collection structure #### - [ ] TestSuite with only nested TestSuites - [ ] TestSuite with only TestCases - [ ] TestSuite with a mix of suites and tests - [ ] Test class without any tests - [ ] PHPT end-to-end - [ ] skipped PHPT end-to-end via a triggered `--SKIPIF--` section - [ ] Test skipped by `@requires` #### Isolation #### - [ ] Test class with `@runClassInSeparateProcess`, all events reach user - [ ] Test with `@runInSeparateProcess`, all events reach user #### Included printers and logging formats #### - [x] `--debug` - [ ] the classic dot-o-matic printer - [ ] TestDox (with all features) - [ ] TeamCity - [ ] JUnit - [ ] JUnit v5 see #3162 - [ ] the formats above with _and_ without colorization - [ ] the formats above with both quiet _and_ verbose
process
à tout de suite fast self test suite as a developer of phpunit i want a fast integration test of all basic use cases so that experimenting with the new logging is easier and more fun inspiration related to the and the updates for and working on the test execution ordering i wished i had a quick test suite collection of tests that shows me the general state of phpunit with a focus on the developer experience the current test collection is too unwieldy to run on my old laptop continuously what functionality to test are all test result status and messages presented to the user example do exceptions and errors bubble to the surface example running order of test collection within current architecture and all its related issues like buffered logging of async test events to decouple execution order and result printer example output format of the loggers included with the basic package implementation create a small test suite with the basics as a part of the main test collection but which can also be reused for individual end to end tests reuse existing tests where possible review existing end to end tests remove duplicates obsolete regression tests make it easier to test loggers via unit tests list of scenarios basics all test result types with standard messages all test result types with verbose messages fixtures testcase class with a failing setupbeforeclass testcase class with a failing teardownafterclass testcase class with a setup that fails always testcase class with a teardown that fails dataproviders test with dataprovider and a provider that returns a simple array test with dataprovider and a provider that returns named rows test with an incompatible dataprovider test with a dataprovider that returns no data test with a generator dataprovider test with a generator dataprovider with a duplicate key test with a malformed dataprovider row skipped dataprovider see dataproviderskippedtest test execution order test that depends on a previous test original no output test that depends on a later test original skipped when no depends test that depends on a test in another class unbuffered logging reordering affects both run order and output order buffered logging reordering affects only run order and result but not output order no difference between buffered and unbuffered logging when in original order output random configuration to make random runs are reproducible see test collection structure testsuite with only nested testsuites testsuite with only testcases testsuite with a mix of suites and tests test class without any tests phpt end to end skipped phpt end to end via a triggered skipif section test skipped by requires isolation test class with runclassinseparateprocess all events reach user test with runinseparateprocess all events reach user included printers and logging formats debug the classic dot o matic printer testdox with all features teamcity junit junit see the formats above with and without colorization the formats above with both quiet and verbose
1
17,550
23,362,692,545
IssuesEvent
2022-08-10 13:04:31
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Span count sampler throws null pointer exception
bug priority:p1 processor/tailsampling
**Describe the bug** Span count sampler is throwing null pointer exception when used alongside with an AND policy in the tailsampling processor **Steps to reproduce** Configure tailsampling processor to utilize an AND policy in conjunction with Span count sampler. **What did you expect to see?** No errors. **What did you see instead?** Null pointer exception. **What version did you use?** v0.54 **Additional context** I'm creating a PR to fix the issue.
1.0
Span count sampler throws null pointer exception - **Describe the bug** Span count sampler is throwing null pointer exception when used alongside with an AND policy in the tailsampling processor **Steps to reproduce** Configure tailsampling processor to utilize an AND policy in conjunction with Span count sampler. **What did you expect to see?** No errors. **What did you see instead?** Null pointer exception. **What version did you use?** v0.54 **Additional context** I'm creating a PR to fix the issue.
process
span count sampler throws null pointer exception describe the bug span count sampler is throwing null pointer exception when used alongside with an and policy in the tailsampling processor steps to reproduce configure tailsampling processor to utilize an and policy in conjunction with span count sampler what did you expect to see no errors what did you see instead null pointer exception what version did you use additional context i m creating a pr to fix the issue
1
2,316
2,607,871,627
IssuesEvent
2015-02-26 00:00:47
chrsmithdemos/google-styleguide
https://api.github.com/repos/chrsmithdemos/google-styleguide
opened
Forced return value in cpplint
auto-migrated Priority-Medium Type-Defect
``` The cpplint returns non-zero when there is at least one message. If the cpplint is started from a make file, the non-zero return value will stop. Proposal is to add "--return=value" option that forces the script return the given value even when there is messages (see PC-lint -zero). Attached batch adds the option to the cpplint and it has been tested with cmake + Visual Studio 2013. ``` ----- Original issue reported on code.google.com by `ksole...@gmail.com` on 1 Oct 2014 at 7:15 Attachments: * [cpplint.py.--return.patch](https://storage.googleapis.com/google-code-attachments/google-styleguide/issue-28/comment-0/cpplint.py.--return.patch)
1.0
Forced return value in cpplint - ``` The cpplint returns non-zero when there is at least one message. If the cpplint is started from a make file, the non-zero return value will stop. Proposal is to add "--return=value" option that forces the script return the given value even when there is messages (see PC-lint -zero). Attached batch adds the option to the cpplint and it has been tested with cmake + Visual Studio 2013. ``` ----- Original issue reported on code.google.com by `ksole...@gmail.com` on 1 Oct 2014 at 7:15 Attachments: * [cpplint.py.--return.patch](https://storage.googleapis.com/google-code-attachments/google-styleguide/issue-28/comment-0/cpplint.py.--return.patch)
non_process
forced return value in cpplint the cpplint returns non zero when there is at least one message if the cpplint is started from a make file the non zero return value will stop proposal is to add return value option that forces the script return the given value even when there is messages see pc lint zero attached batch adds the option to the cpplint and it has been tested with cmake visual studio original issue reported on code google com by ksole gmail com on oct at attachments
0
218,549
16,762,775,013
IssuesEvent
2021-06-14 03:08:48
ARMmaster17/Captain
https://api.github.com/repos/ARMmaster17/Captain
closed
Option to specify Ansible playbook name in formation data
component:ATC component:Radar difficulty:medium documentation enhancement
## General Information - This is a: feature request - Captain version: `main` - OS: Debian 10 x64 LXC - Proxmox version: `v6.4-8` ## Issue Report Currently when a plane is created, only `test.yml` is run. This mean that all new planes will have the same provisioned state. For now, the relative playbook path should be specified at the formation level for now. In the future, there could be some kind of nested playbook structure where a playbook could be specified for each level of the organization of the Captain stack. ## Related issues Related to #165
1.0
Option to specify Ansible playbook name in formation data - ## General Information - This is a: feature request - Captain version: `main` - OS: Debian 10 x64 LXC - Proxmox version: `v6.4-8` ## Issue Report Currently when a plane is created, only `test.yml` is run. This mean that all new planes will have the same provisioned state. For now, the relative playbook path should be specified at the formation level for now. In the future, there could be some kind of nested playbook structure where a playbook could be specified for each level of the organization of the Captain stack. ## Related issues Related to #165
non_process
option to specify ansible playbook name in formation data general information this is a feature request captain version main os debian lxc proxmox version issue report currently when a plane is created only test yml is run this mean that all new planes will have the same provisioned state for now the relative playbook path should be specified at the formation level for now in the future there could be some kind of nested playbook structure where a playbook could be specified for each level of the organization of the captain stack related issues related to
0
20,713
27,409,982,549
IssuesEvent
2023-03-01 09:53:06
Deltares/Ribasim.jl
https://api.github.com/repos/Deltares/Ribasim.jl
opened
Determine areal shortage(supply)/excess(drainage) condition in timestep
physical process
Control actions mya differen when you are in a shortage/supply state of an excess/discharge state. A method is needed to determine on a subnetwork whether the system is for this subnetwork in a supply or drainage situation, such that proper controll settings can be triggered
1.0
Determine areal shortage(supply)/excess(drainage) condition in timestep - Control actions mya differen when you are in a shortage/supply state of an excess/discharge state. A method is needed to determine on a subnetwork whether the system is for this subnetwork in a supply or drainage situation, such that proper controll settings can be triggered
process
determine areal shortage supply excess drainage condition in timestep control actions mya differen when you are in a shortage supply state of an excess discharge state a method is needed to determine on a subnetwork whether the system is for this subnetwork in a supply or drainage situation such that proper controll settings can be triggered
1
18
2,496,245,328
IssuesEvent
2015-01-06 18:06:16
vivo-isf/vivo-isf-ontology
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
closed
parturition
biological_process imported
_From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on December 12, 2012 17:26:53_ GO:0007567 parent: \<a href="http://purl.obolibrary.org/obo/GO_0022414" rel="nofollow">http://purl.obolibrary.org/obo/GO_0022414</a>&#13; reproductive process&#13; _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=176_
1.0
parturition - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on December 12, 2012 17:26:53_ GO:0007567 parent: \<a href="http://purl.obolibrary.org/obo/GO_0022414" rel="nofollow">http://purl.obolibrary.org/obo/GO_0022414</a>&#13; reproductive process&#13; _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=176_
process
parturition from on december go parent reproductive process original issue
1
342,455
10,317,403,706
IssuesEvent
2019-08-30 12:38:43
geosolutions-it/StatutoTerritorioRT
https://api.github.com/repos/geosolutions-it/StatutoTerritorioRT
opened
L'azione "Emissione Provvedimento di Verifica" non viene chiusa dopo essere stata eseguita
Avvio Priority: High bug
Tipo di Piano testato: - con verifica assoggettamento VAS - con richiesta C.C. iniziale
1.0
L'azione "Emissione Provvedimento di Verifica" non viene chiusa dopo essere stata eseguita - Tipo di Piano testato: - con verifica assoggettamento VAS - con richiesta C.C. iniziale
non_process
l azione emissione provvedimento di verifica non viene chiusa dopo essere stata eseguita tipo di piano testato con verifica assoggettamento vas con richiesta c c iniziale
0
681,636
23,319,037,999
IssuesEvent
2022-08-08 14:49:39
responsible-ai-collaborative/aiid
https://api.github.com/repos/responsible-ai-collaborative/aiid
closed
Listing of all Incidents Bad References
Type:Bug current backlog Priority:High
It looks like the https://incidentdatabase.ai/summaries/incidents is presenting the references on the incidents collection, and that this is producing a bunch of missing reports. See the screen capture below. Incident 161 doesn't even have a single report properly linked. <img width="1116" alt="Screen Shot 2022-08-02 at 8 42 12 AM" src="https://user-images.githubusercontent.com/64780/182416012-4e181874-fb48-48c1-a5e6-f74c634c1aa5.png"> * `Missing Report 23` * `Missing Report 208` * `Missing Report 233` * `Missing Report 1522` * `Missing Report 1530` * `Missing Report 1605`
1.0
Listing of all Incidents Bad References - It looks like the https://incidentdatabase.ai/summaries/incidents is presenting the references on the incidents collection, and that this is producing a bunch of missing reports. See the screen capture below. Incident 161 doesn't even have a single report properly linked. <img width="1116" alt="Screen Shot 2022-08-02 at 8 42 12 AM" src="https://user-images.githubusercontent.com/64780/182416012-4e181874-fb48-48c1-a5e6-f74c634c1aa5.png"> * `Missing Report 23` * `Missing Report 208` * `Missing Report 233` * `Missing Report 1522` * `Missing Report 1530` * `Missing Report 1605`
non_process
listing of all incidents bad references it looks like the is presenting the references on the incidents collection and that this is producing a bunch of missing reports see the screen capture below incident doesn t even have a single report properly linked img width alt screen shot at am src missing report missing report missing report missing report missing report missing report
0
75,106
20,627,438,336
IssuesEvent
2022-03-08 00:45:45
finos/legend-studio
https://api.github.com/repos/finos/legend-studio
closed
Feature request: Add function explorer panel in Query Builder for user defined functions
Type: Feature Request Studio Core Team Component: Query Builder
### Similar requests - [X] I have searched and found no existing similar requests ### How are you using Studio? Legend Studio ### What problems are you trying to solve? To enhance our user experience in using user-defined functions in the query builder. Adding a function explorer panel would give users ability to explorer functions defined in the project and drag and drop into fetch structure panel. ### Implementation Ideas/Steps - [x] Enable a new panel for user-defined functions - [x] Two modes of displaying by package or flatten out by function name - [x] `CONSIDER` Leverage the `doc.doc` tag value so producers of function can add documentation for the function and consumers can read tag value and learn what function does - [x] For functions whom return type is primitive users can DND directly into fetch structure. - [x] Drag and drop into derivation panel - [ ] `CONSIDER` include on the fly `calculated attributes` where users can reuse defined attributes within query builder: @akphi - I have thought about this and I think we can support it but it doesn't have to persistable, meaning that within one editing session, the user can create their calculated attribute to save time, then DnD it etc. but this will not exist in the next editing session (I have seen this behavior in Google Sheets before with number format) - [x] `CONSIDER` add a dropdown somewhere to enable/disable panels to show as we start to have too many panels (parameters, setup, explorer, functions, calculated attributes, etc.)
1.0
Feature request: Add function explorer panel in Query Builder for user defined functions - ### Similar requests - [X] I have searched and found no existing similar requests ### How are you using Studio? Legend Studio ### What problems are you trying to solve? To enhance our user experience in using user-defined functions in the query builder. Adding a function explorer panel would give users ability to explorer functions defined in the project and drag and drop into fetch structure panel. ### Implementation Ideas/Steps - [x] Enable a new panel for user-defined functions - [x] Two modes of displaying by package or flatten out by function name - [x] `CONSIDER` Leverage the `doc.doc` tag value so producers of function can add documentation for the function and consumers can read tag value and learn what function does - [x] For functions whom return type is primitive users can DND directly into fetch structure. - [x] Drag and drop into derivation panel - [ ] `CONSIDER` include on the fly `calculated attributes` where users can reuse defined attributes within query builder: @akphi - I have thought about this and I think we can support it but it doesn't have to persistable, meaning that within one editing session, the user can create their calculated attribute to save time, then DnD it etc. but this will not exist in the next editing session (I have seen this behavior in Google Sheets before with number format) - [x] `CONSIDER` add a dropdown somewhere to enable/disable panels to show as we start to have too many panels (parameters, setup, explorer, functions, calculated attributes, etc.)
non_process
feature request add function explorer panel in query builder for user defined functions similar requests i have searched and found no existing similar requests how are you using studio legend studio what problems are you trying to solve to enhance our user experience in using user defined functions in the query builder adding a function explorer panel would give users ability to explorer functions defined in the project and drag and drop into fetch structure panel implementation ideas steps enable a new panel for user defined functions two modes of displaying by package or flatten out by function name consider leverage the doc doc tag value so producers of function can add documentation for the function and consumers can read tag value and learn what function does for functions whom return type is primitive users can dnd directly into fetch structure drag and drop into derivation panel consider include on the fly calculated attributes where users can reuse defined attributes within query builder akphi i have thought about this and i think we can support it but it doesn t have to persistable meaning that within one editing session the user can create their calculated attribute to save time then dnd it etc but this will not exist in the next editing session i have seen this behavior in google sheets before with number format consider add a dropdown somewhere to enable disable panels to show as we start to have too many panels parameters setup explorer functions calculated attributes etc
0
17,088
22,596,210,733
IssuesEvent
2022-06-29 03:29:31
camunda/feel-scala
https://api.github.com/repos/camunda/feel-scala
closed
Boolean Expressions: Parsing error when using Conjunction ("And") or Disjunction ("Or")
type: bug team/process-automation
**Describe the bug** We use FEEL in Camunda's DMN tables to evaluate expressions. After we upgraded from Camunda `7.16.7` (with FEEL `1.13.3`) to `7.17.1` (FEEL `1.14.2`) we discovered a parsing error in FEEL when it comes to evaluating boolean expressions. ![image](https://user-images.githubusercontent.com/1492127/169830172-3bc84bf0-f36d-46b2-bf6e-e2f025fd30d3.png) `FEEL/SCALA-01008 Error while evaluating expression: failed to parse expression 'true and null': Expected ("," | end-of-input):1:6, found "and null"` Before the upgrade the expression "true and null" could be evaulated without problems and resulted in "null" (which is in line with https://camunda.github.io/feel-scala/docs/1.14/reference/language-guide/feel-boolean-expressions/#conjunction--and). When the expression is enclosed in paretheses tha parsing error does not appear. So this is OK: `(true and null)` The same problem also arises when using `or` in a boolean expression, e.g. `true or true` **To Reproduce** Steps to reproduce the behavior: 1. Create a DMN table in Camunda modeler like in the screenshot above. 2. Call `DmnDecisionTableResult.getDecisionResult(..)` **Expected behavior** No parsing error should occur when two booleans are combined with `and` or `or`. **Environment** * FEEL engine version: [1.14.2] * Affects: * Camunda Automation Platform 7: [7.17.1]
1.0
Boolean Expressions: Parsing error when using Conjunction ("And") or Disjunction ("Or") - **Describe the bug** We use FEEL in Camunda's DMN tables to evaluate expressions. After we upgraded from Camunda `7.16.7` (with FEEL `1.13.3`) to `7.17.1` (FEEL `1.14.2`) we discovered a parsing error in FEEL when it comes to evaluating boolean expressions. ![image](https://user-images.githubusercontent.com/1492127/169830172-3bc84bf0-f36d-46b2-bf6e-e2f025fd30d3.png) `FEEL/SCALA-01008 Error while evaluating expression: failed to parse expression 'true and null': Expected ("," | end-of-input):1:6, found "and null"` Before the upgrade the expression "true and null" could be evaulated without problems and resulted in "null" (which is in line with https://camunda.github.io/feel-scala/docs/1.14/reference/language-guide/feel-boolean-expressions/#conjunction--and). When the expression is enclosed in paretheses tha parsing error does not appear. So this is OK: `(true and null)` The same problem also arises when using `or` in a boolean expression, e.g. `true or true` **To Reproduce** Steps to reproduce the behavior: 1. Create a DMN table in Camunda modeler like in the screenshot above. 2. Call `DmnDecisionTableResult.getDecisionResult(..)` **Expected behavior** No parsing error should occur when two booleans are combined with `and` or `or`. **Environment** * FEEL engine version: [1.14.2] * Affects: * Camunda Automation Platform 7: [7.17.1]
process
boolean expressions parsing error when using conjunction and or disjunction or describe the bug we use feel in camunda s dmn tables to evaluate expressions after we upgraded from camunda with feel to feel we discovered a parsing error in feel when it comes to evaluating boolean expressions feel scala error while evaluating expression failed to parse expression true and null expected end of input found and null before the upgrade the expression true and null could be evaulated without problems and resulted in null which is in line with when the expression is enclosed in paretheses tha parsing error does not appear so this is ok true and null the same problem also arises when using or in a boolean expression e g true or true to reproduce steps to reproduce the behavior create a dmn table in camunda modeler like in the screenshot above call dmndecisiontableresult getdecisionresult expected behavior no parsing error should occur when two booleans are combined with and or or environment feel engine version affects camunda automation platform
1
407,604
11,924,455,436
IssuesEvent
2020-04-01 09:33:28
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
members.naughtyamerica.com - video or audio doesn't play
browser-firefox-reality engine-gecko priority-normal
<!-- @browser: Android 7.1.1 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile VR; rv:75.0) Gecko/75.0 Firefox/75.0 --> <!-- @reported_with: browser-fxr --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/50906 --> <!-- @extra_labels: browser-firefox-reality --> **URL**: https://members.naughtyamerica.com/scene/carmen-caliente-emma-hix-and-tiffany-watsons-spring-break-foursome-25904? **Browser / Version**: Android 7.1.1 **Operating System**: Android 7.1.1 **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: cant visualise and stream file **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
members.naughtyamerica.com - video or audio doesn't play - <!-- @browser: Android 7.1.1 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile VR; rv:75.0) Gecko/75.0 Firefox/75.0 --> <!-- @reported_with: browser-fxr --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/50906 --> <!-- @extra_labels: browser-firefox-reality --> **URL**: https://members.naughtyamerica.com/scene/carmen-caliente-emma-hix-and-tiffany-watsons-spring-break-foursome-25904? **Browser / Version**: Android 7.1.1 **Operating System**: Android 7.1.1 **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: cant visualise and stream file **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
members naughtyamerica com video or audio doesn t play url browser version android operating system android tested another browser no problem type video or audio doesn t play description cant visualise and stream file steps to reproduce browser configuration none from with ❤️
0
680,741
23,284,153,548
IssuesEvent
2022-08-05 14:50:02
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.radiofrance.fr - video or audio doesn't play
browser-firefox priority-normal os-mac engine-gecko
<!-- @browser: Firefox 64.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.11; rv:64.0) Gecko/20100101 Firefox/64.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/108506 --> **URL**: https://www.radiofrance.fr/franceinter/podcasts/sous-le-soleil-de-platon/sous-le-soleil-de-platon-du-mercredi-03-aout-2022-5865357 **Browser / Version**: Firefox 64.0 **Operating System**: Mac OS X 10.11 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Le bouton play ne s'active plus, je ne peux plus écouter les podcast ni avec firefox ni avec chrome <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/070486ea-16f1-4aed-a851-65a42a42e3ba.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190108160530</li><li>channel: release</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/8/28180b84-583b-4627-b37b-9c13e5116759) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.radiofrance.fr - video or audio doesn't play - <!-- @browser: Firefox 64.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.11; rv:64.0) Gecko/20100101 Firefox/64.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/108506 --> **URL**: https://www.radiofrance.fr/franceinter/podcasts/sous-le-soleil-de-platon/sous-le-soleil-de-platon-du-mercredi-03-aout-2022-5865357 **Browser / Version**: Firefox 64.0 **Operating System**: Mac OS X 10.11 **Tested Another Browser**: Yes Chrome **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: Le bouton play ne s'active plus, je ne peux plus écouter les podcast ni avec firefox ni avec chrome <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/8/070486ea-16f1-4aed-a851-65a42a42e3ba.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190108160530</li><li>channel: release</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/8/28180b84-583b-4627-b37b-9c13e5116759) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
video or audio doesn t play url browser version firefox operating system mac os x tested another browser yes chrome problem type video or audio doesn t play description the video or audio does not play steps to reproduce le bouton play ne s active plus je ne peux plus écouter les podcast ni avec firefox ni avec chrome view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
5,198
7,974,035,319
IssuesEvent
2018-07-17 02:51:39
pelias/pelias
https://api.github.com/repos/pelias/pelias
closed
Build Reporting [PL-IMO7]
processed story
Save metrics after each build so we can measure improvement/regression. related issues: - https://trello.com/c/aMEkdmdH/132-discuss-build-reporting
1.0
Build Reporting [PL-IMO7] - Save metrics after each build so we can measure improvement/regression. related issues: - https://trello.com/c/aMEkdmdH/132-discuss-build-reporting
process
build reporting save metrics after each build so we can measure improvement regression related issues
1
385
2,829,717,662
IssuesEvent
2015-05-23 04:30:23
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Suspected memory leak in OT
bug preprocess
We're building a bookmap with just over 1000 topics, and we're seeing what appears to be evidence of a memory leak in the OT 1.7 and 1.7.5. We allocated 5MB in ANT_OPTS. Physical memory usage started at 2GB and then during the gen-list and filter phases it grew linearly to 6MB, then leveled off and started decreasing slightly. We saw a number of instances of the CPU pinned at 100% during this memory climb. If we allocate less than 5MB, the OT spends about half its time at 100% CPU and the build takes very long. Allocating less than 4MB causes the too-much-time-in-garbage-collection abort. Enclosed is a screenshot of the memory usage graph during just the gen-list and filter phases. This is reproduceable on 3 different machines with this source input. ![leak](https://f.cloud.github.com/assets/2956867/558668/9c90593e-c429-11e2-9a3c-e27af694c6ab.gif)
1.0
Suspected memory leak in OT - We're building a bookmap with just over 1000 topics, and we're seeing what appears to be evidence of a memory leak in the OT 1.7 and 1.7.5. We allocated 5MB in ANT_OPTS. Physical memory usage started at 2GB and then during the gen-list and filter phases it grew linearly to 6MB, then leveled off and started decreasing slightly. We saw a number of instances of the CPU pinned at 100% during this memory climb. If we allocate less than 5MB, the OT spends about half its time at 100% CPU and the build takes very long. Allocating less than 4MB causes the too-much-time-in-garbage-collection abort. Enclosed is a screenshot of the memory usage graph during just the gen-list and filter phases. This is reproduceable on 3 different machines with this source input. ![leak](https://f.cloud.github.com/assets/2956867/558668/9c90593e-c429-11e2-9a3c-e27af694c6ab.gif)
process
suspected memory leak in ot we re building a bookmap with just over topics and we re seeing what appears to be evidence of a memory leak in the ot and we allocated in ant opts physical memory usage started at and then during the gen list and filter phases it grew linearly to then leveled off and started decreasing slightly we saw a number of instances of the cpu pinned at during this memory climb if we allocate less than the ot spends about half its time at cpu and the build takes very long allocating less than causes the too much time in garbage collection abort enclosed is a screenshot of the memory usage graph during just the gen list and filter phases this is reproduceable on different machines with this source input
1
11,992
14,737,215,782
IssuesEvent
2021-01-07 01:12:00
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Towne - question about payments
anc-external anc-process anp-1 ant-support
In GitLab by @kdjstudios on Apr 25, 2018, 16:29 **Submitted by:** Deb Crown <dcrown@towneanswering.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-25-59434/conversation **Server:** External **Client/Site:** Towne **Account:** NA **Issue:** This is a general question about posting payments and how SA posts them. I know that –at some time in the past- SA changed how payments were posted. I believe it used to break up the payment and post it to outstanding invoices instead of keeping it as one payment. I have a client to whom I am going to provide his payment history and I know it is going to be a problem for him to match things up. Could you please tell me the date of when that change was made in SA? Looking at the history, it looks like it might be the end of 2015?
1.0
Towne - question about payments - In GitLab by @kdjstudios on Apr 25, 2018, 16:29 **Submitted by:** Deb Crown <dcrown@towneanswering.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-25-59434/conversation **Server:** External **Client/Site:** Towne **Account:** NA **Issue:** This is a general question about posting payments and how SA posts them. I know that –at some time in the past- SA changed how payments were posted. I believe it used to break up the payment and post it to outstanding invoices instead of keeping it as one payment. I have a client to whom I am going to provide his payment history and I know it is going to be a problem for him to match things up. Could you please tell me the date of when that change was made in SA? Looking at the history, it looks like it might be the end of 2015?
process
towne question about payments in gitlab by kdjstudios on apr submitted by deb crown helpdesk server external client site towne account na issue this is a general question about posting payments and how sa posts them i know that –at some time in the past sa changed how payments were posted i believe it used to break up the payment and post it to outstanding invoices instead of keeping it as one payment i have a client to whom i am going to provide his payment history and i know it is going to be a problem for him to match things up could you please tell me the date of when that change was made in sa looking at the history it looks like it might be the end of
1
285,878
21,557,285,503
IssuesEvent
2022-04-30 16:37:46
skotlex/ffxiv-material-ui
https://api.github.com/repos/skotlex/ffxiv-material-ui
closed
Correct the URL for Dotnet in the Guide.
documentation
Textools is built on the 4.x version of the framework and the URL in the Guide is for 5.x, there is no cross compatibility between the two.
1.0
Correct the URL for Dotnet in the Guide. - Textools is built on the 4.x version of the framework and the URL in the Guide is for 5.x, there is no cross compatibility between the two.
non_process
correct the url for dotnet in the guide textools is built on the x version of the framework and the url in the guide is for x there is no cross compatibility between the two
0
164,420
25,963,918,870
IssuesEvent
2022-12-19 04:06:29
psf/black
https://api.github.com/repos/psf/black
closed
Unnecessary split in left hand side of assignment due to multiline string
T: design F: linebreak
Hi, I'm experience strange behavior when I use dicts and triple-quotes strings, for example: ```python a["b"] = f"""Some really long long string.""" ``` ends up like this: ```python a[ "b" ] = f"""Some really long long string.""" ``` Is it expected behavior or? If yes, how can one disable it? Operating system: Linux Ubuntu 18.04 Python version: 3.6 *Black* version: 18.9b0 Does also happen on master: yes
1.0
Unnecessary split in left hand side of assignment due to multiline string - Hi, I'm experience strange behavior when I use dicts and triple-quotes strings, for example: ```python a["b"] = f"""Some really long long string.""" ``` ends up like this: ```python a[ "b" ] = f"""Some really long long string.""" ``` Is it expected behavior or? If yes, how can one disable it? Operating system: Linux Ubuntu 18.04 Python version: 3.6 *Black* version: 18.9b0 Does also happen on master: yes
non_process
unnecessary split in left hand side of assignment due to multiline string hi i m experience strange behavior when i use dicts and triple quotes strings for example python a f some really long long string ends up like this python a b f some really long long string is it expected behavior or if yes how can one disable it operating system linux ubuntu python version black version does also happen on master yes
0
17,497
23,305,508,086
IssuesEvent
2022-08-07 23:50:05
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Million Dollar Heads or Tails from HIMYM
suggested title in process
Please add as much of the following info as you can: Title: Million Dollar Heads or Tails Type (film/tv show):TV show Film or show in which it appears: How I Met Your Mother Is the parent film/show streaming anywhere? Clip at https://www.youtube.com/watch?v=TDmcPUQj_G0&ab_channel=CopyrightStriked About when in the parent film/show does it appear? Beginning of this clip Actual footage of the film/show can be seen (yes/no)? Yes "Marnie, flip that coin!"
1.0
Add Million Dollar Heads or Tails from HIMYM - Please add as much of the following info as you can: Title: Million Dollar Heads or Tails Type (film/tv show):TV show Film or show in which it appears: How I Met Your Mother Is the parent film/show streaming anywhere? Clip at https://www.youtube.com/watch?v=TDmcPUQj_G0&ab_channel=CopyrightStriked About when in the parent film/show does it appear? Beginning of this clip Actual footage of the film/show can be seen (yes/no)? Yes "Marnie, flip that coin!"
process
add million dollar heads or tails from himym please add as much of the following info as you can title million dollar heads or tails type film tv show tv show film or show in which it appears how i met your mother is the parent film show streaming anywhere clip at about when in the parent film show does it appear beginning of this clip actual footage of the film show can be seen yes no yes marnie flip that coin
1
21,323
28,932,882,618
IssuesEvent
2023-05-09 02:00:08
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Tue, 9 May 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast ### Structural and Statistical Texture Knowledge Distillation for Semantic Segmentation - **Authors:** Deyi Ji, Haoran Wang, Mingyuan Tao, Jianqiang Huang, Xian-Sheng Hua, Hongtao Lu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.03944 - **Pdf link:** https://arxiv.org/pdf/2305.03944 - **Abstract** Existing knowledge distillation works for semantic segmentation mainly focus on transferring high-level contextual knowledge from teacher to student. However, low-level texture knowledge is also of vital importance for characterizing the local structural pattern and global statistical property, such as boundary, smoothness, regularity and color contrast, which may not be well addressed by high-level deep features. In this paper, we are intended to take full advantage of both structural and statistical texture knowledge and propose a novel Structural and Statistical Texture Knowledge Distillation (SSTKD) framework for semantic segmentation. Specifically, for structural texture knowledge, we introduce a Contourlet Decomposition Module (CDM) that decomposes low-level features with iterative Laplacian pyramid and directional filter bank to mine the structural texture knowledge. For statistical knowledge, we propose a Denoised Texture Intensity Equalization Module (DTIEM) to adaptively extract and enhance statistical texture knowledge through heuristics iterative quantization and denoised operation. Finally, each knowledge learning is supervised by an individual loss function, forcing the student network to mimic the teacher better from a broader perspective. Experiments show that the proposed method achieves state-of-the-art performance on Cityscapes, Pascal VOC 2012 and ADE20K datasets. ## Keyword: AWB ### NL-CS Net: Deep Learning with Non-Local Prior for Image Compressive Sensing - **Authors:** Shuai Bian, Shouliang Qi, Chen Li, Yudong Yao, Yueyang Teng - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2305.03899 - **Pdf link:** https://arxiv.org/pdf/2305.03899 - **Abstract** Deep learning has been applied to compressive sensing (CS) of images successfully in recent years. However, existing network-based methods are often trained as the black box, in which the lack of prior knowledge is often the bottleneck for further performance improvement. To overcome this drawback, this paper proposes a novel CS method using non-local prior which combines the interpretability of the traditional optimization methods with the speed of network-based methods, called NL-CS Net. We unroll each phase from iteration of the augmented Lagrangian method solving non-local and sparse regularized optimization problem by a network. NL-CS Net is composed of the up-sampling module and the recovery module. In the up-sampling module, we use learnable up-sampling matrix instead of a predefined one. In the recovery module, patch-wise non-local network is employed to capture long-range feature correspondences. Important parameters involved (e.g. sampling matrix, nonlinear transforms, shrinkage thresholds, step size, $etc.$) are learned end-to-end, rather than hand-crafted. Furthermore, to facilitate practical implementation, orthogonal and binary constraints on the sampling matrix are simultaneously adopted. Extensive experiments on natural images and magnetic resonance imaging (MRI) demonstrate that the proposed method outperforms the state-of-the-art methods while maintaining great interpretability and speed. ### Scene Text Recognition with Image-Text Matching-guided Dictionary - **Authors:** Jiajun Wei, Hongjian Zhan, Xiao Tu, Yue Lu, Umapada Pal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04524 - **Pdf link:** https://arxiv.org/pdf/2305.04524 - **Abstract** Employing a dictionary can efficiently rectify the deviation between the visual prediction and the ground truth in scene text recognition methods. However, the independence of the dictionary on the visual features may lead to incorrect rectification of accurate visual predictions. In this paper, we propose a new dictionary language model leveraging the Scene Image-Text Matching(SITM) network, which avoids the drawbacks of the explicit dictionary language model: 1) the independence of the visual features; 2) noisy choice in candidates etc. The SITM network accomplishes this by using Image-Text Contrastive (ITC) Learning to match an image with its corresponding text among candidates in the inference stage. ITC is widely used in vision-language learning to pull the positive image-text pair closer in feature space. Inspired by ITC, the SITM network combines the visual features and the text features of all candidates to identify the candidate with the minimum distance in the feature space. Our lexicon method achieves better results(93.8\% accuracy) than the ordinary method results(92.1\% accuracy) on six mainstream benchmarks. Additionally, we integrate our method with ABINet and establish new state-of-the-art results on several benchmarks. ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### YOLOCS: Object Detection based on Dense Channel Compression for Feature Spatial Solidification - **Authors:** Lin Huang, Weisheng Li, Linlin Shen, Haojie Fu, Xue Xiao, Suihan Xiao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04170 - **Pdf link:** https://arxiv.org/pdf/2305.04170 - **Abstract** In this study, we examine the associations between channel features and convolutional kernels during the processes of feature purification and gradient backpropagation, with a focus on the forward and backward propagation within the network. Consequently, we propose a method called Dense Channel Compression for Feature Spatial Solidification. Drawing upon the central concept of this method, we introduce two innovative modules for backbone and head networks: the Dense Channel Compression for Feature Spatial Solidification Structure (DCFS) and the Asymmetric Multi-Level Compression Decoupled Head (ADH). When integrated into the YOLOv5 model, these two modules demonstrate exceptional performance, resulting in a modified model referred to as YOLOCS. Evaluated on the MSCOCO dataset, the large, medium, and small YOLOCS models yield AP of 50.1%, 47.6%, and 42.5%, respectively. Maintaining inference speeds remarkably similar to those of the YOLOv5 model, the large, medium, and small YOLOCS models surpass the YOLOv5 model's AP by 1.1%, 2.3%, and 5.2%, respectively. ### Adversarial Examples Detection with Enhanced Image Difference Features based on Local Histogram Equalization - **Authors:** Zhaoxia Yin, Shaowei Zhu, Hang Su, Jianteng Peng, Wanli Lyu, Bin Luo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04436 - **Pdf link:** https://arxiv.org/pdf/2305.04436 - **Abstract** Deep Neural Networks (DNNs) have recently made significant progress in many fields. However, studies have shown that DNNs are vulnerable to adversarial examples, where imperceptible perturbations can greatly mislead DNNs even if the full underlying model parameters are not accessible. Various defense methods have been proposed, such as feature compression and gradient masking. However, numerous studies have proven that previous methods create detection or defense against certain attacks, which renders the method ineffective in the face of the latest unknown attack methods. The invisibility of adversarial perturbations is one of the evaluation indicators for adversarial example attacks, which also means that the difference in the local correlation of high-frequency information in adversarial examples and normal examples can be used as an effective feature to distinguish the two. Therefore, we propose an adversarial example detection framework based on a high-frequency information enhancement strategy, which can effectively extract and amplify the feature differences between adversarial examples and normal examples. Experimental results show that the feature augmentation module can be combined with existing detection models in a modular way under this framework. Improve the detector's performance and reduce the deployment cost without modifying the existing detection model. ### SNT: Sharpness-Minimizing Network Transformation for Fast Compression-friendly Pretraining - **Authors:** Jung Hwan Heo, Seyedarmin Azizi, Arash Fayyazi, Mahdi Nazemi, Massoud Pedram - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04526 - **Pdf link:** https://arxiv.org/pdf/2305.04526 - **Abstract** Model compression has become the de-facto approach for optimizing the efficiency of vision models. Recently, the focus of most compression efforts has shifted to post-training scenarios due to the very high cost of large-scale pretraining. This has created the need to build compressible models from scratch, which can effectively be compressed after training. In this work, we present a sharpness-minimizing network transformation (SNT) method applied during pretraining that can create models with desirable compressibility and generalizability features. We compare our approach to a well-known sharpness-minimizing optimizer to validate its efficacy in creating a flat loss landscape. To the best of our knowledge, SNT is the first pretraining method that uses an architectural transformation to generate compression-friendly networks. We find that SNT generalizes across different compression tasks and network backbones, delivering consistent improvements over the ADAM baseline with up to 2% accuracy improvement on weight pruning and 5.4% accuracy improvement on quantization. Code to reproduce our results will be made publicly available. ## Keyword: RAW ### NL-CS Net: Deep Learning with Non-Local Prior for Image Compressive Sensing - **Authors:** Shuai Bian, Shouliang Qi, Chen Li, Yudong Yao, Yueyang Teng - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2305.03899 - **Pdf link:** https://arxiv.org/pdf/2305.03899 - **Abstract** Deep learning has been applied to compressive sensing (CS) of images successfully in recent years. However, existing network-based methods are often trained as the black box, in which the lack of prior knowledge is often the bottleneck for further performance improvement. To overcome this drawback, this paper proposes a novel CS method using non-local prior which combines the interpretability of the traditional optimization methods with the speed of network-based methods, called NL-CS Net. We unroll each phase from iteration of the augmented Lagrangian method solving non-local and sparse regularized optimization problem by a network. NL-CS Net is composed of the up-sampling module and the recovery module. In the up-sampling module, we use learnable up-sampling matrix instead of a predefined one. In the recovery module, patch-wise non-local network is employed to capture long-range feature correspondences. Important parameters involved (e.g. sampling matrix, nonlinear transforms, shrinkage thresholds, step size, $etc.$) are learned end-to-end, rather than hand-crafted. Furthermore, to facilitate practical implementation, orthogonal and binary constraints on the sampling matrix are simultaneously adopted. Extensive experiments on natural images and magnetic resonance imaging (MRI) demonstrate that the proposed method outperforms the state-of-the-art methods while maintaining great interpretability and speed. ### Prompt What You Need: Enhancing Segmentation in Rainy Scenes with Anchor-based Prompting - **Authors:** Xiaoyu Guo, Xiang Wei, Qi Su, Huiqin Zhao, Shunli Zhan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.03902 - **Pdf link:** https://arxiv.org/pdf/2305.03902 - **Abstract** Semantic segmentation in rainy scenes is a challenging task due to the complex environment, class distribution imbalance, and limited annotated data. To address these challenges, we propose a novel framework that utilizes semi-supervised learning and pre-trained segmentation foundation model to achieve superior performance. Specifically, our framework leverages the semi-supervised model as the basis for generating raw semantic segmentation results, while also serving as a guiding force to prompt pre-trained foundation model to compensate for knowledge gaps with entropy-based anchors. In addition, to minimize the impact of irrelevant segmentation masks generated by the pre-trained foundation model, we also propose a mask filtering and fusion mechanism that optimizes raw semantic segmentation results based on the principle of minimum risk. The proposed framework achieves superior segmentation performance on the Rainy WCity dataset and is awarded the first prize in the sub-track of STRAIN in ICME 2023 Grand Challenges. ### YOLOCS: Object Detection based on Dense Channel Compression for Feature Spatial Solidification - **Authors:** Lin Huang, Weisheng Li, Linlin Shen, Haojie Fu, Xue Xiao, Suihan Xiao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04170 - **Pdf link:** https://arxiv.org/pdf/2305.04170 - **Abstract** In this study, we examine the associations between channel features and convolutional kernels during the processes of feature purification and gradient backpropagation, with a focus on the forward and backward propagation within the network. Consequently, we propose a method called Dense Channel Compression for Feature Spatial Solidification. Drawing upon the central concept of this method, we introduce two innovative modules for backbone and head networks: the Dense Channel Compression for Feature Spatial Solidification Structure (DCFS) and the Asymmetric Multi-Level Compression Decoupled Head (ADH). When integrated into the YOLOv5 model, these two modules demonstrate exceptional performance, resulting in a modified model referred to as YOLOCS. Evaluated on the MSCOCO dataset, the large, medium, and small YOLOCS models yield AP of 50.1%, 47.6%, and 42.5%, respectively. Maintaining inference speeds remarkably similar to those of the YOLOv5 model, the large, medium, and small YOLOCS models surpass the YOLOv5 model's AP by 1.1%, 2.3%, and 5.2%, respectively. ### Segmentation of the veterinary cytological images for fast neoplastic tumors diagnosis - **Authors:** Jakub Grzeszczyk, Michał Karwatowski, Daria Łukasik, Maciej Wielgosz, Paweł Russek, Szymon Mazurek, Jakub Caputa, Rafał Frączek, Anna Śmiech, Ernest Jamro, Sebastian Koryciak, Agnieszka Dąbrowska-Boruch, Marcin Pietroń, Kazimierz Wiatr - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04332 - **Pdf link:** https://arxiv.org/pdf/2305.04332 - **Abstract** This paper shows the machine learning system which performs instance segmentation of cytological images in veterinary medicine. Eleven cell types were used directly and indirectly in the experiments, including damaged and unrecognized categories. The deep learning models employed in the system achieve a high score of average precision and recall metrics, i.e. 0.94 and 0.8 respectively, for the selected three types of tumors. This variety of label types allowed us to draw a meaningful conclusion that there are relatively few mistakes for tumor cell types. Additionally, the model learned tumor cell features well enough to avoid misclassification mistakes of one tumor type into another. The experiments also revealed that the quality of the results improves with the dataset size (excluding the damaged cells). It is worth noting that all the experiments were done using a custom dedicated dataset provided by the cooperating vet doctors. ### Scene Text Recognition with Image-Text Matching-guided Dictionary - **Authors:** Jiajun Wei, Hongjian Zhan, Xiao Tu, Yue Lu, Umapada Pal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04524 - **Pdf link:** https://arxiv.org/pdf/2305.04524 - **Abstract** Employing a dictionary can efficiently rectify the deviation between the visual prediction and the ground truth in scene text recognition methods. However, the independence of the dictionary on the visual features may lead to incorrect rectification of accurate visual predictions. In this paper, we propose a new dictionary language model leveraging the Scene Image-Text Matching(SITM) network, which avoids the drawbacks of the explicit dictionary language model: 1) the independence of the visual features; 2) noisy choice in candidates etc. The SITM network accomplishes this by using Image-Text Contrastive (ITC) Learning to match an image with its corresponding text among candidates in the inference stage. ITC is widely used in vision-language learning to pull the positive image-text pair closer in feature space. Inspired by ITC, the SITM network combines the visual features and the text features of all candidates to identify the candidate with the minimum distance in the feature space. Our lexicon method achieves better results(93.8\% accuracy) than the ordinary method results(92.1\% accuracy) on six mainstream benchmarks. Additionally, we integrate our method with ABINet and establish new state-of-the-art results on several benchmarks. ### High Quality Large-Scale 3-D Urban Mapping with Multi-Master TomoSAR - **Authors:** Yilei Shi, Richard Bamler, Yuanyuan Wang, Xiao Xiang Zhu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04541 - **Pdf link:** https://arxiv.org/pdf/2305.04541 - **Abstract** Multi-baseline interferometric synthetic aperture radar (InSAR) techniques are effective approaches for retrieving the 3-D information of urban areas. In order to obtain a plausible reconstruction, it is necessary to use large-stack interferograms. Hence, these methods are commonly not appropriate for large-scale 3-D urban mapping using TanDEM-X data where only a few acquisitions are available in average for each city. This work proposes a new SAR tomographic processing framework to work with those extremely small stacks, which integrates the non-local filtering into SAR tomography inversion. The applicability of the algorithm is demonstrated using a TanDEM-X multi-baseline stack with 5 bistatic interferograms over the whole city of Munich, Germany. Systematic comparison of our result with airborne LiDAR data shows that the relative height accuracy of two third buildings is within two meters, which outperforms the TanDEM-X raw DEM. The promising performance of the proposed algorithm paved the first step towards high quality large-scale 3-D urban mapping. ### MARS: Mask Attention Refinement with Sequential Quadtree Nodes for Car Damage Instance Segmentation - **Authors:** Teerapong Panboonyuen, Naphat Nithisopa, Panin Pienroj, Laphonchai Jirachuphun, Chaiwasut Watthanasirikrit, Naruepon Pornwiriyakul - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04743 - **Pdf link:** https://arxiv.org/pdf/2305.04743 - **Abstract** Evaluating car damages from misfortune is critical to the car insurance industry. However, the accuracy is still insufficient for real-world applications since the deep learning network is not designed for car damage images as inputs, and its segmented masks are still very coarse. This paper presents MARS (Mask Attention Refinement with Sequential quadtree nodes) for car damage instance segmentation. Our MARS represents self-attention mechanisms to draw global dependencies between the sequential quadtree nodes layer and quadtree transformer to recalibrate channel weights and predict highly accurate instance masks. Our extensive experiments demonstrate that MARS outperforms state-of-the-art (SOTA) instance segmentation methods on three popular benchmarks such as Mask R-CNN [9], PointRend [14], and Mask Transfiner [13], by a large margin of +1.3 maskAP-based R50-FPN backbone and +2.3 maskAP-based R101-FPN backbone on Thai car-damage dataset. Our demos are available at https://www.marssolution.io. ### BiRT: Bio-inspired Replay in Vision Transformers for Continual Learning - **Authors:** Kishaan Jeeveswaran, Prashant Bhat, Bahram Zonooz, Elahe Arani - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2305.04769 - **Pdf link:** https://arxiv.org/pdf/2305.04769 - **Abstract** The ability of deep neural networks to continually learn and adapt to a sequence of tasks has remained challenging due to catastrophic forgetting of previously learned tasks. Humans, on the other hand, have a remarkable ability to acquire, assimilate, and transfer knowledge across tasks throughout their lifetime without catastrophic forgetting. The versatility of the brain can be attributed to the rehearsal of abstract experiences through a complementary learning system. However, representation rehearsal in vision transformers lacks diversity, resulting in overfitting and consequently, performance drops significantly compared to raw image rehearsal. Therefore, we propose BiRT, a novel representation rehearsal-based continual learning approach using vision transformers. Specifically, we introduce constructive noises at various stages of the vision transformer and enforce consistency in predictions with respect to an exponential moving average of the working model. Our method provides consistent performance gain over raw image and vanilla representation rehearsal on several challenging CL benchmarks, while being memory efficient and robust to natural and adversarial corruptions. ### PillarNeXt: Rethinking Network Designs for 3D Object Detection in LiDAR Point Clouds - **Authors:** Jinyu Li, Chenxu Luo, Xiaodong Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04925 - **Pdf link:** https://arxiv.org/pdf/2305.04925 - **Abstract** In order to deal with the sparse and unstructured raw point clouds, LiDAR based 3D object detection research mostly focuses on designing dedicated local point aggregators for fine-grained geometrical modeling. In this paper, we revisit the local point aggregators from the perspective of allocating computational resources. We find that the simplest pillar based models perform surprisingly well considering both accuracy and latency. Additionally, we show that minimal adaptions from the success of 2D object detection, such as enlarging receptive field, significantly boost the performance. Extensive experiments reveal that our pillar based networks with modernized designs in terms of architecture and training render the state-of-the-art performance on the two popular benchmarks: Waymo Open Dataset and nuScenes. Our results challenge the common intuition that the detailed geometry modeling is essential to achieve high performance for 3D object detection. ## Keyword: raw image ### BiRT: Bio-inspired Replay in Vision Transformers for Continual Learning - **Authors:** Kishaan Jeeveswaran, Prashant Bhat, Bahram Zonooz, Elahe Arani - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2305.04769 - **Pdf link:** https://arxiv.org/pdf/2305.04769 - **Abstract** The ability of deep neural networks to continually learn and adapt to a sequence of tasks has remained challenging due to catastrophic forgetting of previously learned tasks. Humans, on the other hand, have a remarkable ability to acquire, assimilate, and transfer knowledge across tasks throughout their lifetime without catastrophic forgetting. The versatility of the brain can be attributed to the rehearsal of abstract experiences through a complementary learning system. However, representation rehearsal in vision transformers lacks diversity, resulting in overfitting and consequently, performance drops significantly compared to raw image rehearsal. Therefore, we propose BiRT, a novel representation rehearsal-based continual learning approach using vision transformers. Specifically, we introduce constructive noises at various stages of the vision transformer and enforce consistency in predictions with respect to an exponential moving average of the working model. Our method provides consistent performance gain over raw image and vanilla representation rehearsal on several challenging CL benchmarks, while being memory efficient and robust to natural and adversarial corruptions.
2.0
New submissions for Tue, 9 May 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast ### Structural and Statistical Texture Knowledge Distillation for Semantic Segmentation - **Authors:** Deyi Ji, Haoran Wang, Mingyuan Tao, Jianqiang Huang, Xian-Sheng Hua, Hongtao Lu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.03944 - **Pdf link:** https://arxiv.org/pdf/2305.03944 - **Abstract** Existing knowledge distillation works for semantic segmentation mainly focus on transferring high-level contextual knowledge from teacher to student. However, low-level texture knowledge is also of vital importance for characterizing the local structural pattern and global statistical property, such as boundary, smoothness, regularity and color contrast, which may not be well addressed by high-level deep features. In this paper, we are intended to take full advantage of both structural and statistical texture knowledge and propose a novel Structural and Statistical Texture Knowledge Distillation (SSTKD) framework for semantic segmentation. Specifically, for structural texture knowledge, we introduce a Contourlet Decomposition Module (CDM) that decomposes low-level features with iterative Laplacian pyramid and directional filter bank to mine the structural texture knowledge. For statistical knowledge, we propose a Denoised Texture Intensity Equalization Module (DTIEM) to adaptively extract and enhance statistical texture knowledge through heuristics iterative quantization and denoised operation. Finally, each knowledge learning is supervised by an individual loss function, forcing the student network to mimic the teacher better from a broader perspective. Experiments show that the proposed method achieves state-of-the-art performance on Cityscapes, Pascal VOC 2012 and ADE20K datasets. ## Keyword: AWB ### NL-CS Net: Deep Learning with Non-Local Prior for Image Compressive Sensing - **Authors:** Shuai Bian, Shouliang Qi, Chen Li, Yudong Yao, Yueyang Teng - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2305.03899 - **Pdf link:** https://arxiv.org/pdf/2305.03899 - **Abstract** Deep learning has been applied to compressive sensing (CS) of images successfully in recent years. However, existing network-based methods are often trained as the black box, in which the lack of prior knowledge is often the bottleneck for further performance improvement. To overcome this drawback, this paper proposes a novel CS method using non-local prior which combines the interpretability of the traditional optimization methods with the speed of network-based methods, called NL-CS Net. We unroll each phase from iteration of the augmented Lagrangian method solving non-local and sparse regularized optimization problem by a network. NL-CS Net is composed of the up-sampling module and the recovery module. In the up-sampling module, we use learnable up-sampling matrix instead of a predefined one. In the recovery module, patch-wise non-local network is employed to capture long-range feature correspondences. Important parameters involved (e.g. sampling matrix, nonlinear transforms, shrinkage thresholds, step size, $etc.$) are learned end-to-end, rather than hand-crafted. Furthermore, to facilitate practical implementation, orthogonal and binary constraints on the sampling matrix are simultaneously adopted. Extensive experiments on natural images and magnetic resonance imaging (MRI) demonstrate that the proposed method outperforms the state-of-the-art methods while maintaining great interpretability and speed. ### Scene Text Recognition with Image-Text Matching-guided Dictionary - **Authors:** Jiajun Wei, Hongjian Zhan, Xiao Tu, Yue Lu, Umapada Pal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04524 - **Pdf link:** https://arxiv.org/pdf/2305.04524 - **Abstract** Employing a dictionary can efficiently rectify the deviation between the visual prediction and the ground truth in scene text recognition methods. However, the independence of the dictionary on the visual features may lead to incorrect rectification of accurate visual predictions. In this paper, we propose a new dictionary language model leveraging the Scene Image-Text Matching(SITM) network, which avoids the drawbacks of the explicit dictionary language model: 1) the independence of the visual features; 2) noisy choice in candidates etc. The SITM network accomplishes this by using Image-Text Contrastive (ITC) Learning to match an image with its corresponding text among candidates in the inference stage. ITC is widely used in vision-language learning to pull the positive image-text pair closer in feature space. Inspired by ITC, the SITM network combines the visual features and the text features of all candidates to identify the candidate with the minimum distance in the feature space. Our lexicon method achieves better results(93.8\% accuracy) than the ordinary method results(92.1\% accuracy) on six mainstream benchmarks. Additionally, we integrate our method with ABINet and establish new state-of-the-art results on several benchmarks. ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### YOLOCS: Object Detection based on Dense Channel Compression for Feature Spatial Solidification - **Authors:** Lin Huang, Weisheng Li, Linlin Shen, Haojie Fu, Xue Xiao, Suihan Xiao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04170 - **Pdf link:** https://arxiv.org/pdf/2305.04170 - **Abstract** In this study, we examine the associations between channel features and convolutional kernels during the processes of feature purification and gradient backpropagation, with a focus on the forward and backward propagation within the network. Consequently, we propose a method called Dense Channel Compression for Feature Spatial Solidification. Drawing upon the central concept of this method, we introduce two innovative modules for backbone and head networks: the Dense Channel Compression for Feature Spatial Solidification Structure (DCFS) and the Asymmetric Multi-Level Compression Decoupled Head (ADH). When integrated into the YOLOv5 model, these two modules demonstrate exceptional performance, resulting in a modified model referred to as YOLOCS. Evaluated on the MSCOCO dataset, the large, medium, and small YOLOCS models yield AP of 50.1%, 47.6%, and 42.5%, respectively. Maintaining inference speeds remarkably similar to those of the YOLOv5 model, the large, medium, and small YOLOCS models surpass the YOLOv5 model's AP by 1.1%, 2.3%, and 5.2%, respectively. ### Adversarial Examples Detection with Enhanced Image Difference Features based on Local Histogram Equalization - **Authors:** Zhaoxia Yin, Shaowei Zhu, Hang Su, Jianteng Peng, Wanli Lyu, Bin Luo - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04436 - **Pdf link:** https://arxiv.org/pdf/2305.04436 - **Abstract** Deep Neural Networks (DNNs) have recently made significant progress in many fields. However, studies have shown that DNNs are vulnerable to adversarial examples, where imperceptible perturbations can greatly mislead DNNs even if the full underlying model parameters are not accessible. Various defense methods have been proposed, such as feature compression and gradient masking. However, numerous studies have proven that previous methods create detection or defense against certain attacks, which renders the method ineffective in the face of the latest unknown attack methods. The invisibility of adversarial perturbations is one of the evaluation indicators for adversarial example attacks, which also means that the difference in the local correlation of high-frequency information in adversarial examples and normal examples can be used as an effective feature to distinguish the two. Therefore, we propose an adversarial example detection framework based on a high-frequency information enhancement strategy, which can effectively extract and amplify the feature differences between adversarial examples and normal examples. Experimental results show that the feature augmentation module can be combined with existing detection models in a modular way under this framework. Improve the detector's performance and reduce the deployment cost without modifying the existing detection model. ### SNT: Sharpness-Minimizing Network Transformation for Fast Compression-friendly Pretraining - **Authors:** Jung Hwan Heo, Seyedarmin Azizi, Arash Fayyazi, Mahdi Nazemi, Massoud Pedram - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04526 - **Pdf link:** https://arxiv.org/pdf/2305.04526 - **Abstract** Model compression has become the de-facto approach for optimizing the efficiency of vision models. Recently, the focus of most compression efforts has shifted to post-training scenarios due to the very high cost of large-scale pretraining. This has created the need to build compressible models from scratch, which can effectively be compressed after training. In this work, we present a sharpness-minimizing network transformation (SNT) method applied during pretraining that can create models with desirable compressibility and generalizability features. We compare our approach to a well-known sharpness-minimizing optimizer to validate its efficacy in creating a flat loss landscape. To the best of our knowledge, SNT is the first pretraining method that uses an architectural transformation to generate compression-friendly networks. We find that SNT generalizes across different compression tasks and network backbones, delivering consistent improvements over the ADAM baseline with up to 2% accuracy improvement on weight pruning and 5.4% accuracy improvement on quantization. Code to reproduce our results will be made publicly available. ## Keyword: RAW ### NL-CS Net: Deep Learning with Non-Local Prior for Image Compressive Sensing - **Authors:** Shuai Bian, Shouliang Qi, Chen Li, Yudong Yao, Yueyang Teng - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Image and Video Processing (eess.IV) - **Arxiv link:** https://arxiv.org/abs/2305.03899 - **Pdf link:** https://arxiv.org/pdf/2305.03899 - **Abstract** Deep learning has been applied to compressive sensing (CS) of images successfully in recent years. However, existing network-based methods are often trained as the black box, in which the lack of prior knowledge is often the bottleneck for further performance improvement. To overcome this drawback, this paper proposes a novel CS method using non-local prior which combines the interpretability of the traditional optimization methods with the speed of network-based methods, called NL-CS Net. We unroll each phase from iteration of the augmented Lagrangian method solving non-local and sparse regularized optimization problem by a network. NL-CS Net is composed of the up-sampling module and the recovery module. In the up-sampling module, we use learnable up-sampling matrix instead of a predefined one. In the recovery module, patch-wise non-local network is employed to capture long-range feature correspondences. Important parameters involved (e.g. sampling matrix, nonlinear transforms, shrinkage thresholds, step size, $etc.$) are learned end-to-end, rather than hand-crafted. Furthermore, to facilitate practical implementation, orthogonal and binary constraints on the sampling matrix are simultaneously adopted. Extensive experiments on natural images and magnetic resonance imaging (MRI) demonstrate that the proposed method outperforms the state-of-the-art methods while maintaining great interpretability and speed. ### Prompt What You Need: Enhancing Segmentation in Rainy Scenes with Anchor-based Prompting - **Authors:** Xiaoyu Guo, Xiang Wei, Qi Su, Huiqin Zhao, Shunli Zhan - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.03902 - **Pdf link:** https://arxiv.org/pdf/2305.03902 - **Abstract** Semantic segmentation in rainy scenes is a challenging task due to the complex environment, class distribution imbalance, and limited annotated data. To address these challenges, we propose a novel framework that utilizes semi-supervised learning and pre-trained segmentation foundation model to achieve superior performance. Specifically, our framework leverages the semi-supervised model as the basis for generating raw semantic segmentation results, while also serving as a guiding force to prompt pre-trained foundation model to compensate for knowledge gaps with entropy-based anchors. In addition, to minimize the impact of irrelevant segmentation masks generated by the pre-trained foundation model, we also propose a mask filtering and fusion mechanism that optimizes raw semantic segmentation results based on the principle of minimum risk. The proposed framework achieves superior segmentation performance on the Rainy WCity dataset and is awarded the first prize in the sub-track of STRAIN in ICME 2023 Grand Challenges. ### YOLOCS: Object Detection based on Dense Channel Compression for Feature Spatial Solidification - **Authors:** Lin Huang, Weisheng Li, Linlin Shen, Haojie Fu, Xue Xiao, Suihan Xiao - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04170 - **Pdf link:** https://arxiv.org/pdf/2305.04170 - **Abstract** In this study, we examine the associations between channel features and convolutional kernels during the processes of feature purification and gradient backpropagation, with a focus on the forward and backward propagation within the network. Consequently, we propose a method called Dense Channel Compression for Feature Spatial Solidification. Drawing upon the central concept of this method, we introduce two innovative modules for backbone and head networks: the Dense Channel Compression for Feature Spatial Solidification Structure (DCFS) and the Asymmetric Multi-Level Compression Decoupled Head (ADH). When integrated into the YOLOv5 model, these two modules demonstrate exceptional performance, resulting in a modified model referred to as YOLOCS. Evaluated on the MSCOCO dataset, the large, medium, and small YOLOCS models yield AP of 50.1%, 47.6%, and 42.5%, respectively. Maintaining inference speeds remarkably similar to those of the YOLOv5 model, the large, medium, and small YOLOCS models surpass the YOLOv5 model's AP by 1.1%, 2.3%, and 5.2%, respectively. ### Segmentation of the veterinary cytological images for fast neoplastic tumors diagnosis - **Authors:** Jakub Grzeszczyk, Michał Karwatowski, Daria Łukasik, Maciej Wielgosz, Paweł Russek, Szymon Mazurek, Jakub Caputa, Rafał Frączek, Anna Śmiech, Ernest Jamro, Sebastian Koryciak, Agnieszka Dąbrowska-Boruch, Marcin Pietroń, Kazimierz Wiatr - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04332 - **Pdf link:** https://arxiv.org/pdf/2305.04332 - **Abstract** This paper shows the machine learning system which performs instance segmentation of cytological images in veterinary medicine. Eleven cell types were used directly and indirectly in the experiments, including damaged and unrecognized categories. The deep learning models employed in the system achieve a high score of average precision and recall metrics, i.e. 0.94 and 0.8 respectively, for the selected three types of tumors. This variety of label types allowed us to draw a meaningful conclusion that there are relatively few mistakes for tumor cell types. Additionally, the model learned tumor cell features well enough to avoid misclassification mistakes of one tumor type into another. The experiments also revealed that the quality of the results improves with the dataset size (excluding the damaged cells). It is worth noting that all the experiments were done using a custom dedicated dataset provided by the cooperating vet doctors. ### Scene Text Recognition with Image-Text Matching-guided Dictionary - **Authors:** Jiajun Wei, Hongjian Zhan, Xiao Tu, Yue Lu, Umapada Pal - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04524 - **Pdf link:** https://arxiv.org/pdf/2305.04524 - **Abstract** Employing a dictionary can efficiently rectify the deviation between the visual prediction and the ground truth in scene text recognition methods. However, the independence of the dictionary on the visual features may lead to incorrect rectification of accurate visual predictions. In this paper, we propose a new dictionary language model leveraging the Scene Image-Text Matching(SITM) network, which avoids the drawbacks of the explicit dictionary language model: 1) the independence of the visual features; 2) noisy choice in candidates etc. The SITM network accomplishes this by using Image-Text Contrastive (ITC) Learning to match an image with its corresponding text among candidates in the inference stage. ITC is widely used in vision-language learning to pull the positive image-text pair closer in feature space. Inspired by ITC, the SITM network combines the visual features and the text features of all candidates to identify the candidate with the minimum distance in the feature space. Our lexicon method achieves better results(93.8\% accuracy) than the ordinary method results(92.1\% accuracy) on six mainstream benchmarks. Additionally, we integrate our method with ABINet and establish new state-of-the-art results on several benchmarks. ### High Quality Large-Scale 3-D Urban Mapping with Multi-Master TomoSAR - **Authors:** Yilei Shi, Richard Bamler, Yuanyuan Wang, Xiao Xiang Zhu - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04541 - **Pdf link:** https://arxiv.org/pdf/2305.04541 - **Abstract** Multi-baseline interferometric synthetic aperture radar (InSAR) techniques are effective approaches for retrieving the 3-D information of urban areas. In order to obtain a plausible reconstruction, it is necessary to use large-stack interferograms. Hence, these methods are commonly not appropriate for large-scale 3-D urban mapping using TanDEM-X data where only a few acquisitions are available in average for each city. This work proposes a new SAR tomographic processing framework to work with those extremely small stacks, which integrates the non-local filtering into SAR tomography inversion. The applicability of the algorithm is demonstrated using a TanDEM-X multi-baseline stack with 5 bistatic interferograms over the whole city of Munich, Germany. Systematic comparison of our result with airborne LiDAR data shows that the relative height accuracy of two third buildings is within two meters, which outperforms the TanDEM-X raw DEM. The promising performance of the proposed algorithm paved the first step towards high quality large-scale 3-D urban mapping. ### MARS: Mask Attention Refinement with Sequential Quadtree Nodes for Car Damage Instance Segmentation - **Authors:** Teerapong Panboonyuen, Naphat Nithisopa, Panin Pienroj, Laphonchai Jirachuphun, Chaiwasut Watthanasirikrit, Naruepon Pornwiriyakul - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04743 - **Pdf link:** https://arxiv.org/pdf/2305.04743 - **Abstract** Evaluating car damages from misfortune is critical to the car insurance industry. However, the accuracy is still insufficient for real-world applications since the deep learning network is not designed for car damage images as inputs, and its segmented masks are still very coarse. This paper presents MARS (Mask Attention Refinement with Sequential quadtree nodes) for car damage instance segmentation. Our MARS represents self-attention mechanisms to draw global dependencies between the sequential quadtree nodes layer and quadtree transformer to recalibrate channel weights and predict highly accurate instance masks. Our extensive experiments demonstrate that MARS outperforms state-of-the-art (SOTA) instance segmentation methods on three popular benchmarks such as Mask R-CNN [9], PointRend [14], and Mask Transfiner [13], by a large margin of +1.3 maskAP-based R50-FPN backbone and +2.3 maskAP-based R101-FPN backbone on Thai car-damage dataset. Our demos are available at https://www.marssolution.io. ### BiRT: Bio-inspired Replay in Vision Transformers for Continual Learning - **Authors:** Kishaan Jeeveswaran, Prashant Bhat, Bahram Zonooz, Elahe Arani - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2305.04769 - **Pdf link:** https://arxiv.org/pdf/2305.04769 - **Abstract** The ability of deep neural networks to continually learn and adapt to a sequence of tasks has remained challenging due to catastrophic forgetting of previously learned tasks. Humans, on the other hand, have a remarkable ability to acquire, assimilate, and transfer knowledge across tasks throughout their lifetime without catastrophic forgetting. The versatility of the brain can be attributed to the rehearsal of abstract experiences through a complementary learning system. However, representation rehearsal in vision transformers lacks diversity, resulting in overfitting and consequently, performance drops significantly compared to raw image rehearsal. Therefore, we propose BiRT, a novel representation rehearsal-based continual learning approach using vision transformers. Specifically, we introduce constructive noises at various stages of the vision transformer and enforce consistency in predictions with respect to an exponential moving average of the working model. Our method provides consistent performance gain over raw image and vanilla representation rehearsal on several challenging CL benchmarks, while being memory efficient and robust to natural and adversarial corruptions. ### PillarNeXt: Rethinking Network Designs for 3D Object Detection in LiDAR Point Clouds - **Authors:** Jinyu Li, Chenxu Luo, Xiaodong Yang - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2305.04925 - **Pdf link:** https://arxiv.org/pdf/2305.04925 - **Abstract** In order to deal with the sparse and unstructured raw point clouds, LiDAR based 3D object detection research mostly focuses on designing dedicated local point aggregators for fine-grained geometrical modeling. In this paper, we revisit the local point aggregators from the perspective of allocating computational resources. We find that the simplest pillar based models perform surprisingly well considering both accuracy and latency. Additionally, we show that minimal adaptions from the success of 2D object detection, such as enlarging receptive field, significantly boost the performance. Extensive experiments reveal that our pillar based networks with modernized designs in terms of architecture and training render the state-of-the-art performance on the two popular benchmarks: Waymo Open Dataset and nuScenes. Our results challenge the common intuition that the detailed geometry modeling is essential to achieve high performance for 3D object detection. ## Keyword: raw image ### BiRT: Bio-inspired Replay in Vision Transformers for Continual Learning - **Authors:** Kishaan Jeeveswaran, Prashant Bhat, Bahram Zonooz, Elahe Arani - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Machine Learning (cs.LG); Neural and Evolutionary Computing (cs.NE) - **Arxiv link:** https://arxiv.org/abs/2305.04769 - **Pdf link:** https://arxiv.org/pdf/2305.04769 - **Abstract** The ability of deep neural networks to continually learn and adapt to a sequence of tasks has remained challenging due to catastrophic forgetting of previously learned tasks. Humans, on the other hand, have a remarkable ability to acquire, assimilate, and transfer knowledge across tasks throughout their lifetime without catastrophic forgetting. The versatility of the brain can be attributed to the rehearsal of abstract experiences through a complementary learning system. However, representation rehearsal in vision transformers lacks diversity, resulting in overfitting and consequently, performance drops significantly compared to raw image rehearsal. Therefore, we propose BiRT, a novel representation rehearsal-based continual learning approach using vision transformers. Specifically, we introduce constructive noises at various stages of the vision transformer and enforce consistency in predictions with respect to an exponential moving average of the working model. Our method provides consistent performance gain over raw image and vanilla representation rehearsal on several challenging CL benchmarks, while being memory efficient and robust to natural and adversarial corruptions.
process
new submissions for tue may keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast structural and statistical texture knowledge distillation for semantic segmentation authors deyi ji haoran wang mingyuan tao jianqiang huang xian sheng hua hongtao lu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract existing knowledge distillation works for semantic segmentation mainly focus on transferring high level contextual knowledge from teacher to student however low level texture knowledge is also of vital importance for characterizing the local structural pattern and global statistical property such as boundary smoothness regularity and color contrast which may not be well addressed by high level deep features in this paper we are intended to take full advantage of both structural and statistical texture knowledge and propose a novel structural and statistical texture knowledge distillation sstkd framework for semantic segmentation specifically for structural texture knowledge we introduce a contourlet decomposition module cdm that decomposes low level features with iterative laplacian pyramid and directional filter bank to mine the structural texture knowledge for statistical knowledge we propose a denoised texture intensity equalization module dtiem to adaptively extract and enhance statistical texture knowledge through heuristics iterative quantization and denoised operation finally each knowledge learning is supervised by an individual loss function forcing the student network to mimic the teacher better from a broader perspective experiments show that the proposed method achieves state of the art performance on cityscapes pascal voc and datasets keyword awb nl cs net deep learning with non local prior for image compressive sensing authors shuai bian shouliang qi chen li yudong yao yueyang teng subjects computer vision and pattern recognition cs cv machine learning cs lg image and video processing eess iv arxiv link pdf link abstract deep learning has been applied to compressive sensing cs of images successfully in recent years however existing network based methods are often trained as the black box in which the lack of prior knowledge is often the bottleneck for further performance improvement to overcome this drawback this paper proposes a novel cs method using non local prior which combines the interpretability of the traditional optimization methods with the speed of network based methods called nl cs net we unroll each phase from iteration of the augmented lagrangian method solving non local and sparse regularized optimization problem by a network nl cs net is composed of the up sampling module and the recovery module in the up sampling module we use learnable up sampling matrix instead of a predefined one in the recovery module patch wise non local network is employed to capture long range feature correspondences important parameters involved e g sampling matrix nonlinear transforms shrinkage thresholds step size etc are learned end to end rather than hand crafted furthermore to facilitate practical implementation orthogonal and binary constraints on the sampling matrix are simultaneously adopted extensive experiments on natural images and magnetic resonance imaging mri demonstrate that the proposed method outperforms the state of the art methods while maintaining great interpretability and speed scene text recognition with image text matching guided dictionary authors jiajun wei hongjian zhan xiao tu yue lu umapada pal subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract employing a dictionary can efficiently rectify the deviation between the visual prediction and the ground truth in scene text recognition methods however the independence of the dictionary on the visual features may lead to incorrect rectification of accurate visual predictions in this paper we propose a new dictionary language model leveraging the scene image text matching sitm network which avoids the drawbacks of the explicit dictionary language model the independence of the visual features noisy choice in candidates etc the sitm network accomplishes this by using image text contrastive itc learning to match an image with its corresponding text among candidates in the inference stage itc is widely used in vision language learning to pull the positive image text pair closer in feature space inspired by itc the sitm network combines the visual features and the text features of all candidates to identify the candidate with the minimum distance in the feature space our lexicon method achieves better results accuracy than the ordinary method results accuracy on six mainstream benchmarks additionally we integrate our method with abinet and establish new state of the art results on several benchmarks keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression yolocs object detection based on dense channel compression for feature spatial solidification authors lin huang weisheng li linlin shen haojie fu xue xiao suihan xiao subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in this study we examine the associations between channel features and convolutional kernels during the processes of feature purification and gradient backpropagation with a focus on the forward and backward propagation within the network consequently we propose a method called dense channel compression for feature spatial solidification drawing upon the central concept of this method we introduce two innovative modules for backbone and head networks the dense channel compression for feature spatial solidification structure dcfs and the asymmetric multi level compression decoupled head adh when integrated into the model these two modules demonstrate exceptional performance resulting in a modified model referred to as yolocs evaluated on the mscoco dataset the large medium and small yolocs models yield ap of and respectively maintaining inference speeds remarkably similar to those of the model the large medium and small yolocs models surpass the model s ap by and respectively adversarial examples detection with enhanced image difference features based on local histogram equalization authors zhaoxia yin shaowei zhu hang su jianteng peng wanli lyu bin luo subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract deep neural networks dnns have recently made significant progress in many fields however studies have shown that dnns are vulnerable to adversarial examples where imperceptible perturbations can greatly mislead dnns even if the full underlying model parameters are not accessible various defense methods have been proposed such as feature compression and gradient masking however numerous studies have proven that previous methods create detection or defense against certain attacks which renders the method ineffective in the face of the latest unknown attack methods the invisibility of adversarial perturbations is one of the evaluation indicators for adversarial example attacks which also means that the difference in the local correlation of high frequency information in adversarial examples and normal examples can be used as an effective feature to distinguish the two therefore we propose an adversarial example detection framework based on a high frequency information enhancement strategy which can effectively extract and amplify the feature differences between adversarial examples and normal examples experimental results show that the feature augmentation module can be combined with existing detection models in a modular way under this framework improve the detector s performance and reduce the deployment cost without modifying the existing detection model snt sharpness minimizing network transformation for fast compression friendly pretraining authors jung hwan heo seyedarmin azizi arash fayyazi mahdi nazemi massoud pedram subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract model compression has become the de facto approach for optimizing the efficiency of vision models recently the focus of most compression efforts has shifted to post training scenarios due to the very high cost of large scale pretraining this has created the need to build compressible models from scratch which can effectively be compressed after training in this work we present a sharpness minimizing network transformation snt method applied during pretraining that can create models with desirable compressibility and generalizability features we compare our approach to a well known sharpness minimizing optimizer to validate its efficacy in creating a flat loss landscape to the best of our knowledge snt is the first pretraining method that uses an architectural transformation to generate compression friendly networks we find that snt generalizes across different compression tasks and network backbones delivering consistent improvements over the adam baseline with up to accuracy improvement on weight pruning and accuracy improvement on quantization code to reproduce our results will be made publicly available keyword raw nl cs net deep learning with non local prior for image compressive sensing authors shuai bian shouliang qi chen li yudong yao yueyang teng subjects computer vision and pattern recognition cs cv machine learning cs lg image and video processing eess iv arxiv link pdf link abstract deep learning has been applied to compressive sensing cs of images successfully in recent years however existing network based methods are often trained as the black box in which the lack of prior knowledge is often the bottleneck for further performance improvement to overcome this drawback this paper proposes a novel cs method using non local prior which combines the interpretability of the traditional optimization methods with the speed of network based methods called nl cs net we unroll each phase from iteration of the augmented lagrangian method solving non local and sparse regularized optimization problem by a network nl cs net is composed of the up sampling module and the recovery module in the up sampling module we use learnable up sampling matrix instead of a predefined one in the recovery module patch wise non local network is employed to capture long range feature correspondences important parameters involved e g sampling matrix nonlinear transforms shrinkage thresholds step size etc are learned end to end rather than hand crafted furthermore to facilitate practical implementation orthogonal and binary constraints on the sampling matrix are simultaneously adopted extensive experiments on natural images and magnetic resonance imaging mri demonstrate that the proposed method outperforms the state of the art methods while maintaining great interpretability and speed prompt what you need enhancing segmentation in rainy scenes with anchor based prompting authors xiaoyu guo xiang wei qi su huiqin zhao shunli zhan subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract semantic segmentation in rainy scenes is a challenging task due to the complex environment class distribution imbalance and limited annotated data to address these challenges we propose a novel framework that utilizes semi supervised learning and pre trained segmentation foundation model to achieve superior performance specifically our framework leverages the semi supervised model as the basis for generating raw semantic segmentation results while also serving as a guiding force to prompt pre trained foundation model to compensate for knowledge gaps with entropy based anchors in addition to minimize the impact of irrelevant segmentation masks generated by the pre trained foundation model we also propose a mask filtering and fusion mechanism that optimizes raw semantic segmentation results based on the principle of minimum risk the proposed framework achieves superior segmentation performance on the rainy wcity dataset and is awarded the first prize in the sub track of strain in icme grand challenges yolocs object detection based on dense channel compression for feature spatial solidification authors lin huang weisheng li linlin shen haojie fu xue xiao suihan xiao subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in this study we examine the associations between channel features and convolutional kernels during the processes of feature purification and gradient backpropagation with a focus on the forward and backward propagation within the network consequently we propose a method called dense channel compression for feature spatial solidification drawing upon the central concept of this method we introduce two innovative modules for backbone and head networks the dense channel compression for feature spatial solidification structure dcfs and the asymmetric multi level compression decoupled head adh when integrated into the model these two modules demonstrate exceptional performance resulting in a modified model referred to as yolocs evaluated on the mscoco dataset the large medium and small yolocs models yield ap of and respectively maintaining inference speeds remarkably similar to those of the model the large medium and small yolocs models surpass the model s ap by and respectively segmentation of the veterinary cytological images for fast neoplastic tumors diagnosis authors jakub grzeszczyk michał karwatowski daria łukasik maciej wielgosz paweł russek szymon mazurek jakub caputa rafał frączek anna śmiech ernest jamro sebastian koryciak agnieszka dąbrowska boruch marcin pietroń kazimierz wiatr subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract this paper shows the machine learning system which performs instance segmentation of cytological images in veterinary medicine eleven cell types were used directly and indirectly in the experiments including damaged and unrecognized categories the deep learning models employed in the system achieve a high score of average precision and recall metrics i e and respectively for the selected three types of tumors this variety of label types allowed us to draw a meaningful conclusion that there are relatively few mistakes for tumor cell types additionally the model learned tumor cell features well enough to avoid misclassification mistakes of one tumor type into another the experiments also revealed that the quality of the results improves with the dataset size excluding the damaged cells it is worth noting that all the experiments were done using a custom dedicated dataset provided by the cooperating vet doctors scene text recognition with image text matching guided dictionary authors jiajun wei hongjian zhan xiao tu yue lu umapada pal subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract employing a dictionary can efficiently rectify the deviation between the visual prediction and the ground truth in scene text recognition methods however the independence of the dictionary on the visual features may lead to incorrect rectification of accurate visual predictions in this paper we propose a new dictionary language model leveraging the scene image text matching sitm network which avoids the drawbacks of the explicit dictionary language model the independence of the visual features noisy choice in candidates etc the sitm network accomplishes this by using image text contrastive itc learning to match an image with its corresponding text among candidates in the inference stage itc is widely used in vision language learning to pull the positive image text pair closer in feature space inspired by itc the sitm network combines the visual features and the text features of all candidates to identify the candidate with the minimum distance in the feature space our lexicon method achieves better results accuracy than the ordinary method results accuracy on six mainstream benchmarks additionally we integrate our method with abinet and establish new state of the art results on several benchmarks high quality large scale d urban mapping with multi master tomosar authors yilei shi richard bamler yuanyuan wang xiao xiang zhu subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract multi baseline interferometric synthetic aperture radar insar techniques are effective approaches for retrieving the d information of urban areas in order to obtain a plausible reconstruction it is necessary to use large stack interferograms hence these methods are commonly not appropriate for large scale d urban mapping using tandem x data where only a few acquisitions are available in average for each city this work proposes a new sar tomographic processing framework to work with those extremely small stacks which integrates the non local filtering into sar tomography inversion the applicability of the algorithm is demonstrated using a tandem x multi baseline stack with bistatic interferograms over the whole city of munich germany systematic comparison of our result with airborne lidar data shows that the relative height accuracy of two third buildings is within two meters which outperforms the tandem x raw dem the promising performance of the proposed algorithm paved the first step towards high quality large scale d urban mapping mars mask attention refinement with sequential quadtree nodes for car damage instance segmentation authors teerapong panboonyuen naphat nithisopa panin pienroj laphonchai jirachuphun chaiwasut watthanasirikrit naruepon pornwiriyakul subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract evaluating car damages from misfortune is critical to the car insurance industry however the accuracy is still insufficient for real world applications since the deep learning network is not designed for car damage images as inputs and its segmented masks are still very coarse this paper presents mars mask attention refinement with sequential quadtree nodes for car damage instance segmentation our mars represents self attention mechanisms to draw global dependencies between the sequential quadtree nodes layer and quadtree transformer to recalibrate channel weights and predict highly accurate instance masks our extensive experiments demonstrate that mars outperforms state of the art sota instance segmentation methods on three popular benchmarks such as mask r cnn pointrend and mask transfiner by a large margin of maskap based fpn backbone and maskap based fpn backbone on thai car damage dataset our demos are available at birt bio inspired replay in vision transformers for continual learning authors kishaan jeeveswaran prashant bhat bahram zonooz elahe arani subjects computer vision and pattern recognition cs cv machine learning cs lg neural and evolutionary computing cs ne arxiv link pdf link abstract the ability of deep neural networks to continually learn and adapt to a sequence of tasks has remained challenging due to catastrophic forgetting of previously learned tasks humans on the other hand have a remarkable ability to acquire assimilate and transfer knowledge across tasks throughout their lifetime without catastrophic forgetting the versatility of the brain can be attributed to the rehearsal of abstract experiences through a complementary learning system however representation rehearsal in vision transformers lacks diversity resulting in overfitting and consequently performance drops significantly compared to raw image rehearsal therefore we propose birt a novel representation rehearsal based continual learning approach using vision transformers specifically we introduce constructive noises at various stages of the vision transformer and enforce consistency in predictions with respect to an exponential moving average of the working model our method provides consistent performance gain over raw image and vanilla representation rehearsal on several challenging cl benchmarks while being memory efficient and robust to natural and adversarial corruptions pillarnext rethinking network designs for object detection in lidar point clouds authors jinyu li chenxu luo xiaodong yang subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract in order to deal with the sparse and unstructured raw point clouds lidar based object detection research mostly focuses on designing dedicated local point aggregators for fine grained geometrical modeling in this paper we revisit the local point aggregators from the perspective of allocating computational resources we find that the simplest pillar based models perform surprisingly well considering both accuracy and latency additionally we show that minimal adaptions from the success of object detection such as enlarging receptive field significantly boost the performance extensive experiments reveal that our pillar based networks with modernized designs in terms of architecture and training render the state of the art performance on the two popular benchmarks waymo open dataset and nuscenes our results challenge the common intuition that the detailed geometry modeling is essential to achieve high performance for object detection keyword raw image birt bio inspired replay in vision transformers for continual learning authors kishaan jeeveswaran prashant bhat bahram zonooz elahe arani subjects computer vision and pattern recognition cs cv machine learning cs lg neural and evolutionary computing cs ne arxiv link pdf link abstract the ability of deep neural networks to continually learn and adapt to a sequence of tasks has remained challenging due to catastrophic forgetting of previously learned tasks humans on the other hand have a remarkable ability to acquire assimilate and transfer knowledge across tasks throughout their lifetime without catastrophic forgetting the versatility of the brain can be attributed to the rehearsal of abstract experiences through a complementary learning system however representation rehearsal in vision transformers lacks diversity resulting in overfitting and consequently performance drops significantly compared to raw image rehearsal therefore we propose birt a novel representation rehearsal based continual learning approach using vision transformers specifically we introduce constructive noises at various stages of the vision transformer and enforce consistency in predictions with respect to an exponential moving average of the working model our method provides consistent performance gain over raw image and vanilla representation rehearsal on several challenging cl benchmarks while being memory efficient and robust to natural and adversarial corruptions
1
3,270
6,344,588,933
IssuesEvent
2017-07-27 20:16:37
ncbo/bioportal-project
https://api.github.com/repos/ncbo/bioportal-project
opened
LCMPT: classes page returns "Not Found"
ontology processing problem
LCMPT, a SKOS ontology appears to have parsed correctly, but when navigating to the classes page: http://bioportal.bioontology.org/ontologies/LCMPT?p=classes BioPortal returns "The page you are looking for wasn't found. Please try again."
1.0
LCMPT: classes page returns "Not Found" - LCMPT, a SKOS ontology appears to have parsed correctly, but when navigating to the classes page: http://bioportal.bioontology.org/ontologies/LCMPT?p=classes BioPortal returns "The page you are looking for wasn't found. Please try again."
process
lcmpt classes page returns not found lcmpt a skos ontology appears to have parsed correctly but when navigating to the classes page bioportal returns the page you are looking for wasn t found please try again
1