Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 4
112
| repo_url
stringlengths 33
141
| action
stringclasses 3
values | title
stringlengths 1
1.02k
| labels
stringlengths 4
1.54k
| body
stringlengths 1
262k
| index
stringclasses 17
values | text_combine
stringlengths 95
262k
| label
stringclasses 2
values | text
stringlengths 96
252k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
145,143
| 22,613,887,568
|
IssuesEvent
|
2022-06-29 19:50:02
|
cityofaustin/atd-data-tech
|
https://api.github.com/repos/cityofaustin/atd-data-tech
|
opened
|
Create master list of possible MPV features
|
Workgroup: DTS Service: Product Need: 1-Must Have Product: Moped Type: Design
|
Based on the Design Discovery research (#9510) conducted in late June 2022, Rebecca will create a universal list of potential features for the Mobility Project Viewer (#1106).
|
1.0
|
Create master list of possible MPV features - Based on the Design Discovery research (#9510) conducted in late June 2022, Rebecca will create a universal list of potential features for the Mobility Project Viewer (#1106).
|
non_test
|
create master list of possible mpv features based on the design discovery research conducted in late june rebecca will create a universal list of potential features for the mobility project viewer
| 0
|
826,565
| 31,654,088,884
|
IssuesEvent
|
2023-09-07 02:30:18
|
craftercms/craftercms
|
https://api.github.com/repos/craftercms/craftercms
|
closed
|
[deployer] Deployer throws NPE on startup with Springframework log level `info`
|
bug priority: medium CI validate
|
### Duplicates
- [X] I have searched the existing issues
### Latest version
- [ ] The issue is in the latest released 4.2.x
- [X] The issue is in the latest released 4.1.x
- [ ] The issue is in the latest released 4.0.x
- [ ] The issue is in the latest released 3.1.x
### Describe the issue
Deployer throws NPE on startup with spring log level info
### Steps to reproduce
Steps:
1. Set log level in `crafter-authoring/bin/crafter-deployer/logging.xml` for spring to info:
```
<Logger name="org.springframework" level="info"/>
```
2. Start authoring
3. See the error
### Relevant log output
```shell
2023-08-18 21:43:34.583 INFO 21 --- [main] ctory.support.DefaultListableBeanFactory : FactoryBean threw exception from getObjectType, despite the contract saying that it should return null if the type of its object cannot be determined yet
java.lang.NullPointerException: Cannot invoke "Object.getClass()" because "this.actualBean" is null
at org.craftercms.commons.spring.ConditionalFactoryBean.getObjectType(ConditionalFactoryBean.java:52)
at org.springframework.beans.factory.support.FactoryBeanRegistrySupport.getTypeForFactoryBean(FactoryBeanRegistrySupport.java:64)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.getTypeForFactoryBean(AbstractAutowireCapableBeanFactory.java:911)
at org.springframework.beans.factory.support.AbstractBeanFactory.isTypeMatch(AbstractBeanFactory.java:637)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doGetBeanNamesForType(DefaultListableBeanFactory.java:583)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.getBeanNamesForType(DefaultListableBeanFactory.java:550)
at org.springframework.beans.factory.BeanFactoryUtils.beanNamesForTypeIncludingAncestors(BeanFactoryUtils.java:265)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1557)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1354)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:229)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1372)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1222)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:955)
at org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:920)
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:583)
at org.craftercms.deployer.impl.TargetServiceImpl.loadApplicationContext(TargetServiceImpl.java:468)
at org.craftercms.deployer.impl.TargetServiceImpl.buildTarget(TargetServiceImpl.java:365)
at org.craftercms.deployer.impl.TargetServiceImpl.loadTarget(TargetServiceImpl.java:378)
at org.craftercms.deployer.impl.TargetServiceImpl.resolveTargetFromConfigFile(TargetServiceImpl.java:348)
at org.craftercms.deployer.impl.TargetServiceImpl.resolveTargets(TargetServiceImpl.java:221)
at org.craftercms.deployer.impl.TargetServiceImpl.onApplicationEvent(TargetServiceImpl.java:166)
at org.craftercms.deployer.impl.TargetServiceImpl.onApplicationEvent(TargetServiceImpl.java:81)
at org.springframework.context.event.SimpleApplicationEventMulticaster.doInvokeListener(SimpleApplicationEventMulticaster.java:176)
at org.springframework.context.event.SimpleApplicationEventMulticaster.invokeListener(SimpleApplicationEventMulticaster.java:169)
at org.springframework.context.event.SimpleApplicationEventMulticaster.multicastEvent(SimpleApplicationEventMulticaster.java:143)
at org.springframework.context.support.AbstractApplicationContext.publishEvent(AbstractApplicationContext.java:421)
at org.springframework.context.support.AbstractApplicationContext.publishEvent(AbstractApplicationContext.java:378)
at org.springframework.boot.context.event.EventPublishingRunListener.ready(EventPublishingRunListener.java:114)
at org.springframework.boot.SpringApplicationRunListeners.lambda$ready$6(SpringApplicationRunListeners.java:82)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1511)
at org.springframework.boot.SpringApplicationRunListeners.doWithListeners(SpringApplicationRunListeners.java:120)
at org.springframework.boot.SpringApplicationRunListeners.doWithListeners(SpringApplicationRunListeners.java:114)
at org.springframework.boot.SpringApplicationRunListeners.ready(SpringApplicationRunListeners.java:82)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:322)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1303)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1292)
at org.craftercms.deployer.DeployerApplication.main(DeployerApplication.java:113)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58)
at org.springframework.boot.loader.PropertiesLauncher.main(PropertiesLauncher.java:467)
```
### Screenshots and/or videos
_No response_
|
1.0
|
[deployer] Deployer throws NPE on startup with Springframework log level `info` - ### Duplicates
- [X] I have searched the existing issues
### Latest version
- [ ] The issue is in the latest released 4.2.x
- [X] The issue is in the latest released 4.1.x
- [ ] The issue is in the latest released 4.0.x
- [ ] The issue is in the latest released 3.1.x
### Describe the issue
Deployer throws NPE on startup with spring log level info
### Steps to reproduce
Steps:
1. Set log level in `crafter-authoring/bin/crafter-deployer/logging.xml` for spring to info:
```
<Logger name="org.springframework" level="info"/>
```
2. Start authoring
3. See the error
### Relevant log output
```shell
2023-08-18 21:43:34.583 INFO 21 --- [main] ctory.support.DefaultListableBeanFactory : FactoryBean threw exception from getObjectType, despite the contract saying that it should return null if the type of its object cannot be determined yet
java.lang.NullPointerException: Cannot invoke "Object.getClass()" because "this.actualBean" is null
at org.craftercms.commons.spring.ConditionalFactoryBean.getObjectType(ConditionalFactoryBean.java:52)
at org.springframework.beans.factory.support.FactoryBeanRegistrySupport.getTypeForFactoryBean(FactoryBeanRegistrySupport.java:64)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.getTypeForFactoryBean(AbstractAutowireCapableBeanFactory.java:911)
at org.springframework.beans.factory.support.AbstractBeanFactory.isTypeMatch(AbstractBeanFactory.java:637)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doGetBeanNamesForType(DefaultListableBeanFactory.java:583)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.getBeanNamesForType(DefaultListableBeanFactory.java:550)
at org.springframework.beans.factory.BeanFactoryUtils.beanNamesForTypeIncludingAncestors(BeanFactoryUtils.java:265)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.findAutowireCandidates(DefaultListableBeanFactory.java:1557)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doResolveDependency(DefaultListableBeanFactory.java:1354)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.resolveDependency(DefaultListableBeanFactory.java:1311)
at org.springframework.beans.factory.support.ConstructorResolver.resolveAutowiredArgument(ConstructorResolver.java:887)
at org.springframework.beans.factory.support.ConstructorResolver.createArgumentArray(ConstructorResolver.java:791)
at org.springframework.beans.factory.support.ConstructorResolver.autowireConstructor(ConstructorResolver.java:229)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.autowireConstructor(AbstractAutowireCapableBeanFactory.java:1372)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBeanInstance(AbstractAutowireCapableBeanFactory.java:1222)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.doCreateBean(AbstractAutowireCapableBeanFactory.java:582)
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.createBean(AbstractAutowireCapableBeanFactory.java:542)
at org.springframework.beans.factory.support.AbstractBeanFactory.lambda$doGetBean$0(AbstractBeanFactory.java:335)
at org.springframework.beans.factory.support.DefaultSingletonBeanRegistry.getSingleton(DefaultSingletonBeanRegistry.java:234)
at org.springframework.beans.factory.support.AbstractBeanFactory.doGetBean(AbstractBeanFactory.java:333)
at org.springframework.beans.factory.support.AbstractBeanFactory.getBean(AbstractBeanFactory.java:208)
at org.springframework.beans.factory.support.DefaultListableBeanFactory.preInstantiateSingletons(DefaultListableBeanFactory.java:955)
at org.springframework.context.support.AbstractApplicationContext.finishBeanFactoryInitialization(AbstractApplicationContext.java:920)
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:583)
at org.craftercms.deployer.impl.TargetServiceImpl.loadApplicationContext(TargetServiceImpl.java:468)
at org.craftercms.deployer.impl.TargetServiceImpl.buildTarget(TargetServiceImpl.java:365)
at org.craftercms.deployer.impl.TargetServiceImpl.loadTarget(TargetServiceImpl.java:378)
at org.craftercms.deployer.impl.TargetServiceImpl.resolveTargetFromConfigFile(TargetServiceImpl.java:348)
at org.craftercms.deployer.impl.TargetServiceImpl.resolveTargets(TargetServiceImpl.java:221)
at org.craftercms.deployer.impl.TargetServiceImpl.onApplicationEvent(TargetServiceImpl.java:166)
at org.craftercms.deployer.impl.TargetServiceImpl.onApplicationEvent(TargetServiceImpl.java:81)
at org.springframework.context.event.SimpleApplicationEventMulticaster.doInvokeListener(SimpleApplicationEventMulticaster.java:176)
at org.springframework.context.event.SimpleApplicationEventMulticaster.invokeListener(SimpleApplicationEventMulticaster.java:169)
at org.springframework.context.event.SimpleApplicationEventMulticaster.multicastEvent(SimpleApplicationEventMulticaster.java:143)
at org.springframework.context.support.AbstractApplicationContext.publishEvent(AbstractApplicationContext.java:421)
at org.springframework.context.support.AbstractApplicationContext.publishEvent(AbstractApplicationContext.java:378)
at org.springframework.boot.context.event.EventPublishingRunListener.ready(EventPublishingRunListener.java:114)
at org.springframework.boot.SpringApplicationRunListeners.lambda$ready$6(SpringApplicationRunListeners.java:82)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1511)
at org.springframework.boot.SpringApplicationRunListeners.doWithListeners(SpringApplicationRunListeners.java:120)
at org.springframework.boot.SpringApplicationRunListeners.doWithListeners(SpringApplicationRunListeners.java:114)
at org.springframework.boot.SpringApplicationRunListeners.ready(SpringApplicationRunListeners.java:82)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:322)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1303)
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1292)
at org.craftercms.deployer.DeployerApplication.main(DeployerApplication.java:113)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.base/java.lang.reflect.Method.invoke(Method.java:568)
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108)
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58)
at org.springframework.boot.loader.PropertiesLauncher.main(PropertiesLauncher.java:467)
```
### Screenshots and/or videos
_No response_
|
non_test
|
deployer throws npe on startup with springframework log level info duplicates i have searched the existing issues latest version the issue is in the latest released x the issue is in the latest released x the issue is in the latest released x the issue is in the latest released x describe the issue deployer throws npe on startup with spring log level info steps to reproduce steps set log level in crafter authoring bin crafter deployer logging xml for spring to info start authoring see the error relevant log output shell info ctory support defaultlistablebeanfactory factorybean threw exception from getobjecttype despite the contract saying that it should return null if the type of its object cannot be determined yet java lang nullpointerexception cannot invoke object getclass because this actualbean is null at org craftercms commons spring conditionalfactorybean getobjecttype conditionalfactorybean java at org springframework beans factory support factorybeanregistrysupport gettypeforfactorybean factorybeanregistrysupport java at org springframework beans factory support abstractautowirecapablebeanfactory gettypeforfactorybean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory istypematch abstractbeanfactory java at org springframework beans factory support defaultlistablebeanfactory dogetbeannamesfortype defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory getbeannamesfortype defaultlistablebeanfactory java at org springframework beans factory beanfactoryutils beannamesfortypeincludingancestors beanfactoryutils java at org springframework beans factory support defaultlistablebeanfactory findautowirecandidates defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory doresolvedependency defaultlistablebeanfactory java at org springframework beans factory support defaultlistablebeanfactory resolvedependency defaultlistablebeanfactory java at org springframework beans factory support constructorresolver resolveautowiredargument constructorresolver java at org springframework beans factory support constructorresolver createargumentarray constructorresolver java at org springframework beans factory support constructorresolver autowireconstructor constructorresolver java at org springframework beans factory support abstractautowirecapablebeanfactory autowireconstructor abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbeaninstance abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory docreatebean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractautowirecapablebeanfactory createbean abstractautowirecapablebeanfactory java at org springframework beans factory support abstractbeanfactory lambda dogetbean abstractbeanfactory java at org springframework beans factory support defaultsingletonbeanregistry getsingleton defaultsingletonbeanregistry java at org springframework beans factory support abstractbeanfactory dogetbean abstractbeanfactory java at org springframework beans factory support abstractbeanfactory getbean abstractbeanfactory java at org springframework beans factory support defaultlistablebeanfactory preinstantiatesingletons defaultlistablebeanfactory java at org springframework context support abstractapplicationcontext finishbeanfactoryinitialization abstractapplicationcontext java at org springframework context support abstractapplicationcontext refresh abstractapplicationcontext java at org craftercms deployer impl targetserviceimpl loadapplicationcontext targetserviceimpl java at org craftercms deployer impl targetserviceimpl buildtarget targetserviceimpl java at org craftercms deployer impl targetserviceimpl loadtarget targetserviceimpl java at org craftercms deployer impl targetserviceimpl resolvetargetfromconfigfile targetserviceimpl java at org craftercms deployer impl targetserviceimpl resolvetargets targetserviceimpl java at org craftercms deployer impl targetserviceimpl onapplicationevent targetserviceimpl java at org craftercms deployer impl targetserviceimpl onapplicationevent targetserviceimpl java at org springframework context event simpleapplicationeventmulticaster doinvokelistener simpleapplicationeventmulticaster java at org springframework context event simpleapplicationeventmulticaster invokelistener simpleapplicationeventmulticaster java at org springframework context event simpleapplicationeventmulticaster multicastevent simpleapplicationeventmulticaster java at org springframework context support abstractapplicationcontext publishevent abstractapplicationcontext java at org springframework context support abstractapplicationcontext publishevent abstractapplicationcontext java at org springframework boot context event eventpublishingrunlistener ready eventpublishingrunlistener java at org springframework boot springapplicationrunlisteners lambda ready springapplicationrunlisteners java at java base java util arraylist foreach arraylist java at org springframework boot springapplicationrunlisteners dowithlisteners springapplicationrunlisteners java at org springframework boot springapplicationrunlisteners dowithlisteners springapplicationrunlisteners java at org springframework boot springapplicationrunlisteners ready springapplicationrunlisteners java at org springframework boot springapplication run springapplication java at org springframework boot springapplication run springapplication java at org springframework boot springapplication run springapplication java at org craftercms deployer deployerapplication main deployerapplication java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at org springframework boot loader mainmethodrunner run mainmethodrunner java at org springframework boot loader launcher launch launcher java at org springframework boot loader launcher launch launcher java at org springframework boot loader propertieslauncher main propertieslauncher java screenshots and or videos no response
| 0
|
164,285
| 12,795,627,802
|
IssuesEvent
|
2020-07-02 09:03:59
|
ComputationalRadiationPhysics/picongpu
|
https://api.github.com/repos/ComputationalRadiationPhysics/picongpu
|
opened
|
PML with uneven domains is not output and check pointed correctly
|
affects latest release bug component: plugin
|
The logic of writing and reading of PML fields relies on the domains being of equal size. It equally concerns output and checkpointing.
Here is the code that assumed it: [1](https://github.com/ComputationalRadiationPhysics/picongpu/blob/fe322db49e69b09258691c4aa781769470fe693d/include/picongpu/plugins/hdf5/writer/Field.hpp#L119-L142), [2](https://github.com/ComputationalRadiationPhysics/picongpu/blob/fe322db49e69b09258691c4aa781769470fe693d/include/picongpu/plugins/hdf5/restart/RestartFieldLoader.hpp#L93-L115), and similar for the other two output plugins. The error is that while the allocated-for PML thickness is indeed the same (even for internal domains), that does not imply the buffer size is the same, as it also depends on the local domain size.
This concerns input/output only, not the correctness of the computational part.
The incorrect behavior discovered by @Anton-Le
|
1.0
|
PML with uneven domains is not output and check pointed correctly - The logic of writing and reading of PML fields relies on the domains being of equal size. It equally concerns output and checkpointing.
Here is the code that assumed it: [1](https://github.com/ComputationalRadiationPhysics/picongpu/blob/fe322db49e69b09258691c4aa781769470fe693d/include/picongpu/plugins/hdf5/writer/Field.hpp#L119-L142), [2](https://github.com/ComputationalRadiationPhysics/picongpu/blob/fe322db49e69b09258691c4aa781769470fe693d/include/picongpu/plugins/hdf5/restart/RestartFieldLoader.hpp#L93-L115), and similar for the other two output plugins. The error is that while the allocated-for PML thickness is indeed the same (even for internal domains), that does not imply the buffer size is the same, as it also depends on the local domain size.
This concerns input/output only, not the correctness of the computational part.
The incorrect behavior discovered by @Anton-Le
|
test
|
pml with uneven domains is not output and check pointed correctly the logic of writing and reading of pml fields relies on the domains being of equal size it equally concerns output and checkpointing here is the code that assumed it and similar for the other two output plugins the error is that while the allocated for pml thickness is indeed the same even for internal domains that does not imply the buffer size is the same as it also depends on the local domain size this concerns input output only not the correctness of the computational part the incorrect behavior discovered by anton le
| 1
|
208,554
| 15,894,514,477
|
IssuesEvent
|
2021-04-11 10:26:58
|
tech256/jobs
|
https://api.github.com/repos/tech256/jobs
|
closed
|
Intermediate Program Analysts (#1620600)
|
Active Clearance Required Hiring Testing/Quality Assurance stale
|
Team PeopleTec is currently seeking multiple Intermediate Program Analysts to support our efforts on MDA's TEAMS NEXT contract.
Required Skills/Experience:
Capable of providing expertise to relevant program analytical principles and practices for developmental and operational programs.
Capable of providing planning, monitoring, acquisition, and implementation support for complex systems.
Knowledge, skills, and abilities in the analysis field.
Capable of utilizing, adapting and developing analytic techniques and processes for both technical and administrative support activities.
Works as a member of team and assists with solving complex problems involving test planning, integration, execution, and analysis.
Establishes priorities and manages planning and readiness activities of MDS test events IAW approved Test issuances (e.g., CONOPS).
Participate in concept planning; design and development; integration, testing, and verification at the MDS system, element or component level. Routinely resolves significant technical matters.
Must be a U.S. Citizen
At minimum, an active DoD Secret clearance is required to perform this work. Candidates are required to possess and maintain this level of clearance during their employment.
Education Requirements:
A BA/BS in with 7 years relevant experience is required. Three years of additional relevant experience may be substituted for the BA/BS degree requirement.
Desired Skills/Experience:
Experience with DoD testing
Experience with DoD test execution
#cjpost #dpost
EOE/Minorities/Females/Veterans/Disabled
PeopleTec, Inc. is an Equal Employment Opportunity employer and provides reasonable accommodation for qualified individuals with disabilities and disabled veterans in its job application procedures. If you have any difficulty using our online system and you need an accommodation due to a disability, you may use the following email address, HR@peopletec.com and/or phone number (256.319.3800) to contact us about your interest in employment with PeopleTec, Inc.
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, age, genetic information, citizenship, ancestry, marital status, protected veteran status, disability status or any other status protected by federal, state, or local law. PeopleTec, Inc. participates in E-Verify
For more information, or to apply now, you must go to the website below. Please DO NOT email your resume to us as we only accept applications through our website.
https://www.applicantpro.com/openings/peopletec/jobs/1620600-424912
|
1.0
|
Intermediate Program Analysts (#1620600) - Team PeopleTec is currently seeking multiple Intermediate Program Analysts to support our efforts on MDA's TEAMS NEXT contract.
Required Skills/Experience:
Capable of providing expertise to relevant program analytical principles and practices for developmental and operational programs.
Capable of providing planning, monitoring, acquisition, and implementation support for complex systems.
Knowledge, skills, and abilities in the analysis field.
Capable of utilizing, adapting and developing analytic techniques and processes for both technical and administrative support activities.
Works as a member of team and assists with solving complex problems involving test planning, integration, execution, and analysis.
Establishes priorities and manages planning and readiness activities of MDS test events IAW approved Test issuances (e.g., CONOPS).
Participate in concept planning; design and development; integration, testing, and verification at the MDS system, element or component level. Routinely resolves significant technical matters.
Must be a U.S. Citizen
At minimum, an active DoD Secret clearance is required to perform this work. Candidates are required to possess and maintain this level of clearance during their employment.
Education Requirements:
A BA/BS in with 7 years relevant experience is required. Three years of additional relevant experience may be substituted for the BA/BS degree requirement.
Desired Skills/Experience:
Experience with DoD testing
Experience with DoD test execution
#cjpost #dpost
EOE/Minorities/Females/Veterans/Disabled
PeopleTec, Inc. is an Equal Employment Opportunity employer and provides reasonable accommodation for qualified individuals with disabilities and disabled veterans in its job application procedures. If you have any difficulty using our online system and you need an accommodation due to a disability, you may use the following email address, HR@peopletec.com and/or phone number (256.319.3800) to contact us about your interest in employment with PeopleTec, Inc.
All qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, age, genetic information, citizenship, ancestry, marital status, protected veteran status, disability status or any other status protected by federal, state, or local law. PeopleTec, Inc. participates in E-Verify
For more information, or to apply now, you must go to the website below. Please DO NOT email your resume to us as we only accept applications through our website.
https://www.applicantpro.com/openings/peopletec/jobs/1620600-424912
|
test
|
intermediate program analysts team peopletec is currently seeking multiple intermediate program analysts to support our efforts on mda s teams next contract required skills experience capable of providing expertise to relevant program analytical principles and practices for developmental and operational programs capable of providing planning monitoring acquisition and implementation support for complex systems knowledge skills and abilities in the analysis field capable of utilizing adapting and developing analytic techniques and processes for both technical and administrative support activities works as a member of team and assists with solving complex problems involving test planning integration execution and analysis establishes priorities and manages planning and readiness activities of mds test events iaw approved test issuances e g conops participate in concept planning design and development integration testing and verification at the mds system element or component level routinely resolves significant technical matters must be a u s citizen at minimum an active dod secret clearance is required to perform this work candidates are required to possess and maintain this level of clearance during their employment education requirements a ba bs in with years relevant experience is required three years of additional relevant experience may be substituted for the ba bs degree requirement desired skills experience experience with dod testing experience with dod test execution cjpost dpost eoe minorities females veterans disabled peopletec inc is an equal employment opportunity employer and provides reasonable accommodation for qualified individuals with disabilities and disabled veterans in its job application procedures if you have any difficulty using our online system and you need an accommodation due to a disability you may use the following email address hr peopletec com and or phone number to contact us about your interest in employment with peopletec inc all qualified applicants will receive consideration for employment without regard to race color religion sex sexual orientation gender identity national origin age genetic information citizenship ancestry marital status protected veteran status disability status or any other status protected by federal state or local law peopletec inc participates in e verify for more information or to apply now you must go to the website below please do not email your resume to us as we only accept applications through our website
| 1
|
253,187
| 27,300,467,086
|
IssuesEvent
|
2023-02-24 01:11:30
|
panasalap/linux-4.19.72_1
|
https://api.github.com/repos/panasalap/linux-4.19.72_1
|
opened
|
CVE-2023-0179 (High) detected in linux-yoctov5.4.51
|
security vulnerability
|
## CVE-2023-0179 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nft_payload.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nft_payload.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
netfilter: nft_payload: incorrect arithmetics when fetching VLAN header bits
<p>Publish Date: 2023-01-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0179>CVE-2023-0179</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0179">https://www.linuxkernelcves.com/cves/CVE-2023-0179</a></p>
<p>Release Date: 2023-01-11</p>
<p>Fix Resolution: v5.10.164,v5.15.89,v6.1.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-0179 (High) detected in linux-yoctov5.4.51 - ## CVE-2023-0179 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nft_payload.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nft_payload.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
netfilter: nft_payload: incorrect arithmetics when fetching VLAN header bits
<p>Publish Date: 2023-01-11
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0179>CVE-2023-0179</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-0179">https://www.linuxkernelcves.com/cves/CVE-2023-0179</a></p>
<p>Release Date: 2023-01-11</p>
<p>Fix Resolution: v5.10.164,v5.15.89,v6.1.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in linux cve high severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in base branch master vulnerable source files net netfilter nft payload c net netfilter nft payload c vulnerability details netfilter nft payload incorrect arithmetics when fetching vlan header bits publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
94,252
| 8,481,826,493
|
IssuesEvent
|
2018-10-25 16:45:21
|
freedomofpress/securedrop-ux
|
https://api.github.com/repos/freedomofpress/securedrop-ux
|
closed
|
SecureDrop Client Error Messaging Inventory
|
Alpha Workstation NEED Thing we need to test with users
|
Need to know what user actions will trigger error messages, and when, in [SD Client](https://github.com/freedomofpress/securedrop-client).
Also need to know what user actions will need to be taken to rectify problems, to develop content. @eloquence volunteered to take a first stab at messaging content. @redshiftzero volunteered to create initial inventory.
Inventory must include Alpha errors, and will need to be expanded at a later date to include Beta functionality errors.
|
1.0
|
SecureDrop Client Error Messaging Inventory - Need to know what user actions will trigger error messages, and when, in [SD Client](https://github.com/freedomofpress/securedrop-client).
Also need to know what user actions will need to be taken to rectify problems, to develop content. @eloquence volunteered to take a first stab at messaging content. @redshiftzero volunteered to create initial inventory.
Inventory must include Alpha errors, and will need to be expanded at a later date to include Beta functionality errors.
|
test
|
securedrop client error messaging inventory need to know what user actions will trigger error messages and when in also need to know what user actions will need to be taken to rectify problems to develop content eloquence volunteered to take a first stab at messaging content redshiftzero volunteered to create initial inventory inventory must include alpha errors and will need to be expanded at a later date to include beta functionality errors
| 1
|
26,434
| 4,226,219,022
|
IssuesEvent
|
2016-07-02 09:42:00
|
ubiquits/ubiquits
|
https://api.github.com/repos/ubiquits/ubiquits
|
closed
|
Travis CI is picking up ts 1.8.x not @next which means there are a lot out errors output.
|
comp: testing effort1: easy (hour) priority3: required severity5: regression type: bug
|
Maybe intermittent, follow up on https://travis-ci.org/ubiquits/ubiquits/builds/141821325 when complete.
|
1.0
|
Travis CI is picking up ts 1.8.x not @next which means there are a lot out errors output. - Maybe intermittent, follow up on https://travis-ci.org/ubiquits/ubiquits/builds/141821325 when complete.
|
test
|
travis ci is picking up ts x not next which means there are a lot out errors output maybe intermittent follow up on when complete
| 1
|
184,949
| 14,291,131,339
|
IssuesEvent
|
2020-11-23 22:06:40
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
wptide/wptide: cmd/phpcs-server/main_test.go; 11 LoC
|
small test
|
Found a possible issue in [wptide/wptide](https://www.github.com/wptide/wptide) at [cmd/phpcs-server/main_test.go](https://github.com/wptide/wptide/blob/24ecc5d569823e858eb1414ce5386b67d3f31d46/cmd/phpcs-server/main_test.go#L58-L68)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/wptide/wptide/blob/24ecc5d569823e858eb1414ce5386b67d3f31d46/cmd/phpcs-server/main_test.go#L58-L68)
<details>
<summary>Click here to show the 11 line(s) of Go which triggered the analyzer.</summary>
```go
for key, val := range envTest {
// Key is set, so retain the value when the test is finished.
if value = os.Getenv(key); value != "" {
os.Unsetenv(key)
defer func() { os.Setenv(key, value) }()
}
// Set the test value.
os.Setenv(key, val)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 24ecc5d569823e858eb1414ce5386b67d3f31d46
|
1.0
|
wptide/wptide: cmd/phpcs-server/main_test.go; 11 LoC -
Found a possible issue in [wptide/wptide](https://www.github.com/wptide/wptide) at [cmd/phpcs-server/main_test.go](https://github.com/wptide/wptide/blob/24ecc5d569823e858eb1414ce5386b67d3f31d46/cmd/phpcs-server/main_test.go#L58-L68)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/wptide/wptide/blob/24ecc5d569823e858eb1414ce5386b67d3f31d46/cmd/phpcs-server/main_test.go#L58-L68)
<details>
<summary>Click here to show the 11 line(s) of Go which triggered the analyzer.</summary>
```go
for key, val := range envTest {
// Key is set, so retain the value when the test is finished.
if value = os.Getenv(key); value != "" {
os.Unsetenv(key)
defer func() { os.Setenv(key, value) }()
}
// Set the test value.
os.Setenv(key, val)
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: 24ecc5d569823e858eb1414ce5386b67d3f31d46
|
test
|
wptide wptide cmd phpcs server main test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for key val range envtest key is set so retain the value when the test is finished if value os getenv key value os unsetenv key defer func os setenv key value set the test value os setenv key val leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
13,599
| 3,754,825,934
|
IssuesEvent
|
2016-03-12 07:21:17
|
Rdatatable/data.table
|
https://api.github.com/repos/Rdatatable/data.table
|
closed
|
[R-Forge #5643] ?data.table should clear up the difference with data.frame on subset on logical column
|
documentation
|
Submitted by: Arun ; Assigned to: Nobody; [R-Forge link](https://r-forge.r-project.org/tracker/index.php?func=detail&aid=5643&group_id=240&atid=5356)
From here:
http://r.789695.n4.nabble.com/Subsetting-with-logical-td4689054.html
The documentation says currently (in ?data.table): "integer and logical vectors work the same way they do in [.data.frame. Other than NAs in logical i are treated as FALSE and a single NA logical is not recycled to match the number of rows, as it is in [.data.frame."
Perhaps the difference in subset on logical column (requiring parenthesis) should be made clear here.
|
1.0
|
[R-Forge #5643] ?data.table should clear up the difference with data.frame on subset on logical column - Submitted by: Arun ; Assigned to: Nobody; [R-Forge link](https://r-forge.r-project.org/tracker/index.php?func=detail&aid=5643&group_id=240&atid=5356)
From here:
http://r.789695.n4.nabble.com/Subsetting-with-logical-td4689054.html
The documentation says currently (in ?data.table): "integer and logical vectors work the same way they do in [.data.frame. Other than NAs in logical i are treated as FALSE and a single NA logical is not recycled to match the number of rows, as it is in [.data.frame."
Perhaps the difference in subset on logical column (requiring parenthesis) should be made clear here.
|
non_test
|
data table should clear up the difference with data frame on subset on logical column submitted by arun assigned to nobody from here the documentation says currently in data table integer and logical vectors work the same way they do in data frame other than nas in logical i are treated as false and a single na logical is not recycled to match the number of rows as it is in data frame perhaps the difference in subset on logical column requiring parenthesis should be made clear here
| 0
|
123,236
| 17,772,192,742
|
IssuesEvent
|
2021-08-30 14:50:25
|
kapseliboi/sqlpad
|
https://api.github.com/repos/kapseliboi/sqlpad
|
opened
|
CVE-2021-32796 (Medium) detected in xmldom-0.5.0.tgz, xmldom-0.6.0.tgz
|
security vulnerability
|
## CVE-2021-32796 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>xmldom-0.5.0.tgz</b>, <b>xmldom-0.6.0.tgz</b></p></summary>
<p>
<details><summary><b>xmldom-0.5.0.tgz</b></p></summary>
<p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.5.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.5.0.tgz</a></p>
<p>Path to dependency file: sqlpad/server/package.json</p>
<p>Path to vulnerable library: sqlpad/server/node_modules/passport-saml/node_modules/xmldom/package.json</p>
<p>
Dependency Hierarchy:
- passport-saml-2.2.0.tgz (Root Library)
- :x: **xmldom-0.5.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>xmldom-0.6.0.tgz</b></p></summary>
<p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz</a></p>
<p>Path to dependency file: sqlpad/server/package.json</p>
<p>Path to vulnerable library: sqlpad/server/node_modules/xmldom/package.json</p>
<p>
Dependency Hierarchy:
- passport-saml-2.2.0.tgz (Root Library)
- xml-crypto-2.1.2.tgz
- :x: **xmldom-0.6.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/sqlpad/commit/95024fc09fd71a1bc52f23bc0709ce5daa7e9f98">95024fc09fd71a1bc52f23bc0709ce5daa7e9f98</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
xmldom is an open source pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module. xmldom versions 0.6.0 and older do not correctly escape special characters when serializing elements removed from their ancestor. This may lead to unexpected syntactic changes during XML processing in some downstream applications. This issue has been resolved in version 0.7.0. As a workaround downstream applications can validate the input and reject the maliciously crafted documents.
<p>Publish Date: 2021-07-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32796>CVE-2021-32796</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q">https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q</a></p>
<p>Release Date: 2021-07-27</p>
<p>Fix Resolution: xmldom - 0.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-32796 (Medium) detected in xmldom-0.5.0.tgz, xmldom-0.6.0.tgz - ## CVE-2021-32796 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>xmldom-0.5.0.tgz</b>, <b>xmldom-0.6.0.tgz</b></p></summary>
<p>
<details><summary><b>xmldom-0.5.0.tgz</b></p></summary>
<p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.5.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.5.0.tgz</a></p>
<p>Path to dependency file: sqlpad/server/package.json</p>
<p>Path to vulnerable library: sqlpad/server/node_modules/passport-saml/node_modules/xmldom/package.json</p>
<p>
Dependency Hierarchy:
- passport-saml-2.2.0.tgz (Root Library)
- :x: **xmldom-0.5.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>xmldom-0.6.0.tgz</b></p></summary>
<p>A pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz">https://registry.npmjs.org/xmldom/-/xmldom-0.6.0.tgz</a></p>
<p>Path to dependency file: sqlpad/server/package.json</p>
<p>Path to vulnerable library: sqlpad/server/node_modules/xmldom/package.json</p>
<p>
Dependency Hierarchy:
- passport-saml-2.2.0.tgz (Root Library)
- xml-crypto-2.1.2.tgz
- :x: **xmldom-0.6.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/sqlpad/commit/95024fc09fd71a1bc52f23bc0709ce5daa7e9f98">95024fc09fd71a1bc52f23bc0709ce5daa7e9f98</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
xmldom is an open source pure JavaScript W3C standard-based (XML DOM Level 2 Core) DOMParser and XMLSerializer module. xmldom versions 0.6.0 and older do not correctly escape special characters when serializing elements removed from their ancestor. This may lead to unexpected syntactic changes during XML processing in some downstream applications. This issue has been resolved in version 0.7.0. As a workaround downstream applications can validate the input and reject the maliciously crafted documents.
<p>Publish Date: 2021-07-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32796>CVE-2021-32796</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q">https://github.com/xmldom/xmldom/security/advisories/GHSA-5fg8-2547-mr8q</a></p>
<p>Release Date: 2021-07-27</p>
<p>Fix Resolution: xmldom - 0.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in xmldom tgz xmldom tgz cve medium severity vulnerability vulnerable libraries xmldom tgz xmldom tgz xmldom tgz a pure javascript standard based xml dom level core domparser and xmlserializer module library home page a href path to dependency file sqlpad server package json path to vulnerable library sqlpad server node modules passport saml node modules xmldom package json dependency hierarchy passport saml tgz root library x xmldom tgz vulnerable library xmldom tgz a pure javascript standard based xml dom level core domparser and xmlserializer module library home page a href path to dependency file sqlpad server package json path to vulnerable library sqlpad server node modules xmldom package json dependency hierarchy passport saml tgz root library xml crypto tgz x xmldom tgz vulnerable library found in head commit a href found in base branch master vulnerability details xmldom is an open source pure javascript standard based xml dom level core domparser and xmlserializer module xmldom versions and older do not correctly escape special characters when serializing elements removed from their ancestor this may lead to unexpected syntactic changes during xml processing in some downstream applications this issue has been resolved in version as a workaround downstream applications can validate the input and reject the maliciously crafted documents publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution xmldom step up your open source security game with whitesource
| 0
|
302,610
| 26,154,295,012
|
IssuesEvent
|
2022-12-30 18:46:05
|
cancervariants/therapy-normalization
|
https://api.github.com/repos/cancervariants/therapy-normalization
|
closed
|
Use Mock classes for testing
|
test
|
(where possible)
https://docs.pytest.org/en/stable/monkeypatch.html?highlight=mocking#monkeypatching-returned-objects-building-mock-classes
* It's probably fine to pull most of the 'mock' db layer out of the existing normalization tests, as long as there's a way to catch update requests before they alter a local DB, validate them, and prevent them from making any changes.
|
1.0
|
Use Mock classes for testing - (where possible)
https://docs.pytest.org/en/stable/monkeypatch.html?highlight=mocking#monkeypatching-returned-objects-building-mock-classes
* It's probably fine to pull most of the 'mock' db layer out of the existing normalization tests, as long as there's a way to catch update requests before they alter a local DB, validate them, and prevent them from making any changes.
|
test
|
use mock classes for testing where possible it s probably fine to pull most of the mock db layer out of the existing normalization tests as long as there s a way to catch update requests before they alter a local db validate them and prevent them from making any changes
| 1
|
233,716
| 17,875,740,309
|
IssuesEvent
|
2021-09-07 03:12:04
|
PyTorchLightning/lightning-bolts
|
https://api.github.com/repos/PyTorchLightning/lightning-bolts
|
closed
|
Different pip installation command on the official website and doc site
|
documentation
|
On the front page of the [lightning website](https://www.pytorchlightning.ai), the installation guide says use `pip install pytorch-lightning` to install pytorch lightning, which results in subsequent failure from attempts to import pl_bolts as instructed by tutorials. While on the documentation site, the installation guide says use `pip install lightning-bolts` to install lightning, which works perfect.
Why on earth are there two different installation commands? Why does pytorch-lightning package exist at all?
|
1.0
|
Different pip installation command on the official website and doc site - On the front page of the [lightning website](https://www.pytorchlightning.ai), the installation guide says use `pip install pytorch-lightning` to install pytorch lightning, which results in subsequent failure from attempts to import pl_bolts as instructed by tutorials. While on the documentation site, the installation guide says use `pip install lightning-bolts` to install lightning, which works perfect.
Why on earth are there two different installation commands? Why does pytorch-lightning package exist at all?
|
non_test
|
different pip installation command on the official website and doc site on the front page of the the installation guide says use pip install pytorch lightning to install pytorch lightning which results in subsequent failure from attempts to import pl bolts as instructed by tutorials while on the documentation site the installation guide says use pip install lightning bolts to install lightning which works perfect why on earth are there two different installation commands why does pytorch lightning package exist at all
| 0
|
109,681
| 4,402,603,480
|
IssuesEvent
|
2016-08-11 02:10:56
|
CorWatts/fasterscale
|
https://api.github.com/repos/CorWatts/fasterscale
|
closed
|
Add an FAQ section
|
enhancement medium priority small
|
Answer common questions and provide links to specific questions at relevant parts of the app.
|
1.0
|
Add an FAQ section - Answer common questions and provide links to specific questions at relevant parts of the app.
|
non_test
|
add an faq section answer common questions and provide links to specific questions at relevant parts of the app
| 0
|
102,671
| 16,577,869,478
|
IssuesEvent
|
2021-05-31 07:49:59
|
AlexRogalskiy/github-action-random-proverb
|
https://api.github.com/repos/AlexRogalskiy/github-action-random-proverb
|
opened
|
CVE-2021-33502 (High) detected in normalize-url-6.0.0.tgz
|
security vulnerability
|
## CVE-2021-33502 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-6.0.0.tgz</b></p></summary>
<p>Normalize a URL</p>
<p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-6.0.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-6.0.0.tgz</a></p>
<p>Path to dependency file: github-action-random-proverb/package.json</p>
<p>Path to vulnerable library: github-action-random-proverb/node_modules/normalize-url/package.json</p>
<p>
Dependency Hierarchy:
- npm-7.1.1.tgz (Root Library)
- :x: **normalize-url-6.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-random-proverb/commit/d092a08d841932bff96facc4556f2ff4cce2c76b">d092a08d841932bff96facc4556f2ff4cce2c76b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs.
<p>Publish Date: 2021-05-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p>
<p>Release Date: 2021-05-24</p>
<p>Fix Resolution: normalize-url - 4.5.1, 5.3.1, 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-33502 (High) detected in normalize-url-6.0.0.tgz - ## CVE-2021-33502 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>normalize-url-6.0.0.tgz</b></p></summary>
<p>Normalize a URL</p>
<p>Library home page: <a href="https://registry.npmjs.org/normalize-url/-/normalize-url-6.0.0.tgz">https://registry.npmjs.org/normalize-url/-/normalize-url-6.0.0.tgz</a></p>
<p>Path to dependency file: github-action-random-proverb/package.json</p>
<p>Path to vulnerable library: github-action-random-proverb/node_modules/normalize-url/package.json</p>
<p>
Dependency Hierarchy:
- npm-7.1.1.tgz (Root Library)
- :x: **normalize-url-6.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-random-proverb/commit/d092a08d841932bff96facc4556f2ff4cce2c76b">d092a08d841932bff96facc4556f2ff4cce2c76b</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The normalize-url package before 4.5.1, 5.x before 5.3.1, and 6.x before 6.0.1 for Node.js has a ReDoS (regular expression denial of service) issue because it has exponential performance for data: URLs.
<p>Publish Date: 2021-05-24
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502>CVE-2021-33502</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33502</a></p>
<p>Release Date: 2021-05-24</p>
<p>Fix Resolution: normalize-url - 4.5.1, 5.3.1, 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve high detected in normalize url tgz cve high severity vulnerability vulnerable library normalize url tgz normalize a url library home page a href path to dependency file github action random proverb package json path to vulnerable library github action random proverb node modules normalize url package json dependency hierarchy npm tgz root library x normalize url tgz vulnerable library found in head commit a href vulnerability details the normalize url package before x before and x before for node js has a redos regular expression denial of service issue because it has exponential performance for data urls publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution normalize url step up your open source security game with whitesource
| 0
|
315,129
| 27,047,221,152
|
IssuesEvent
|
2023-02-13 10:37:30
|
stargate/stargate
|
https://api.github.com/repos/stargate/stargate
|
opened
|
Avoid using profiles in REST integration tests
|
test
|
There are two integration tests in REST that test the same thing based on the properties:
* RestApiV2QCqlDisabledIT
* RestApiV2QCqlIT
The second one specifies a profile that sets `stargate.rest.cql.disabled=false`.
However, although this works on the integration tests, it fails when we are using the integration tests with the running app. Simply, you can not switch the properties of the running app. IN that sense, the integration tests, which are effectively end-to-end tests should not specify profiles.
I am not sure what's the best way to solve this. Imo the `RestApiV2QCqlDisabledIT` could be classical `@QuarkusTest` with concrete profile and check that method is not existing. For the other, we could enable by default in the test resources that the option is `false` and then have some kind of a protection in the situation when we run against the running app. Not sure though, maybe a small discussion on this should be done.
|
1.0
|
Avoid using profiles in REST integration tests - There are two integration tests in REST that test the same thing based on the properties:
* RestApiV2QCqlDisabledIT
* RestApiV2QCqlIT
The second one specifies a profile that sets `stargate.rest.cql.disabled=false`.
However, although this works on the integration tests, it fails when we are using the integration tests with the running app. Simply, you can not switch the properties of the running app. IN that sense, the integration tests, which are effectively end-to-end tests should not specify profiles.
I am not sure what's the best way to solve this. Imo the `RestApiV2QCqlDisabledIT` could be classical `@QuarkusTest` with concrete profile and check that method is not existing. For the other, we could enable by default in the test resources that the option is `false` and then have some kind of a protection in the situation when we run against the running app. Not sure though, maybe a small discussion on this should be done.
|
test
|
avoid using profiles in rest integration tests there are two integration tests in rest that test the same thing based on the properties the second one specifies a profile that sets stargate rest cql disabled false however although this works on the integration tests it fails when we are using the integration tests with the running app simply you can not switch the properties of the running app in that sense the integration tests which are effectively end to end tests should not specify profiles i am not sure what s the best way to solve this imo the could be classical quarkustest with concrete profile and check that method is not existing for the other we could enable by default in the test resources that the option is false and then have some kind of a protection in the situation when we run against the running app not sure though maybe a small discussion on this should be done
| 1
|
43,157
| 7,028,752,787
|
IssuesEvent
|
2017-12-25 13:33:24
|
androguard/androguard
|
https://api.github.com/repos/androguard/androguard
|
closed
|
Deprecated documentation
|
documentation
|
I can't find the `get_AndroidManifest()` method in the apk.py script that is mentioned in the documentation. Maybe the documentation is deprecated?
APK class can't call `get_AndroidManifest()` method that is listed in the [documentation](http://androguard.readthedocs.io/en/latest/api/androguard.core.bytecodes.html#module-androguard.core.bytecodes.apk).
- Androguard Version: 2.0
- Python Version: 3.6
- Operating System: macOS v10.12.6
|
1.0
|
Deprecated documentation - I can't find the `get_AndroidManifest()` method in the apk.py script that is mentioned in the documentation. Maybe the documentation is deprecated?
APK class can't call `get_AndroidManifest()` method that is listed in the [documentation](http://androguard.readthedocs.io/en/latest/api/androguard.core.bytecodes.html#module-androguard.core.bytecodes.apk).
- Androguard Version: 2.0
- Python Version: 3.6
- Operating System: macOS v10.12.6
|
non_test
|
deprecated documentation i can t find the get androidmanifest method in the apk py script that is mentioned in the documentation maybe the documentation is deprecated apk class can t call get androidmanifest method that is listed in the androguard version python version operating system macos
| 0
|
596,920
| 18,150,460,501
|
IssuesEvent
|
2021-09-26 07:08:49
|
harvester/harvester
|
https://api.github.com/repos/harvester/harvester
|
opened
|
[BUG] A force-formatted disk was not added as an additional data volume
|
bug priority/1 area/node-disk-manager
|
**Describe the bug**
A force-formatted disk was not added as an additional data volume successfully.

**To Reproduce**
Steps to reproduce the behavior:
1. Select an additional disk that already has a partition and filesystem type
2. Set the `force-formatted` to true
3. click save and check the node disks again
4. no additional disk was added to the LH side
Issue:
We should provisioning the single root partition after force-format instead of the original block device.

**Expected behavior**
The force-formatted disk should be added as an additional data disk.
**Support bundle**
<!-- You can generate a support bundle in the bottom of Harvester UI. It includes logs and configurations that help diagnose the issue. -->
**Environment:**
- Harvester ISO version: master-head
- Underlying Infrastructure (e.g. Baremetal with Dell PowerEdge R630):
|
1.0
|
[BUG] A force-formatted disk was not added as an additional data volume - **Describe the bug**
A force-formatted disk was not added as an additional data volume successfully.

**To Reproduce**
Steps to reproduce the behavior:
1. Select an additional disk that already has a partition and filesystem type
2. Set the `force-formatted` to true
3. click save and check the node disks again
4. no additional disk was added to the LH side
Issue:
We should provisioning the single root partition after force-format instead of the original block device.

**Expected behavior**
The force-formatted disk should be added as an additional data disk.
**Support bundle**
<!-- You can generate a support bundle in the bottom of Harvester UI. It includes logs and configurations that help diagnose the issue. -->
**Environment:**
- Harvester ISO version: master-head
- Underlying Infrastructure (e.g. Baremetal with Dell PowerEdge R630):
|
non_test
|
a force formatted disk was not added as an additional data volume describe the bug a force formatted disk was not added as an additional data volume successfully to reproduce steps to reproduce the behavior select an additional disk that already has a partition and filesystem type set the force formatted to true click save and check the node disks again no additional disk was added to the lh side issue we should provisioning the single root partition after force format instead of the original block device expected behavior the force formatted disk should be added as an additional data disk support bundle environment harvester iso version master head underlying infrastructure e g baremetal with dell poweredge
| 0
|
162,768
| 12,690,388,857
|
IssuesEvent
|
2020-06-21 11:53:05
|
Oldes/Rebol-issues
|
https://api.github.com/repos/Oldes/Rebol-issues
|
closed
|
READ/string crashes for UCS4 ("UTF-32") LE/BE files with a BOM
|
Oldes.resolved Ren.important Status.important Test.written Type.bug
|
_Submitted by:_ **abolka**
See the example code.
``` rebol
>> read %fixtures/umlauts-utf32le.txt
== #{FFFE0000E4000000F6000000FC0000000A000000}
>> read/string %fixtures/umlauts-utf32le.txt
(R3 crashes) ;; Expected: "äöü^/"
--
>> read %fixtures/umlauts-utf32be.txt
== #{0000FEFF000000E4000000F6000000FC0000000A}
>> read %fixtures/umlauts-utf32be.txt
(R3 crashes) ;; Expected: "äöü^/"
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2186)** [ Version: alpha 111 Type: Bug Platform: All Category: Ports Reproduce: Always Fixed-in:none ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2186</sup>
Comments:
---
> **Rebolbot** commented on Nov 18, 2014:
_Submitted by:_ **abolka**
As we neither internally support UCS4 at the moment nor have a decoder ready (although the latter would be easy to remedy), causing an "unsupported encoding" error is probably the best solution at the moment.
---
> **Rebolbot** commented on Jan 20, 2015:
_Submitted by:_ **abolka**
In the core-tests suite.
---
> **Rebolbot** added **Type.bug**, **Status.important**, **Test.written** on Jan 12, 2016
---
> **Hostilefork** added **Type.bug**, **Status.important**, **Test.written**, **Ren.important** on Mar 22, 2018
---
> **Hostilefork** commented on Mar 22, 2018:
I've tagged this as important not because it is particularly *important* as a priority feature, but because the resolution of the non-passing tests was to remove them and let this issue track the problem. *(It is not the practice in Ren-C to commit tests which do not pass, but to leave them as issues, whose tests will be added when they do pass.)*
These were the deleted lines:
["äöü" == read/string %fixtures/umlauts-utf32le.txt]
[["äöü"] == read/lines %fixtures/umlauts-utf32le.txt]
["äöü" == read/string %fixtures/umlauts-utf32be.txt]
[["äöü"] == read/lines %fixtures/umlauts-utf32be.txt]
My general opinion is that one must distinguish between low-level, foundational byte operations *(for which I propose READ to be used, since that seems to be what's happening)* and higher-level metadata-driven operations, which might use LOAD *(or some other word, if LOAD is too load-ed a term for meaning Rebol source)*
---
> **Oldes** added a commit to **[Oldes/Rebol3](https://github.com/Oldes/Rebol3/)** that referenced this issue on Dec 18, 2018:
> [FIX: READ/string crashes for UCS4 ("UTF-32") LE/BE files with a BOM](https://github.com/Oldes/Rebol3/commit/9190f1ca96214e530f0d1c22fa390d08a78fb746)
---
|
1.0
|
READ/string crashes for UCS4 ("UTF-32") LE/BE files with a BOM - _Submitted by:_ **abolka**
See the example code.
``` rebol
>> read %fixtures/umlauts-utf32le.txt
== #{FFFE0000E4000000F6000000FC0000000A000000}
>> read/string %fixtures/umlauts-utf32le.txt
(R3 crashes) ;; Expected: "äöü^/"
--
>> read %fixtures/umlauts-utf32be.txt
== #{0000FEFF000000E4000000F6000000FC0000000A}
>> read %fixtures/umlauts-utf32be.txt
(R3 crashes) ;; Expected: "äöü^/"
```
---
<sup>**Imported from:** **[CureCode](https://www.curecode.org/rebol3/ticket.rsp?id=2186)** [ Version: alpha 111 Type: Bug Platform: All Category: Ports Reproduce: Always Fixed-in:none ]</sup>
<sup>**Imported from**: https://github.com/rebol/rebol-issues/issues/2186</sup>
Comments:
---
> **Rebolbot** commented on Nov 18, 2014:
_Submitted by:_ **abolka**
As we neither internally support UCS4 at the moment nor have a decoder ready (although the latter would be easy to remedy), causing an "unsupported encoding" error is probably the best solution at the moment.
---
> **Rebolbot** commented on Jan 20, 2015:
_Submitted by:_ **abolka**
In the core-tests suite.
---
> **Rebolbot** added **Type.bug**, **Status.important**, **Test.written** on Jan 12, 2016
---
> **Hostilefork** added **Type.bug**, **Status.important**, **Test.written**, **Ren.important** on Mar 22, 2018
---
> **Hostilefork** commented on Mar 22, 2018:
I've tagged this as important not because it is particularly *important* as a priority feature, but because the resolution of the non-passing tests was to remove them and let this issue track the problem. *(It is not the practice in Ren-C to commit tests which do not pass, but to leave them as issues, whose tests will be added when they do pass.)*
These were the deleted lines:
["äöü" == read/string %fixtures/umlauts-utf32le.txt]
[["äöü"] == read/lines %fixtures/umlauts-utf32le.txt]
["äöü" == read/string %fixtures/umlauts-utf32be.txt]
[["äöü"] == read/lines %fixtures/umlauts-utf32be.txt]
My general opinion is that one must distinguish between low-level, foundational byte operations *(for which I propose READ to be used, since that seems to be what's happening)* and higher-level metadata-driven operations, which might use LOAD *(or some other word, if LOAD is too load-ed a term for meaning Rebol source)*
---
> **Oldes** added a commit to **[Oldes/Rebol3](https://github.com/Oldes/Rebol3/)** that referenced this issue on Dec 18, 2018:
> [FIX: READ/string crashes for UCS4 ("UTF-32") LE/BE files with a BOM](https://github.com/Oldes/Rebol3/commit/9190f1ca96214e530f0d1c22fa390d08a78fb746)
---
|
test
|
read string crashes for utf le be files with a bom submitted by abolka see the example code rebol read fixtures umlauts txt read string fixtures umlauts txt crashes expected äöü read fixtures umlauts txt read fixtures umlauts txt crashes expected äöü imported from imported from comments rebolbot commented on nov submitted by abolka as we neither internally support at the moment nor have a decoder ready although the latter would be easy to remedy causing an unsupported encoding error is probably the best solution at the moment rebolbot commented on jan submitted by abolka in the core tests suite rebolbot added type bug status important test written on jan hostilefork added type bug status important test written ren important on mar hostilefork commented on mar i ve tagged this as important not because it is particularly important as a priority feature but because the resolution of the non passing tests was to remove them and let this issue track the problem it is not the practice in ren c to commit tests which do not pass but to leave them as issues whose tests will be added when they do pass these were the deleted lines read lines fixtures umlauts txt read lines fixtures umlauts txt my general opinion is that one must distinguish between low level foundational byte operations for which i propose read to be used since that seems to be what s happening and higher level metadata driven operations which might use load or some other word if load is too load ed a term for meaning rebol source oldes added a commit to that referenced this issue on dec
| 1
|
641,747
| 20,833,630,199
|
IssuesEvent
|
2022-03-19 21:15:23
|
SoftwareEngineeringGroup3-3/recipe-app-backend
|
https://api.github.com/repos/SoftwareEngineeringGroup3-3/recipe-app-backend
|
closed
|
[BACKEND] [TESTING] Add better tests for API/Ingredients endpoint
|
enhancement priority:high
|
Add more comprehensive tests for API/Ingredients endpoint which will test the actual response instead of just the validation functions by themselves.
|
1.0
|
[BACKEND] [TESTING] Add better tests for API/Ingredients endpoint - Add more comprehensive tests for API/Ingredients endpoint which will test the actual response instead of just the validation functions by themselves.
|
non_test
|
add better tests for api ingredients endpoint add more comprehensive tests for api ingredients endpoint which will test the actual response instead of just the validation functions by themselves
| 0
|
288,398
| 21,702,890,081
|
IssuesEvent
|
2022-05-10 06:54:25
|
GenericMappingTools/pygmt
|
https://api.github.com/repos/GenericMappingTools/pygmt
|
closed
|
Point of view in docstrings
|
question documentation
|
Some docstrings for PyGMT and GMT modules use plural first person when describing what the function does (e.g "we create labels" in `grd2cpt`, "we will write a GeoTiff image" in `grdimage`). Personally, I'm not a fan of this, as there is no "we" doing the operations; it's PyGMT/GMT that is carrying them out. While I don't think any users would misinterpret this as the PyGMT team doing the work behind-the-scenes, I still think the docstrings should be changed to accurately portray who or what is making assumptions and running calculations. Should I submit a PR with changes for these types of docstrings?
This originally came up in [my comment](https://github.com/GenericMappingTools/pygmt/pull/1122#discussion_r734937993) for `project`.
**Examples:**
`color_model` in `grd2cpt`
Old
```
If *label* is appended then we create labels for each category to be used when the CPT is plotted.
```
New
```
Append *label* to create labels for each category to be used when the CPT is plotted.
```
`nearest_multiple` in `grdinfo`
Old
```
We report the result via the text string *zmin/zmax* or *zmin/zmax/dz*
```
New
```
The result is returned in the text string format *zmin/zmax* or *zmin/zmax/dz*
```
**Are you willing to help implement and maintain this feature?** Yes
|
1.0
|
Point of view in docstrings - Some docstrings for PyGMT and GMT modules use plural first person when describing what the function does (e.g "we create labels" in `grd2cpt`, "we will write a GeoTiff image" in `grdimage`). Personally, I'm not a fan of this, as there is no "we" doing the operations; it's PyGMT/GMT that is carrying them out. While I don't think any users would misinterpret this as the PyGMT team doing the work behind-the-scenes, I still think the docstrings should be changed to accurately portray who or what is making assumptions and running calculations. Should I submit a PR with changes for these types of docstrings?
This originally came up in [my comment](https://github.com/GenericMappingTools/pygmt/pull/1122#discussion_r734937993) for `project`.
**Examples:**
`color_model` in `grd2cpt`
Old
```
If *label* is appended then we create labels for each category to be used when the CPT is plotted.
```
New
```
Append *label* to create labels for each category to be used when the CPT is plotted.
```
`nearest_multiple` in `grdinfo`
Old
```
We report the result via the text string *zmin/zmax* or *zmin/zmax/dz*
```
New
```
The result is returned in the text string format *zmin/zmax* or *zmin/zmax/dz*
```
**Are you willing to help implement and maintain this feature?** Yes
|
non_test
|
point of view in docstrings some docstrings for pygmt and gmt modules use plural first person when describing what the function does e g we create labels in we will write a geotiff image in grdimage personally i m not a fan of this as there is no we doing the operations it s pygmt gmt that is carrying them out while i don t think any users would misinterpret this as the pygmt team doing the work behind the scenes i still think the docstrings should be changed to accurately portray who or what is making assumptions and running calculations should i submit a pr with changes for these types of docstrings this originally came up in for project examples color model in old if label is appended then we create labels for each category to be used when the cpt is plotted new append label to create labels for each category to be used when the cpt is plotted nearest multiple in grdinfo old we report the result via the text string zmin zmax or zmin zmax dz new the result is returned in the text string format zmin zmax or zmin zmax dz are you willing to help implement and maintain this feature yes
| 0
|
15,065
| 9,466,171,018
|
IssuesEvent
|
2019-04-18 03:09:12
|
PennyLin1127/Remediate
|
https://api.github.com/repos/PennyLin1127/Remediate
|
opened
|
CVE-2018-14719 High Severity Vulnerability detected by WhiteSource
|
security vulnerability
|
## CVE-2018-14719 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /Remediate/WebGoat-7.1/webgoat-standalone/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.3/jackson-databind-2.6.3.jar,2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.3/jackson-databind-2.6.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PennyLin1127/Remediate/commit/9a0657d4132e44ac3cb2e85ef738091d934b87ac">9a0657d4132e44ac3cb2e85ef738091d934b87ac</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-14719>CVE-2018-14719</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14719">https://nvd.nist.gov/vuln/detail/CVE-2018-14719</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isOpenPROnNewVersion":false,"isPackageBased":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.3","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"vulnerabilityIdentifier":"CVE-2018-14719","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2018-14719 High Severity Vulnerability detected by WhiteSource - ## CVE-2018-14719 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.3.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /Remediate/WebGoat-7.1/webgoat-standalone/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.3/jackson-databind-2.6.3.jar,2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.3/jackson-databind-2.6.3.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.3.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/PennyLin1127/Remediate/commit/9a0657d4132e44ac3cb2e85ef738091d934b87ac">9a0657d4132e44ac3cb2e85ef738091d934b87ac</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.
<p>Publish Date: 2019-01-02
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-14719>CVE-2018-14719</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-14719">https://nvd.nist.gov/vuln/detail/CVE-2018-14719</a></p>
<p>Release Date: 2019-01-02</p>
<p>Fix Resolution: 2.9.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isOpenPROnNewVersion":false,"isPackageBased":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.3","isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.7"}],"vulnerabilityIdentifier":"CVE-2018-14719","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.7 might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze-ds-opt and blaze-ds-core classes from polymorphic deserialization.","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_test
|
cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file remediate webgoat webgoat standalone pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze ds opt and blaze ds core classes from polymorphic deserialization publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource isopenpronvulnerability true isopenpronnewversion false ispackagebased true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before might allow remote attackers to execute arbitrary code by leveraging failure to block the blaze ds opt and blaze ds core classes from polymorphic deserialization high a high ac low pr none s unchanged c high ui none av network i high extradata
| 0
|
587,629
| 17,627,309,281
|
IssuesEvent
|
2021-08-19 00:28:22
|
zephyrproject-rtos/zephyr
|
https://api.github.com/repos/zephyrproject-rtos/zephyr
|
closed
|
enc424j600 driver unusable/broken on stm32l552
|
bug priority: low Stale
|
**Describe the bug**
Testing with the dumb_http_server_mt sample with an STM32 nucleo board and enc424j600 ethernet results in an unusable application which outputs lots of error
**To Reproduce**
Steps to reproduce the behavior:
1. Configure ethernet for nucleo_l552ze_q with following:
```
&spi1 {
pinctrl-0 = <&spi1_nss_pa4 &spi1_sck_pa5
&spi1_miso_pa6 &spi1_mosi_pa7>;
status = "okay";
cs-gpios = <&gpiod 14 GPIO_ACTIVE_LOW>;
enc424j600@0 {
compatible = "microchip,enc424j600";
reg = <0>;
spi-max-frequency = <4000000>;
label = "ETHERNET";
int-gpios = <&gpiod 15 GPIO_ACTIVE_LOW>;
};
};
```
2. Build dumb_http_server_mt
3. Flash to board
4. Attempt to ping or connect to board and a multitude of errors is emitted
**Expected behavior**
Connectivity to work
**Impact**
Network is useless
**Logs and console output**
```
*** Booting Zephyr OS build v2.6.0-rc1-300-g6ce0f2ee6606 ***
[00:00:00.003,000] <dbg> ethdrv.enc424j600_init: EIE: 0x0850
[00:00:00.003,000] <dbg> ethdrv.enc424j600_init_filters: ERXFCON: 0x005b
[00:00:00.003,000] <dbg> ethdrv.enc424j600_init_phy: PHANA: 0x05e1
[00:00:00.004,000] <dbg> ethdrv.enc424j600_init_phy: PHCON1: 0x1200
[00:00:00.004,000] <dbg> ethdrv.enc424j600_init: ECON1: 0x0001
[00:00:00.004,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xda00
[00:00:00.004,000] <inf> ethdrv: Link down
[00:00:00.004,000] <inf> ethdrv: ENC424J600 Initialized
[00:00:01.561,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf00
[00:00:01.561,000] <inf> ethdrv: Link up
[00:00:01.562,000] <dbg> ethdrv.enc424j600_setup_mac: PHANLPA: 0x85e1
[00:00:01.562,000] <inf> ethdrv: 100Mbps
[00:00:01.562,000] <inf> ethdrv: full duplex
[00:00:01.562,000] <dbg> ethdrv.enc424j600_setup_mac: MACON2: 0x40b3
[00:00:01.562,000] <dbg> ethdrv.enc424j600_setup_mac: MAMXFL (maximum frame length): 1518
[00:00:01.562,000] <inf> ethdrv: Not suspended
[00:00:01.562,000] <inf> net_dumb_http_srv_mt_sample: Network connected
[00:00:01.563,000] <dbg> net_dumb_http_srv_mt_sample.process_tcp4: Waiting for IPv4 HTTP connections on port 8080, sock 0
uart:~$ net iface
Interface 0x200016dc (Ethernet) [1]
===================================
Link addr : 68:27:19:EF:56:10
MTU : 1500
Flags : NO_AUTO_START,IPv4
Ethernet capabilities supported:
10 Mbits
100 Mbits
IPv4 unicast addresses (max 1):
192.168.1.55 manual preferred infinite
IPv4 multicast addresses (max 1):
<none>
IPv4 gateway : 0.0.0.0
IPv4 netmask : 255.255.255.0
[00:00:09.547,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x3000
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x3008 now
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx: npp 0x4800, length 16432, status 0x6003c000
[00:00:09.548,000] <err> ethdrv: Maximum frame length exceeded
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:10.259,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:10.259,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x4800
[00:00:10.260,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x4808 now
[00:00:10.260,000] <dbg> ethdrv.enc424j600_rx: npp 0xd100, length 40739, status 0xe257ca55
[00:00:10.260,000] <err> ethdrv: Maximum frame length exceeded
[00:00:10.260,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0xd100
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0xd108 now
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx: npp 0xa900, length 43966, status 0x8b581d15
[00:00:11.258,000] <err> ethdrv: Maximum frame length exceeded
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0xa900
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0xa908 now
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx: npp 0x9e00, length 42698, status 0xd774e5f4
[00:00:12.273,000] <err> ethdrv: Maximum frame length exceeded
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x9e00
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x9e08 now
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx: npp 0x7c00, length 4449, status 0xc0d460d3
[00:00:13.241,000] <err> ethdrv: Maximum frame length exceeded
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:14.242,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x7c00
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x7c08 now
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx: npp 0x3400, length 17654, status 0x537f51c5
[00:00:14.243,000] <err> ethdrv: Maximum frame length exceeded
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:19.899,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:19.899,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x3400
--- 3 messages dropped ---
```
**Environment (please complete the following information):**
- OS: Windows
- Toolchain: GNU Arm Embedded Toolchain 9-2020-q2-update 9.3.1
- Commit SHA: 6ce0f2ee6606915c75e17753e34db71cb053c119
|
1.0
|
enc424j600 driver unusable/broken on stm32l552 - **Describe the bug**
Testing with the dumb_http_server_mt sample with an STM32 nucleo board and enc424j600 ethernet results in an unusable application which outputs lots of error
**To Reproduce**
Steps to reproduce the behavior:
1. Configure ethernet for nucleo_l552ze_q with following:
```
&spi1 {
pinctrl-0 = <&spi1_nss_pa4 &spi1_sck_pa5
&spi1_miso_pa6 &spi1_mosi_pa7>;
status = "okay";
cs-gpios = <&gpiod 14 GPIO_ACTIVE_LOW>;
enc424j600@0 {
compatible = "microchip,enc424j600";
reg = <0>;
spi-max-frequency = <4000000>;
label = "ETHERNET";
int-gpios = <&gpiod 15 GPIO_ACTIVE_LOW>;
};
};
```
2. Build dumb_http_server_mt
3. Flash to board
4. Attempt to ping or connect to board and a multitude of errors is emitted
**Expected behavior**
Connectivity to work
**Impact**
Network is useless
**Logs and console output**
```
*** Booting Zephyr OS build v2.6.0-rc1-300-g6ce0f2ee6606 ***
[00:00:00.003,000] <dbg> ethdrv.enc424j600_init: EIE: 0x0850
[00:00:00.003,000] <dbg> ethdrv.enc424j600_init_filters: ERXFCON: 0x005b
[00:00:00.003,000] <dbg> ethdrv.enc424j600_init_phy: PHANA: 0x05e1
[00:00:00.004,000] <dbg> ethdrv.enc424j600_init_phy: PHCON1: 0x1200
[00:00:00.004,000] <dbg> ethdrv.enc424j600_init: ECON1: 0x0001
[00:00:00.004,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xda00
[00:00:00.004,000] <inf> ethdrv: Link down
[00:00:00.004,000] <inf> ethdrv: ENC424J600 Initialized
[00:00:01.561,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf00
[00:00:01.561,000] <inf> ethdrv: Link up
[00:00:01.562,000] <dbg> ethdrv.enc424j600_setup_mac: PHANLPA: 0x85e1
[00:00:01.562,000] <inf> ethdrv: 100Mbps
[00:00:01.562,000] <inf> ethdrv: full duplex
[00:00:01.562,000] <dbg> ethdrv.enc424j600_setup_mac: MACON2: 0x40b3
[00:00:01.562,000] <dbg> ethdrv.enc424j600_setup_mac: MAMXFL (maximum frame length): 1518
[00:00:01.562,000] <inf> ethdrv: Not suspended
[00:00:01.562,000] <inf> net_dumb_http_srv_mt_sample: Network connected
[00:00:01.563,000] <dbg> net_dumb_http_srv_mt_sample.process_tcp4: Waiting for IPv4 HTTP connections on port 8080, sock 0
uart:~$ net iface
Interface 0x200016dc (Ethernet) [1]
===================================
Link addr : 68:27:19:EF:56:10
MTU : 1500
Flags : NO_AUTO_START,IPv4
Ethernet capabilities supported:
10 Mbits
100 Mbits
IPv4 unicast addresses (max 1):
192.168.1.55 manual preferred infinite
IPv4 multicast addresses (max 1):
<none>
IPv4 gateway : 0.0.0.0
IPv4 netmask : 255.255.255.0
[00:00:09.547,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x3000
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x3008 now
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx: npp 0x4800, length 16432, status 0x6003c000
[00:00:09.548,000] <err> ethdrv: Maximum frame length exceeded
[00:00:09.548,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:10.259,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:10.259,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x4800
[00:00:10.260,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x4808 now
[00:00:10.260,000] <dbg> ethdrv.enc424j600_rx: npp 0xd100, length 40739, status 0xe257ca55
[00:00:10.260,000] <err> ethdrv: Maximum frame length exceeded
[00:00:10.260,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0xd100
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0xd108 now
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx: npp 0xa900, length 43966, status 0x8b581d15
[00:00:11.258,000] <err> ethdrv: Maximum frame length exceeded
[00:00:11.258,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0xa900
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0xa908 now
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx: npp 0x9e00, length 42698, status 0xd774e5f4
[00:00:12.273,000] <err> ethdrv: Maximum frame length exceeded
[00:00:12.273,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x9e00
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x9e08 now
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx: npp 0x7c00, length 4449, status 0xc0d460d3
[00:00:13.241,000] <err> ethdrv: Maximum frame length exceeded
[00:00:13.241,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:14.242,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x7c00
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx: ERXRDPT is 0x7c08 now
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx: npp 0x3400, length 17654, status 0x537f51c5
[00:00:14.243,000] <err> ethdrv: Maximum frame length exceeded
[00:00:14.243,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0x5f00
[00:00:19.899,000] <dbg> ethdrv.enc424j600_rx_thread: ESTAT: 0xdf01
[00:00:19.899,000] <dbg> ethdrv.enc424j600_rx: set ERXRDPT to 0x3400
--- 3 messages dropped ---
```
**Environment (please complete the following information):**
- OS: Windows
- Toolchain: GNU Arm Embedded Toolchain 9-2020-q2-update 9.3.1
- Commit SHA: 6ce0f2ee6606915c75e17753e34db71cb053c119
|
non_test
|
driver unusable broken on describe the bug testing with the dumb http server mt sample with an nucleo board and ethernet results in an unusable application which outputs lots of error to reproduce steps to reproduce the behavior configure ethernet for nucleo q with following pinctrl nss sck miso mosi status okay cs gpios compatible microchip reg spi max frequency label ethernet int gpios build dumb http server mt flash to board attempt to ping or connect to board and a multitude of errors is emitted expected behavior connectivity to work impact network is useless logs and console output booting zephyr os build ethdrv init eie ethdrv init filters erxfcon ethdrv init phy phana ethdrv init phy ethdrv init ethdrv rx thread estat ethdrv link down ethdrv initialized ethdrv rx thread estat ethdrv link up ethdrv setup mac phanlpa ethdrv ethdrv full duplex ethdrv setup mac ethdrv setup mac mamxfl maximum frame length ethdrv not suspended net dumb http srv mt sample network connected net dumb http srv mt sample process waiting for http connections on port sock uart net iface interface ethernet link addr ef mtu flags no auto start ethernet capabilities supported mbits mbits unicast addresses max manual preferred infinite multicast addresses max gateway netmask ethdrv rx thread estat ethdrv rx set erxrdpt to ethdrv rx erxrdpt is now ethdrv rx npp length status ethdrv maximum frame length exceeded ethdrv rx thread estat ethdrv rx thread estat ethdrv rx set erxrdpt to ethdrv rx erxrdpt is now ethdrv rx npp length status ethdrv maximum frame length exceeded ethdrv rx thread estat ethdrv rx thread estat ethdrv rx set erxrdpt to ethdrv rx erxrdpt is now ethdrv rx npp length status ethdrv maximum frame length exceeded ethdrv rx thread estat ethdrv rx thread estat ethdrv rx set erxrdpt to ethdrv rx erxrdpt is now ethdrv rx npp length status ethdrv maximum frame length exceeded ethdrv rx thread estat ethdrv rx thread estat ethdrv rx set erxrdpt to ethdrv rx erxrdpt is now ethdrv rx npp length status ethdrv maximum frame length exceeded ethdrv rx thread estat ethdrv rx thread estat ethdrv rx set erxrdpt to ethdrv rx erxrdpt is now ethdrv rx npp length status ethdrv maximum frame length exceeded ethdrv rx thread estat ethdrv rx thread estat ethdrv rx set erxrdpt to messages dropped environment please complete the following information os windows toolchain gnu arm embedded toolchain update commit sha
| 0
|
6,253
| 7,543,678,655
|
IssuesEvent
|
2018-04-17 16:08:00
|
ga4gh/dockstore
|
https://api.github.com/repos/ga4gh/dockstore
|
closed
|
Base command is optional, post CWL 1.0
|
bug cli gui web service
|
## Feature Request
### Desired behaviour
It should probably be discouraged, but `baseCommand` is optional.
Ensure that tools with no baseCommand validate as valid CWL.
For now, workaround by providing `baseCommand: []`
|
1.0
|
Base command is optional, post CWL 1.0 - ## Feature Request
### Desired behaviour
It should probably be discouraged, but `baseCommand` is optional.
Ensure that tools with no baseCommand validate as valid CWL.
For now, workaround by providing `baseCommand: []`
|
non_test
|
base command is optional post cwl feature request desired behaviour it should probably be discouraged but basecommand is optional ensure that tools with no basecommand validate as valid cwl for now workaround by providing basecommand
| 0
|
238,602
| 26,140,610,638
|
IssuesEvent
|
2022-12-29 17:52:07
|
vectordotdev/vector
|
https://api.github.com/repos/vectordotdev/vector
|
closed
|
[RUSTSEC-2020-0095]: difference is unmaintained
|
domain: security meta: blocked domain: deps
|
Added to `deny.toml` in https://github.com/timberio/vector/pull/6226
```
┌─ /home/kirill/tmp/vector-deny/Cargo.lock:139:1
│
139 │ difference 2.0.0 registry+https://github.com/rust-lang/crates.io-index
│ ---------------------------------------------------------------------- unmaintained advisory detected
│
= ID: RUSTSEC-2020-0095
= Advisory: https://rustsec.org/advisories/RUSTSEC-2020-0095
= The author of the `difference` crate is unresponsive.
Maintained alternatives:
- [`dissimilar`](https://crates.io/crates/dissimilar)
- [`similar`](https://crates.io/crates/similar)
- [`treediff`](https://crates.io/crates/treediff)
- [`diffus`](https://crates.io/crates/diffus)
= Announcement: https://github.com/johannhof/difference.rs/issues/45
= Solution: No safe upgrade is available!
= difference v2.0.0
├── predicates v1.0.6
│ └── assert_cmd v1.0.2
│ └── (dev) vector v0.12.0
└── pretty_assertions v0.6.1
└── (dev) vector v0.12.0 (*)
```
|
True
|
[RUSTSEC-2020-0095]: difference is unmaintained - Added to `deny.toml` in https://github.com/timberio/vector/pull/6226
```
┌─ /home/kirill/tmp/vector-deny/Cargo.lock:139:1
│
139 │ difference 2.0.0 registry+https://github.com/rust-lang/crates.io-index
│ ---------------------------------------------------------------------- unmaintained advisory detected
│
= ID: RUSTSEC-2020-0095
= Advisory: https://rustsec.org/advisories/RUSTSEC-2020-0095
= The author of the `difference` crate is unresponsive.
Maintained alternatives:
- [`dissimilar`](https://crates.io/crates/dissimilar)
- [`similar`](https://crates.io/crates/similar)
- [`treediff`](https://crates.io/crates/treediff)
- [`diffus`](https://crates.io/crates/diffus)
= Announcement: https://github.com/johannhof/difference.rs/issues/45
= Solution: No safe upgrade is available!
= difference v2.0.0
├── predicates v1.0.6
│ └── assert_cmd v1.0.2
│ └── (dev) vector v0.12.0
└── pretty_assertions v0.6.1
└── (dev) vector v0.12.0 (*)
```
|
non_test
|
difference is unmaintained added to deny toml in ┌─ home kirill tmp vector deny cargo lock │ │ difference registry │ unmaintained advisory detected │ id rustsec advisory the author of the difference crate is unresponsive maintained alternatives announcement solution no safe upgrade is available difference ├── predicates │ └── assert cmd │ └── dev vector └── pretty assertions └── dev vector
| 0
|
10,480
| 8,066,788,420
|
IssuesEvent
|
2018-08-04 20:14:52
|
kenger-dk/DAW
|
https://api.github.com/repos/kenger-dk/DAW
|
closed
|
Error pages
|
enhancement security
|
Make nice error pages and handle all errors this way, no .net error pages.
Error logging?
|
True
|
Error pages - Make nice error pages and handle all errors this way, no .net error pages.
Error logging?
|
non_test
|
error pages make nice error pages and handle all errors this way no net error pages error logging
| 0
|
61,491
| 15,014,694,337
|
IssuesEvent
|
2021-02-01 07:05:43
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[SB] Consent > Auto-created consent document > Reduce the font size of the title
|
Bug P2 Process: Fixed Process: Tested dev Study builder
|
SB > Consent > Auto-created consent document > Reduce the font size of the titles of the consent sections in the auto-created document. They should proportional to the font size of the section's content.

|
1.0
|
[SB] Consent > Auto-created consent document > Reduce the font size of the title - SB > Consent > Auto-created consent document > Reduce the font size of the titles of the consent sections in the auto-created document. They should proportional to the font size of the section's content.

|
non_test
|
consent auto created consent document reduce the font size of the title sb consent auto created consent document reduce the font size of the titles of the consent sections in the auto created document they should proportional to the font size of the section s content
| 0
|
216,826
| 16,820,367,941
|
IssuesEvent
|
2021-06-17 12:27:33
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: tpcc/w=100/nodes=3/chaos=true failed
|
C-test-failure O-roachtest O-robot branch-master release-blocker
|
roachtest.tpcc/w=100/nodes=3/chaos=true [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3087613&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3087613&tab=artifacts#/tpcc/w=100/nodes=3/chaos=true) on master @ [ec36bc66f5147a8703719a9c890be12dd0c08945](https://github.com/cockroachdb/cockroach/commits/ec36bc66f5147a8703719a9c890be12dd0c08945):
```
| 53.0s 237715 0.0 108.7 0.0 0.0 0.0 0.0 newOrder
| 53.0s 237715 0.0 10.9 0.0 0.0 0.0 0.0 orderStatus
| 53.0s 237715 0.0 109.6 0.0 0.0 0.0 0.0 payment
| 53.0s 237715 0.0 10.9 0.0 0.0 0.0 0.0 stockLevel
| 54.0s 264391 0.0 10.7 0.0 0.0 0.0 0.0 delivery
| 54.0s 264391 0.0 106.8 0.0 0.0 0.0 0.0 newOrder
| 54.0s 264391 0.0 10.7 0.0 0.0 0.0 0.0 orderStatus
| 54.0s 264391 0.0 107.6 0.0 0.0 0.0 0.0 payment
| 54.0s 264391 0.0 10.7 0.0 0.0 0.0 0.0 stockLevel
| 55.0s 291233 1.0 10.5 7784.6 7784.6 7784.6 7784.6 delivery
| 55.0s 291233 2.0 104.9 8321.5 11274.3 11274.3 11274.3 newOrder
| 55.0s 291233 2.0 10.6 10.5 10737.4 10737.4 10737.4 orderStatus
| 55.0s 291233 6.0 105.8 10737.4 11811.2 11811.2 11811.2 payment
| 55.0s 291233 2.0 10.6 19.9 25.2 25.2 25.2 stockLevel
| 56.0s 316833 4.0 10.4 604.0 12884.9 12884.9 12884.9 delivery
| 56.0s 316833 107.9 104.9 2281.7 12348.0 12884.9 12884.9 newOrder
| 56.0s 316833 7.0 10.5 83.9 109.1 109.1 109.1 orderStatus
| 56.0s 316833 92.9 105.6 1006.6 12348.0 12348.0 12884.9 payment
| 56.0s 316833 13.0 10.6 209.7 11274.3 11811.2 11811.2 stockLevel
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 57.0s 342058 7.0 10.4 738.2 906.0 906.0 906.0 delivery
| 57.0s 342058 110.0 105.0 486.5 1040.2 12348.0 12884.9 newOrder
| 57.0s 342058 14.0 10.6 71.3 159.4 201.3 201.3 orderStatus
| 57.0s 342058 124.0 105.9 385.9 872.4 8321.5 10737.4 payment
| 57.0s 342058 9.0 10.6 113.2 335.5 335.5 335.5 stockLevel
Wraps: (4) secondary error attachment
| signal: killed
| (1) signal: killed
| Error types: (1) *exec.ExitError
Wraps: (5) context canceled
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *secondary.withSecondaryError (5) *errors.errorString
cluster.go:2484,tpcc.go:164,tpcc.go:410,test_runner.go:757: monitor failure: unexpected node event: 2: dead (exit status 0)
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2472
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2480
| main.runTPCC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:164
| main.registerTPCC.func6
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:410
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:757
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) monitor failure
Wraps: (3) unexpected node event: 2: dead (exit status 0)
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString
```
<details><summary>Reproduce</summary>
<p>
<p>To reproduce, try:
```bash
# From https://go.crdb.dev/p/roachstress, perhaps edited lightly.
caffeinate ./roachstress.sh tpcc/w=100/nodes=3/chaos=true
```
</p>
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tpcc/w=100/nodes=3/chaos=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
2.0
|
roachtest: tpcc/w=100/nodes=3/chaos=true failed - roachtest.tpcc/w=100/nodes=3/chaos=true [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=3087613&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=3087613&tab=artifacts#/tpcc/w=100/nodes=3/chaos=true) on master @ [ec36bc66f5147a8703719a9c890be12dd0c08945](https://github.com/cockroachdb/cockroach/commits/ec36bc66f5147a8703719a9c890be12dd0c08945):
```
| 53.0s 237715 0.0 108.7 0.0 0.0 0.0 0.0 newOrder
| 53.0s 237715 0.0 10.9 0.0 0.0 0.0 0.0 orderStatus
| 53.0s 237715 0.0 109.6 0.0 0.0 0.0 0.0 payment
| 53.0s 237715 0.0 10.9 0.0 0.0 0.0 0.0 stockLevel
| 54.0s 264391 0.0 10.7 0.0 0.0 0.0 0.0 delivery
| 54.0s 264391 0.0 106.8 0.0 0.0 0.0 0.0 newOrder
| 54.0s 264391 0.0 10.7 0.0 0.0 0.0 0.0 orderStatus
| 54.0s 264391 0.0 107.6 0.0 0.0 0.0 0.0 payment
| 54.0s 264391 0.0 10.7 0.0 0.0 0.0 0.0 stockLevel
| 55.0s 291233 1.0 10.5 7784.6 7784.6 7784.6 7784.6 delivery
| 55.0s 291233 2.0 104.9 8321.5 11274.3 11274.3 11274.3 newOrder
| 55.0s 291233 2.0 10.6 10.5 10737.4 10737.4 10737.4 orderStatus
| 55.0s 291233 6.0 105.8 10737.4 11811.2 11811.2 11811.2 payment
| 55.0s 291233 2.0 10.6 19.9 25.2 25.2 25.2 stockLevel
| 56.0s 316833 4.0 10.4 604.0 12884.9 12884.9 12884.9 delivery
| 56.0s 316833 107.9 104.9 2281.7 12348.0 12884.9 12884.9 newOrder
| 56.0s 316833 7.0 10.5 83.9 109.1 109.1 109.1 orderStatus
| 56.0s 316833 92.9 105.6 1006.6 12348.0 12348.0 12884.9 payment
| 56.0s 316833 13.0 10.6 209.7 11274.3 11811.2 11811.2 stockLevel
| _elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
| 57.0s 342058 7.0 10.4 738.2 906.0 906.0 906.0 delivery
| 57.0s 342058 110.0 105.0 486.5 1040.2 12348.0 12884.9 newOrder
| 57.0s 342058 14.0 10.6 71.3 159.4 201.3 201.3 orderStatus
| 57.0s 342058 124.0 105.9 385.9 872.4 8321.5 10737.4 payment
| 57.0s 342058 9.0 10.6 113.2 335.5 335.5 335.5 stockLevel
Wraps: (4) secondary error attachment
| signal: killed
| (1) signal: killed
| Error types: (1) *exec.ExitError
Wraps: (5) context canceled
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *main.withCommandDetails (4) *secondary.withSecondaryError (5) *errors.errorString
cluster.go:2484,tpcc.go:164,tpcc.go:410,test_runner.go:757: monitor failure: unexpected node event: 2: dead (exit status 0)
(1) attached stack trace
-- stack trace:
| main.(*monitor).WaitE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2472
| main.(*monitor).Wait
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2480
| main.runTPCC
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:164
| main.registerTPCC.func6
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tpcc.go:410
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:757
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) monitor failure
Wraps: (3) unexpected node event: 2: dead (exit status 0)
Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *errors.errorString
```
<details><summary>Reproduce</summary>
<p>
<p>To reproduce, try:
```bash
# From https://go.crdb.dev/p/roachstress, perhaps edited lightly.
caffeinate ./roachstress.sh tpcc/w=100/nodes=3/chaos=true
```
</p>
</p>
</details>
/cc @cockroachdb/kv
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*tpcc/w=100/nodes=3/chaos=true.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
|
test
|
roachtest tpcc w nodes chaos true failed roachtest tpcc w nodes chaos true with on master neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel delivery neworder orderstatus payment stocklevel elapsed errors ops sec inst ops sec cum ms ms ms pmax ms delivery neworder orderstatus payment stocklevel wraps secondary error attachment signal killed signal killed error types exec exiterror wraps context canceled error types withstack withstack errutil withprefix main withcommanddetails secondary withsecondaryerror errors errorstring cluster go tpcc go tpcc go test runner go monitor failure unexpected node event dead exit status attached stack trace stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runtpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main registertpcc home agent work go src github com cockroachdb cockroach pkg cmd roachtest tpcc go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s wraps monitor failure wraps unexpected node event dead exit status error types withstack withstack errutil withprefix errors errorstring reproduce to reproduce try bash from perhaps edited lightly caffeinate roachstress sh tpcc w nodes chaos true cc cockroachdb kv
| 1
|
104,970
| 9,013,378,685
|
IssuesEvent
|
2019-02-05 19:20:42
|
elastic/elasticsearch
|
https://api.github.com/repos/elastic/elasticsearch
|
closed
|
DateMathExpressionResolverTests fails on master
|
:Core/Infra/Core >test-failure
|
Three tests have started failing over the weekend, I believe it has to do with the end of the year approaching. Not sure if the failures may be related to datetime changes. They happen only in master:
```
org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests testExpression_CustomFormat
org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests testExpression_CustomTimeZoneInIndexName
org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests testExpression_EscapeDateFormat
```
https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+periodic/447/console
They easily reproduce:
```
06:36:03 2> REPRODUCE WITH: ./gradlew :server:unitTest -Dtests.seed=57AEC52CB025A748 -Dtests.class=org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests -Dtests.method="testExpression_EscapeDateFormat" -Dtests.security.manager=true -Dtests.locale=da-DK -Dtests.timezone=Indian/Mayotte -Dcompiler.java=11 -Druntime.java=8
06:36:03 2> REPRODUCE WITH: ./gradlew :server:unitTest -Dtests.seed=57AEC52CB025A748 -Dtests.class=org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests -Dtests.method="testExpression_CustomFormat" -Dtests.security.manager=true -Dtests.locale=da-DK -Dtests.timezone=Indian/Mayotte -Dcompiler.java=11 -Druntime.java=8
06:36:03 2> REPRODUCE WITH: ./gradlew :server:unitTest -Dtests.seed=57AEC52CB025A748 -Dtests.class=org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests -Dtests.method="testExpression_CustomTimeZoneInIndexName" -Dtests.security.manager=true -Dtests.locale=da-DK -Dtests.timezone=Indian/Mayotte -Dcompiler.java=11 -Druntime.java=8
```
```
06:36:03 FAILURE 0.02s J5 | DateMathExpressionResolverTests.testExpression_EscapeDateFormat <<< FAILURES!
06:36:03 > Throwable #1: java.lang.AssertionError:
06:36:03 > Expected: ".marvel-{year}2018"
06:36:03 > but: was ".marvel-{year}2019"
06:36:03 > at __randomizedtesting.SeedInfo.seed([57AEC52CB025A748:C795FA0EBA4BCE4B]:0)
06:36:03 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
06:36:03 > at org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests.testExpression_EscapeDateFormat(DateMathExpressionResolverTests.java:111)
06:36:03 FAILURE 0.00s J5 | DateMathExpressionResolverTests.testExpression_CustomFormat <<< FAILURES!
06:36:03 > Throwable #1: java.lang.AssertionError:
06:36:03 > Expected: ".marvel-2018.12.30"
06:36:03 > but: was ".marvel-2019.12.30"
06:36:03 > at __randomizedtesting.SeedInfo.seed([57AEC52CB025A748:6DA2B3B50CD96D0C]:0)
06:36:03 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
06:36:03 > at org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests.testExpression_CustomFormat(DateMathExpressionResolverTests.java:97)
06:36:03 FAILURE 0.01s J5 | DateMathExpressionResolverTests.testExpression_CustomTimeZoneInIndexName <<< FAILURES!
06:36:03 > Throwable #1: java.lang.AssertionError:
06:36:03 > Expected: ".marvel-2018.12.30"
06:36:03 > but: was ".marvel-2019.12.30"
06:36:03 > at __randomizedtesting.SeedInfo.seed([57AEC52CB025A748:C7B07BB86A595B07]:0)
06:36:03 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
06:36:03 > at org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests.testExpression_CustomTimeZoneInIndexName(DateMathExpressionResolverTests.java:153)
```
|
1.0
|
DateMathExpressionResolverTests fails on master - Three tests have started failing over the weekend, I believe it has to do with the end of the year approaching. Not sure if the failures may be related to datetime changes. They happen only in master:
```
org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests testExpression_CustomFormat
org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests testExpression_CustomTimeZoneInIndexName
org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests testExpression_EscapeDateFormat
```
https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+master+periodic/447/console
They easily reproduce:
```
06:36:03 2> REPRODUCE WITH: ./gradlew :server:unitTest -Dtests.seed=57AEC52CB025A748 -Dtests.class=org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests -Dtests.method="testExpression_EscapeDateFormat" -Dtests.security.manager=true -Dtests.locale=da-DK -Dtests.timezone=Indian/Mayotte -Dcompiler.java=11 -Druntime.java=8
06:36:03 2> REPRODUCE WITH: ./gradlew :server:unitTest -Dtests.seed=57AEC52CB025A748 -Dtests.class=org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests -Dtests.method="testExpression_CustomFormat" -Dtests.security.manager=true -Dtests.locale=da-DK -Dtests.timezone=Indian/Mayotte -Dcompiler.java=11 -Druntime.java=8
06:36:03 2> REPRODUCE WITH: ./gradlew :server:unitTest -Dtests.seed=57AEC52CB025A748 -Dtests.class=org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests -Dtests.method="testExpression_CustomTimeZoneInIndexName" -Dtests.security.manager=true -Dtests.locale=da-DK -Dtests.timezone=Indian/Mayotte -Dcompiler.java=11 -Druntime.java=8
```
```
06:36:03 FAILURE 0.02s J5 | DateMathExpressionResolverTests.testExpression_EscapeDateFormat <<< FAILURES!
06:36:03 > Throwable #1: java.lang.AssertionError:
06:36:03 > Expected: ".marvel-{year}2018"
06:36:03 > but: was ".marvel-{year}2019"
06:36:03 > at __randomizedtesting.SeedInfo.seed([57AEC52CB025A748:C795FA0EBA4BCE4B]:0)
06:36:03 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
06:36:03 > at org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests.testExpression_EscapeDateFormat(DateMathExpressionResolverTests.java:111)
06:36:03 FAILURE 0.00s J5 | DateMathExpressionResolverTests.testExpression_CustomFormat <<< FAILURES!
06:36:03 > Throwable #1: java.lang.AssertionError:
06:36:03 > Expected: ".marvel-2018.12.30"
06:36:03 > but: was ".marvel-2019.12.30"
06:36:03 > at __randomizedtesting.SeedInfo.seed([57AEC52CB025A748:6DA2B3B50CD96D0C]:0)
06:36:03 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
06:36:03 > at org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests.testExpression_CustomFormat(DateMathExpressionResolverTests.java:97)
06:36:03 FAILURE 0.01s J5 | DateMathExpressionResolverTests.testExpression_CustomTimeZoneInIndexName <<< FAILURES!
06:36:03 > Throwable #1: java.lang.AssertionError:
06:36:03 > Expected: ".marvel-2018.12.30"
06:36:03 > but: was ".marvel-2019.12.30"
06:36:03 > at __randomizedtesting.SeedInfo.seed([57AEC52CB025A748:C7B07BB86A595B07]:0)
06:36:03 > at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
06:36:03 > at org.elasticsearch.cluster.metadata.DateMathExpressionResolverTests.testExpression_CustomTimeZoneInIndexName(DateMathExpressionResolverTests.java:153)
```
|
test
|
datemathexpressionresolvertests fails on master three tests have started failing over the weekend i believe it has to do with the end of the year approaching not sure if the failures may be related to datetime changes they happen only in master org elasticsearch cluster metadata datemathexpressionresolvertests testexpression customformat org elasticsearch cluster metadata datemathexpressionresolvertests testexpression customtimezoneinindexname org elasticsearch cluster metadata datemathexpressionresolvertests testexpression escapedateformat they easily reproduce reproduce with gradlew server unittest dtests seed dtests class org elasticsearch cluster metadata datemathexpressionresolvertests dtests method testexpression escapedateformat dtests security manager true dtests locale da dk dtests timezone indian mayotte dcompiler java druntime java reproduce with gradlew server unittest dtests seed dtests class org elasticsearch cluster metadata datemathexpressionresolvertests dtests method testexpression customformat dtests security manager true dtests locale da dk dtests timezone indian mayotte dcompiler java druntime java reproduce with gradlew server unittest dtests seed dtests class org elasticsearch cluster metadata datemathexpressionresolvertests dtests method testexpression customtimezoneinindexname dtests security manager true dtests locale da dk dtests timezone indian mayotte dcompiler java druntime java failure datemathexpressionresolvertests testexpression escapedateformat failures throwable java lang assertionerror expected marvel year but was marvel year at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch cluster metadata datemathexpressionresolvertests testexpression escapedateformat datemathexpressionresolvertests java failure datemathexpressionresolvertests testexpression customformat failures throwable java lang assertionerror expected marvel but was marvel at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch cluster metadata datemathexpressionresolvertests testexpression customformat datemathexpressionresolvertests java failure datemathexpressionresolvertests testexpression customtimezoneinindexname failures throwable java lang assertionerror expected marvel but was marvel at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org elasticsearch cluster metadata datemathexpressionresolvertests testexpression customtimezoneinindexname datemathexpressionresolvertests java
| 1
|
199,522
| 15,046,434,732
|
IssuesEvent
|
2021-02-03 07:23:04
|
celo-org/celo-monorepo
|
https://api.github.com/repos/celo-org/celo-monorepo
|
opened
|
[FLAKEY TEST] cli-test -> cli -> account metadata cmds -> Modifying the metadata file -> account:create-metadata cmd
|
FLAKEY cli cli-test
|
Discovered at commit 8b36d4953b957afbad2a67384d3dbf1a2dc438c7
Attempt No. 1:
Error: thrown: "Exceeded timeout of 10000 ms for a test.
Use jest.setTimeout(newTimeout) to increase the timeout value, if this is a long-running test."
at describe (/home/circleci/app/packages/cli/src/commands/account/claims.test.ts:43:9)
at _dispatchDescribe (/home/circleci/app/node_modules/jest-circus/build/index.js:67:26)
at describe (/home/circleci/app/node_modules/jest-circus/build/index.js:30:5)
at Object.<anonymous>.ganache_test_1.testWithGanache (/home/circleci/app/packages/cli/src/commands/account/claims.test.ts:35:5)
at /home/circleci/app/packages/dev-utils/src/ganache-test.ts:79:5
at _dispatchDescribe (/home/circleci/app/node_modules/jest-circus/build/index.js:67:26)
at describe (/home/circleci/app/node_modules/jest-circus/build/index.js:30:5)
at Object.testWithGanache (/home/circleci/app/packages/dev-utils/src/ganache-test.ts:63:3)
at Object.<anonymous> (/home/circleci/app/packages/cli/src/commands/account/claims.test.ts:26:16)
at Runtime._execModule (/home/circleci/app/node_modules/jest-cli/node_modules/jest-runtime/build/index.js:1277:24)
at Runtime._loadModule (/home/circleci/app/node_modules/jest-cli/node_modules/jest-runtime/build/index.js:876:12)
at Runtime.requireModule (/home/circleci/app/node_modules/jest-cli/node_modules/jest-runtime/build/index.js:726:10)
at jestAdapter (/home/circleci/app/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:106:13)
Attempt No. 2:
Test Passed!
|
1.0
|
[FLAKEY TEST] cli-test -> cli -> account metadata cmds -> Modifying the metadata file -> account:create-metadata cmd - Discovered at commit 8b36d4953b957afbad2a67384d3dbf1a2dc438c7
Attempt No. 1:
Error: thrown: "Exceeded timeout of 10000 ms for a test.
Use jest.setTimeout(newTimeout) to increase the timeout value, if this is a long-running test."
at describe (/home/circleci/app/packages/cli/src/commands/account/claims.test.ts:43:9)
at _dispatchDescribe (/home/circleci/app/node_modules/jest-circus/build/index.js:67:26)
at describe (/home/circleci/app/node_modules/jest-circus/build/index.js:30:5)
at Object.<anonymous>.ganache_test_1.testWithGanache (/home/circleci/app/packages/cli/src/commands/account/claims.test.ts:35:5)
at /home/circleci/app/packages/dev-utils/src/ganache-test.ts:79:5
at _dispatchDescribe (/home/circleci/app/node_modules/jest-circus/build/index.js:67:26)
at describe (/home/circleci/app/node_modules/jest-circus/build/index.js:30:5)
at Object.testWithGanache (/home/circleci/app/packages/dev-utils/src/ganache-test.ts:63:3)
at Object.<anonymous> (/home/circleci/app/packages/cli/src/commands/account/claims.test.ts:26:16)
at Runtime._execModule (/home/circleci/app/node_modules/jest-cli/node_modules/jest-runtime/build/index.js:1277:24)
at Runtime._loadModule (/home/circleci/app/node_modules/jest-cli/node_modules/jest-runtime/build/index.js:876:12)
at Runtime.requireModule (/home/circleci/app/node_modules/jest-cli/node_modules/jest-runtime/build/index.js:726:10)
at jestAdapter (/home/circleci/app/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:106:13)
Attempt No. 2:
Test Passed!
|
test
|
cli test cli account metadata cmds modifying the metadata file account create metadata cmd discovered at commit attempt no error thrown exceeded timeout of ms for a test use jest settimeout newtimeout to increase the timeout value if this is a long running test at describe home circleci app packages cli src commands account claims test ts at dispatchdescribe home circleci app node modules jest circus build index js at describe home circleci app node modules jest circus build index js at object ganache test testwithganache home circleci app packages cli src commands account claims test ts at home circleci app packages dev utils src ganache test ts at dispatchdescribe home circleci app node modules jest circus build index js at describe home circleci app node modules jest circus build index js at object testwithganache home circleci app packages dev utils src ganache test ts at object home circleci app packages cli src commands account claims test ts at runtime execmodule home circleci app node modules jest cli node modules jest runtime build index js at runtime loadmodule home circleci app node modules jest cli node modules jest runtime build index js at runtime requiremodule home circleci app node modules jest cli node modules jest runtime build index js at jestadapter home circleci app node modules jest circus build legacy code todo rewrite jestadapter js attempt no test passed
| 1
|
90,796
| 8,272,319,541
|
IssuesEvent
|
2018-09-16 18:53:59
|
hyperledger/composer
|
https://api.github.com/repos/hyperledger/composer
|
closed
|
New BusCard round trip scenarios required
|
integration test playground protractor qa top10
|
With the inbound changes relating to business network deployment, it will be necessary to modify the existing e2e test that tests the integration with the runtime/fabric/playground.
This is because is will not be possible to use the existing inclusion of an npmrc file in the runtime package
## Required Scenarios
- [ ] Import of card with secret (from CLI to Playground)
- [ ] Import of card with certificates (from CLI to Playground)
- [ ] Export of card with secret (from Playground to CLI)
- [ ] Export of card with certificates (from Playground to CLI)
- [ ] Use non-local card store (redis container?)
## Suggested Test Path
1. Import of card with secret
- Existing test. Import a created PeerAdmin card for use within Playground
2. Import of card with certificates
- Pre req for subsequent tests
- Deploy busnet and create busnet admin card.
- Ping BusNet with card to enrol
- Import card within Playground and log into BusNet with card
3. Export of card with secret
- Log into busnet with card (as above)
- Create Participant
- Bind to identity
- Export card (do not use!)
- Import with CLI
- Ping network with card
4. Export of card with certificates
- As above but ensure card has been used within Playground prior to export.
|
1.0
|
New BusCard round trip scenarios required - With the inbound changes relating to business network deployment, it will be necessary to modify the existing e2e test that tests the integration with the runtime/fabric/playground.
This is because is will not be possible to use the existing inclusion of an npmrc file in the runtime package
## Required Scenarios
- [ ] Import of card with secret (from CLI to Playground)
- [ ] Import of card with certificates (from CLI to Playground)
- [ ] Export of card with secret (from Playground to CLI)
- [ ] Export of card with certificates (from Playground to CLI)
- [ ] Use non-local card store (redis container?)
## Suggested Test Path
1. Import of card with secret
- Existing test. Import a created PeerAdmin card for use within Playground
2. Import of card with certificates
- Pre req for subsequent tests
- Deploy busnet and create busnet admin card.
- Ping BusNet with card to enrol
- Import card within Playground and log into BusNet with card
3. Export of card with secret
- Log into busnet with card (as above)
- Create Participant
- Bind to identity
- Export card (do not use!)
- Import with CLI
- Ping network with card
4. Export of card with certificates
- As above but ensure card has been used within Playground prior to export.
|
test
|
new buscard round trip scenarios required with the inbound changes relating to business network deployment it will be necessary to modify the existing test that tests the integration with the runtime fabric playground this is because is will not be possible to use the existing inclusion of an npmrc file in the runtime package required scenarios import of card with secret from cli to playground import of card with certificates from cli to playground export of card with secret from playground to cli export of card with certificates from playground to cli use non local card store redis container suggested test path import of card with secret existing test import a created peeradmin card for use within playground import of card with certificates pre req for subsequent tests deploy busnet and create busnet admin card ping busnet with card to enrol import card within playground and log into busnet with card export of card with secret log into busnet with card as above create participant bind to identity export card do not use import with cli ping network with card export of card with certificates as above but ensure card has been used within playground prior to export
| 1
|
181,000
| 14,849,358,170
|
IssuesEvent
|
2021-01-18 00:58:45
|
davtorcue/decide
|
https://api.github.com/repos/davtorcue/decide
|
closed
|
Video Presentación Decide
|
Accepted New documentation priority:high rol: ALL
|
Video para presentar el proyecto Decide:
- Que cambios hemos hecho.
- Que herramientas hemos utilizado.
- Como hemos utilizado dichas herramientas.
- Ejemplos.
|
1.0
|
Video Presentación Decide - Video para presentar el proyecto Decide:
- Que cambios hemos hecho.
- Que herramientas hemos utilizado.
- Como hemos utilizado dichas herramientas.
- Ejemplos.
|
non_test
|
video presentación decide video para presentar el proyecto decide que cambios hemos hecho que herramientas hemos utilizado como hemos utilizado dichas herramientas ejemplos
| 0
|
168,387
| 13,082,840,495
|
IssuesEvent
|
2020-08-01 15:50:17
|
ForgottenGlory/Living-Skyrim-2
|
https://api.github.com/repos/ForgottenGlory/Living-Skyrim-2
|
closed
|
vampire lord cape inside body
|
bug help wanted need testers question
|
**If you are reporting a crash to desktop, please attach your NET Script Framework crash log. This can be found in MO2's Overwrite folder.
If possible, please also attach a copy of your most recent save before the issue occurred.**
Please check back on your bug request periodically. I will ask for more details sometimes and if they are not provided I will eventually close the bug report.
**LS Version**
2.02
**Describe the bug**
the cape is in the vampire lords body or just mostly gone
**To Reproduce**
become vampire lord get the very top perk sais equip royal crown
**Expected behavior**
cape to be fully there
**Screenshots**
If applicable, add screenshots to help explain your problem. **IF YOUR SCREENSHOT CONTAINS NUDITY DO NOT PUT IT ON GITHUB.** Additionally, if your report applies to a specific object, make sure it is selected with the console open. Reports without this will be closed.

**Additional context**
Light has the same issue
|
1.0
|
vampire lord cape inside body - **If you are reporting a crash to desktop, please attach your NET Script Framework crash log. This can be found in MO2's Overwrite folder.
If possible, please also attach a copy of your most recent save before the issue occurred.**
Please check back on your bug request periodically. I will ask for more details sometimes and if they are not provided I will eventually close the bug report.
**LS Version**
2.02
**Describe the bug**
the cape is in the vampire lords body or just mostly gone
**To Reproduce**
become vampire lord get the very top perk sais equip royal crown
**Expected behavior**
cape to be fully there
**Screenshots**
If applicable, add screenshots to help explain your problem. **IF YOUR SCREENSHOT CONTAINS NUDITY DO NOT PUT IT ON GITHUB.** Additionally, if your report applies to a specific object, make sure it is selected with the console open. Reports without this will be closed.

**Additional context**
Light has the same issue
|
test
|
vampire lord cape inside body if you are reporting a crash to desktop please attach your net script framework crash log this can be found in s overwrite folder if possible please also attach a copy of your most recent save before the issue occurred please check back on your bug request periodically i will ask for more details sometimes and if they are not provided i will eventually close the bug report ls version describe the bug the cape is in the vampire lords body or just mostly gone to reproduce become vampire lord get the very top perk sais equip royal crown expected behavior cape to be fully there screenshots if applicable add screenshots to help explain your problem if your screenshot contains nudity do not put it on github additionally if your report applies to a specific object make sure it is selected with the console open reports without this will be closed additional context light has the same issue
| 1
|
149,457
| 13,281,628,955
|
IssuesEvent
|
2020-08-23 18:18:28
|
No-Budget-Science-Hack-Week-2020/miguezometro
|
https://api.github.com/repos/No-Budget-Science-Hack-Week-2020/miguezometro
|
closed
|
Transferir documentação do Google Docs para documentação permanente
|
documentation help wanted nbs hack week 2020
|
Uma coisa importante é deixar as contribuições e discussões de vocês disponíveis para a posteridade.
Uma opção é colocar aqui no GitHub mesmo, seja na Wiki do repositório, seja como documentos de texto.
Exemplos de coisas boas a se documentar são:
- [x] Referências da literatura utilizadas
- [x] Atas e discussões de reuniões
- [x] Etapas para o futuro (para vocês ou pessoas aleatórias que queiram dar continuidade)
- [x] Produtos feitos (questionário? add on? )
- [x] Dificuldades encontradas
|
1.0
|
Transferir documentação do Google Docs para documentação permanente - Uma coisa importante é deixar as contribuições e discussões de vocês disponíveis para a posteridade.
Uma opção é colocar aqui no GitHub mesmo, seja na Wiki do repositório, seja como documentos de texto.
Exemplos de coisas boas a se documentar são:
- [x] Referências da literatura utilizadas
- [x] Atas e discussões de reuniões
- [x] Etapas para o futuro (para vocês ou pessoas aleatórias que queiram dar continuidade)
- [x] Produtos feitos (questionário? add on? )
- [x] Dificuldades encontradas
|
non_test
|
transferir documentação do google docs para documentação permanente uma coisa importante é deixar as contribuições e discussões de vocês disponíveis para a posteridade uma opção é colocar aqui no github mesmo seja na wiki do repositório seja como documentos de texto exemplos de coisas boas a se documentar são referências da literatura utilizadas atas e discussões de reuniões etapas para o futuro para vocês ou pessoas aleatórias que queiram dar continuidade produtos feitos questionário add on dificuldades encontradas
| 0
|
173,792
| 13,443,846,713
|
IssuesEvent
|
2020-09-08 08:57:19
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Consider incorporating cbor official spec suite
|
area-System.Security test enhancement
|
CBOR has an official spec test suite which validates handful of corner cases: https://github.com/cbor/test-vectors.
In order to fully incorporate that test data as is in `System.Formats.Cbor`'s test project, it would require implementing RFC 7049 [<kbd>**§6 - Diagnostic Notation**</kbd>](https://tools.ietf.org/html/rfc7049#section-6). Diagnostic Notation API will also help comparing the dotnet's cbor implementation with that of other languages which provide APIs for diagnostic notation.
Playground for diagnostics notation and byte-by-byte explanation: http://cbor.me/
Playground's source code in Ruby: [diagnostics notation](https://github.com/cabo/cbor-diag/blob/0a38a7c926e71e1ac6c73a3e15302a9bf9d31962/lib/cbor-diagnostic.rb) and [byte-by-byte explanation](https://github.com/cabo/cbor-diag/blob/0a38a7c926e71e1ac6c73a3e15302a9bf9d31962/lib/cbor-pretty.rb)
|
1.0
|
Consider incorporating cbor official spec suite - CBOR has an official spec test suite which validates handful of corner cases: https://github.com/cbor/test-vectors.
In order to fully incorporate that test data as is in `System.Formats.Cbor`'s test project, it would require implementing RFC 7049 [<kbd>**§6 - Diagnostic Notation**</kbd>](https://tools.ietf.org/html/rfc7049#section-6). Diagnostic Notation API will also help comparing the dotnet's cbor implementation with that of other languages which provide APIs for diagnostic notation.
Playground for diagnostics notation and byte-by-byte explanation: http://cbor.me/
Playground's source code in Ruby: [diagnostics notation](https://github.com/cabo/cbor-diag/blob/0a38a7c926e71e1ac6c73a3e15302a9bf9d31962/lib/cbor-diagnostic.rb) and [byte-by-byte explanation](https://github.com/cabo/cbor-diag/blob/0a38a7c926e71e1ac6c73a3e15302a9bf9d31962/lib/cbor-pretty.rb)
|
test
|
consider incorporating cbor official spec suite cbor has an official spec test suite which validates handful of corner cases in order to fully incorporate that test data as is in system formats cbor s test project it would require implementing rfc diagnostic notation api will also help comparing the dotnet s cbor implementation with that of other languages which provide apis for diagnostic notation playground for diagnostics notation and byte by byte explanation playground s source code in ruby and
| 1
|
20,979
| 6,971,776,529
|
IssuesEvent
|
2017-12-11 15:04:52
|
cnr-ibf-pa/hbp-bsp-issues
|
https://api.github.com/repos/cnr-ibf-pa/hbp-bsp-issues
|
opened
|
Usage of '/home/jupyter' instead `/home/jovyan`
|
Critical Type_BUG UC_HippoCellModel_Rebuild
|
### Expected behavior
No error
### Actual Behavior
OSError: [Errno 2] No such file or directory, line 62:
```
<ipython-input-7-cdc45f67112e> in on_w3_clicked(change)
60 outputHTML.layout.display='none'
61 plotly.offline.init_notebook_mode()
---> 62 os.chdir("/home/jupyter/BPOPTANALYSIS")
63 if c1.value=='GitHub':
64 os.chdir(w3.value)
```
### Steps to reproduce the problem
Run the Analysis notebook on storage results.
### Optional infrastructural data (user, platform, browser, environment, ...)
|
1.0
|
Usage of '/home/jupyter' instead `/home/jovyan` - ### Expected behavior
No error
### Actual Behavior
OSError: [Errno 2] No such file or directory, line 62:
```
<ipython-input-7-cdc45f67112e> in on_w3_clicked(change)
60 outputHTML.layout.display='none'
61 plotly.offline.init_notebook_mode()
---> 62 os.chdir("/home/jupyter/BPOPTANALYSIS")
63 if c1.value=='GitHub':
64 os.chdir(w3.value)
```
### Steps to reproduce the problem
Run the Analysis notebook on storage results.
### Optional infrastructural data (user, platform, browser, environment, ...)
|
non_test
|
usage of home jupyter instead home jovyan expected behavior no error actual behavior oserror no such file or directory line in on clicked change outputhtml layout display none plotly offline init notebook mode os chdir home jupyter bpoptanalysis if value github os chdir value steps to reproduce the problem run the analysis notebook on storage results optional infrastructural data user platform browser environment
| 0
|
3,971
| 2,698,759,729
|
IssuesEvent
|
2015-04-03 10:41:43
|
bedita/bedita
|
https://api.github.com/repos/bedita/bedita
|
closed
|
Inconsistency in new Section creation form
|
Module - Publications Priority - Low Status - Test
|
When attempting to create a new Section, there are some minor issues in view that seem to depend on which is the default tree location of the Section being created.
@rzecchini
|
1.0
|
Inconsistency in new Section creation form - When attempting to create a new Section, there are some minor issues in view that seem to depend on which is the default tree location of the Section being created.
@rzecchini
|
test
|
inconsistency in new section creation form when attempting to create a new section there are some minor issues in view that seem to depend on which is the default tree location of the section being created rzecchini
| 1
|
227,933
| 18,111,032,901
|
IssuesEvent
|
2021-09-23 04:01:06
|
E3SM-Project/scream
|
https://api.github.com/repos/E3SM-Project/scream
|
opened
|
Erratic testing beavior on weaver
|
bug testing priority:high
|
There seem to be issues with testing on weaver. Mainly, there are two sub-issues.
1. The nightly testing sometimes reports an internal compiler error, when building `atmosphere_microphysics.cpp`. This behavior has been recorded when running on weaver1 and weaver4 compute nodes, but did not happen on weaver2 (no data for the other ones), although on weaver2 we observed bad initial state for Homme (negative dp/dphi/vtheta in EOS during init). The internal compiler error does *not* happen for AT runs. I am not sure what builds these errors belong to (nightlies jobs use the same folder every time, so I can't go and inspect different builds outputs separately).
2. The AT runs are failing the homme_standalone test in release build on weaver1, due to negative/nan layer thickness in homme, either after 3 or 6 timesteps. On weaver2, on top of the release failure, the same bad initial state as in the nightly test runs is observed for the full_debug build.
For the 2nd sub-issue: when running on weaver1, after the failure occurs, running test-all-scream with the `--quick-rerun` option, or simply running the test manually from the folder, does not yield a failure anymore.
|
1.0
|
Erratic testing beavior on weaver - There seem to be issues with testing on weaver. Mainly, there are two sub-issues.
1. The nightly testing sometimes reports an internal compiler error, when building `atmosphere_microphysics.cpp`. This behavior has been recorded when running on weaver1 and weaver4 compute nodes, but did not happen on weaver2 (no data for the other ones), although on weaver2 we observed bad initial state for Homme (negative dp/dphi/vtheta in EOS during init). The internal compiler error does *not* happen for AT runs. I am not sure what builds these errors belong to (nightlies jobs use the same folder every time, so I can't go and inspect different builds outputs separately).
2. The AT runs are failing the homme_standalone test in release build on weaver1, due to negative/nan layer thickness in homme, either after 3 or 6 timesteps. On weaver2, on top of the release failure, the same bad initial state as in the nightly test runs is observed for the full_debug build.
For the 2nd sub-issue: when running on weaver1, after the failure occurs, running test-all-scream with the `--quick-rerun` option, or simply running the test manually from the folder, does not yield a failure anymore.
|
test
|
erratic testing beavior on weaver there seem to be issues with testing on weaver mainly there are two sub issues the nightly testing sometimes reports an internal compiler error when building atmosphere microphysics cpp this behavior has been recorded when running on and compute nodes but did not happen on no data for the other ones although on we observed bad initial state for homme negative dp dphi vtheta in eos during init the internal compiler error does not happen for at runs i am not sure what builds these errors belong to nightlies jobs use the same folder every time so i can t go and inspect different builds outputs separately the at runs are failing the homme standalone test in release build on due to negative nan layer thickness in homme either after or timesteps on on top of the release failure the same bad initial state as in the nightly test runs is observed for the full debug build for the sub issue when running on after the failure occurs running test all scream with the quick rerun option or simply running the test manually from the folder does not yield a failure anymore
| 1
|
20,734
| 6,923,265,095
|
IssuesEvent
|
2017-11-30 08:18:56
|
spack/spack
|
https://api.github.com/repos/spack/spack
|
closed
|
cgal 4.9.1 checksum wrong.
|
build-error
|
Tried to install a package and the `cgal` chosen by concretisation wasn't installed as the checksum was wrong:
```sh
spack install gcc@7.2.0
spack load gcc@7.2.0
spack compilers find
spack unload gcc@7.2.0
spack install openfoam-com@1706 ^openmpi@3.0 fabrics=psm2 schedulers=sge
...
ChecksumError: ChecksumError: md5 checksum failed for /data/cs1wf/spack/var/spack/stage/cgal-4.9.1-djhtft2maevtiffwtsak5glyhlb7v7o7/CGAL-4.9.1.tar.gz
Expected df6517df3320bf6c9de2e1b0361738b9 but got 2c427da202ea4a6f5c8c8fbc6987022e
```
Using Spack commit `1b3b2627761a1f5a792749ff9d6b9c263f5203f9` on Centos 7.4.
Get the same issue if explicitly requesting cgal 4.9 instead.
|
1.0
|
cgal 4.9.1 checksum wrong. - Tried to install a package and the `cgal` chosen by concretisation wasn't installed as the checksum was wrong:
```sh
spack install gcc@7.2.0
spack load gcc@7.2.0
spack compilers find
spack unload gcc@7.2.0
spack install openfoam-com@1706 ^openmpi@3.0 fabrics=psm2 schedulers=sge
...
ChecksumError: ChecksumError: md5 checksum failed for /data/cs1wf/spack/var/spack/stage/cgal-4.9.1-djhtft2maevtiffwtsak5glyhlb7v7o7/CGAL-4.9.1.tar.gz
Expected df6517df3320bf6c9de2e1b0361738b9 but got 2c427da202ea4a6f5c8c8fbc6987022e
```
Using Spack commit `1b3b2627761a1f5a792749ff9d6b9c263f5203f9` on Centos 7.4.
Get the same issue if explicitly requesting cgal 4.9 instead.
|
non_test
|
cgal checksum wrong tried to install a package and the cgal chosen by concretisation wasn t installed as the checksum was wrong sh spack install gcc spack load gcc spack compilers find spack unload gcc spack install openfoam com openmpi fabrics schedulers sge checksumerror checksumerror checksum failed for data spack var spack stage cgal cgal tar gz expected but got using spack commit on centos get the same issue if explicitly requesting cgal instead
| 0
|
92,242
| 8,356,584,793
|
IssuesEvent
|
2018-10-02 18:55:28
|
brave/brave-browser
|
https://api.github.com/repos/brave/brave-browser
|
closed
|
stats are not updated on private window when default tab stats update
|
QA/Test-Plan-Specified QA/Yes priority/P5
|
## Test Plan
See https://github.com/brave/brave-core/pull/540
## Description
1. open a default window and a private window both on new tab page
2. visit wired.com on the default window
3. watch that stats on the private tab page didn't update
First reported by @simonhong and confirmed by @yrliou.
|
1.0
|
stats are not updated on private window when default tab stats update - ## Test Plan
See https://github.com/brave/brave-core/pull/540
## Description
1. open a default window and a private window both on new tab page
2. visit wired.com on the default window
3. watch that stats on the private tab page didn't update
First reported by @simonhong and confirmed by @yrliou.
|
test
|
stats are not updated on private window when default tab stats update test plan see description open a default window and a private window both on new tab page visit wired com on the default window watch that stats on the private tab page didn t update first reported by simonhong and confirmed by yrliou
| 1
|
151,458
| 12,036,765,331
|
IssuesEvent
|
2020-04-13 20:26:47
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
closed
|
Failing test: X-Pack Jest Tests.x-pack/plugins/endpoint/public/applications/endpoint/store/policy_list - policy list store concerns it sets `isLoading` when `userPaginatedPolicyListTable`
|
Feature:Endpoint Team:Endpoint Data Visibility Team:Endpoint Management Team:Endpoint Response failed-test
|
A test failed on a tracked branch
```
Error: expect(received).toBe(expected) // Object.is equality
Expected: false
Received: true
at Object.test (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/x-pack/plugins/endpoint/public/applications/endpoint/store/policy_list/index.test.ts:52:41)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/3285/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Jest Tests.x-pack/plugins/endpoint/public/applications/endpoint/store/policy_list","test.name":"policy list store concerns it sets `isLoading` when `userPaginatedPolicyListTable`","test.failCount":3}} -->
|
1.0
|
Failing test: X-Pack Jest Tests.x-pack/plugins/endpoint/public/applications/endpoint/store/policy_list - policy list store concerns it sets `isLoading` when `userPaginatedPolicyListTable` - A test failed on a tracked branch
```
Error: expect(received).toBe(expected) // Object.is equality
Expected: false
Received: true
at Object.test (/var/lib/jenkins/workspace/elastic+kibana+master/kibana/x-pack/plugins/endpoint/public/applications/endpoint/store/policy_list/index.test.ts:52:41)
```
First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/3285/)
<!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Jest Tests.x-pack/plugins/endpoint/public/applications/endpoint/store/policy_list","test.name":"policy list store concerns it sets `isLoading` when `userPaginatedPolicyListTable`","test.failCount":3}} -->
|
test
|
failing test x pack jest tests x pack plugins endpoint public applications endpoint store policy list policy list store concerns it sets isloading when userpaginatedpolicylisttable a test failed on a tracked branch error expect received tobe expected object is equality expected false received true at object test var lib jenkins workspace elastic kibana master kibana x pack plugins endpoint public applications endpoint store policy list index test ts first failure
| 1
|
304,450
| 23,066,881,635
|
IssuesEvent
|
2022-07-25 14:35:50
|
NDragneelL9/the-undermine-bot
|
https://api.github.com/repos/NDragneelL9/the-undermine-bot
|
closed
|
Documentation
|
documentation
|
Write SSD documentation.
What should be inside Readme file:
- [x] Table of content
- [x] About application (or goal of the project)
- [x] Requirements (both functional and non-functional)
- [x] Glossary
- [x] Stakeholders
- [x] Design section with diagrams
- [x] Architecture with different viewpoints
- [x] Technology stack
- [x] Code quality and test coverage badges
- [x] Gettind started section
- [x] Contributing section
|
1.0
|
Documentation - Write SSD documentation.
What should be inside Readme file:
- [x] Table of content
- [x] About application (or goal of the project)
- [x] Requirements (both functional and non-functional)
- [x] Glossary
- [x] Stakeholders
- [x] Design section with diagrams
- [x] Architecture with different viewpoints
- [x] Technology stack
- [x] Code quality and test coverage badges
- [x] Gettind started section
- [x] Contributing section
|
non_test
|
documentation write ssd documentation what should be inside readme file table of content about application or goal of the project requirements both functional and non functional glossary stakeholders design section with diagrams architecture with different viewpoints technology stack code quality and test coverage badges gettind started section contributing section
| 0
|
11,594
| 3,211,062,903
|
IssuesEvent
|
2015-10-06 08:38:41
|
CoderDojo/community-platform
|
https://api.github.com/repos/CoderDojo/community-platform
|
closed
|
Show number of events a user has attended
|
backlog Ready For Testing
|
In the manage users section show the number of events the user has attended for that Dojo.
Requested by myself
|
1.0
|
Show number of events a user has attended - In the manage users section show the number of events the user has attended for that Dojo.
Requested by myself
|
test
|
show number of events a user has attended in the manage users section show the number of events the user has attended for that dojo requested by myself
| 1
|
71,052
| 9,477,307,224
|
IssuesEvent
|
2019-04-19 18:09:11
|
golang/go
|
https://api.github.com/repos/golang/go
|
closed
|
cmd/go/internal/modfetch: document known bug in `isVendoredPackage`
|
Documentation NeedsFix Unfortunate modules
|
### What version of Go are you using (`go version`)?
<pre>
$ go version
go version go1.12.1 linux/amd64
</pre>
### Does this issue reproduce with the latest release?
Yes
### What operating system and processor architecture are you using (`go env`)?
<details><summary><code>go env</code> Output</summary><br><pre>
$ go env
GOARCH="amd64"
GOBIN="/home/gheibi/.go/bin/"
GOCACHE="/home/gheibi/.cache/go-build"
GOEXE=""
GOFLAGS=""
GOHOSTARCH="amd64"
GOHOSTOS="linux"
GOOS="linux"
GOPATH="/home/gheibi/.go/"
GOPROXY=""
GORACE=""
GOROOT="/local/go1.12.1.linux_amd64/go"
GOTMPDIR=""
GOTOOLDIR="/local/go1.12.1.linux_amd64/go /pkg/tool/linux_amd64"
GCCGO="gccgo"
CC="gcc"
CXX="g++"
CGO_ENABLED="1"
GOMOD=""
CGO_CFLAGS="-g -O2"
CGO_CPPFLAGS=""
CGO_CXXFLAGS="-g -O2"
CGO_FFLAGS="-g -O2"
CGO_LDFLAGS="-g -O2"
PKG_CONFIG="pkg-config"
GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build865790926=/tmp/go-build"
</pre></details>
### What did you do?
```
$grep -A12 "func isVendoredPackage" $GOROOT/src/cmd/go/internal/modfetch/coderepo.go
```
https://github.com/golang/go/blob/c8aaec2f70c5ccbca1ec2152c57d19981ac09133/src/cmd/go/internal/modfetch/coderepo.go#L632-L642
### What did you expect to see?
```
func isVendoredPackage(name string) bool {
var i int
if strings.HasPrefix(name, "vendor/") {
i += len("vendor/")
} else if j := strings.Index(name, "/vendor/"); j >= 0 {
i += j + len("/vendor/")
} else {
return false
}
return strings.Contains(name[i:], "/")
}
```
### What did you see instead?
The function is not documented, but it is obvious that it is not doing what the intention is.
It seems that the intention is to match against this regex: `(^vendor/|/vendor/).*/`
But in fact it does random match by skipping first 8 bytes in case the input contains `/vendor/`.
The fix is to add the j to the i as states above.
It is quite important to fix it early on, because it may cause go-modules with different hashes and a late fix can cause more hash mismatch of the same content (defy the immutability logic).
|
1.0
|
cmd/go/internal/modfetch: document known bug in `isVendoredPackage` - ### What version of Go are you using (`go version`)?
<pre>
$ go version
go version go1.12.1 linux/amd64
</pre>
### Does this issue reproduce with the latest release?
Yes
### What operating system and processor architecture are you using (`go env`)?
<details><summary><code>go env</code> Output</summary><br><pre>
$ go env
GOARCH="amd64"
GOBIN="/home/gheibi/.go/bin/"
GOCACHE="/home/gheibi/.cache/go-build"
GOEXE=""
GOFLAGS=""
GOHOSTARCH="amd64"
GOHOSTOS="linux"
GOOS="linux"
GOPATH="/home/gheibi/.go/"
GOPROXY=""
GORACE=""
GOROOT="/local/go1.12.1.linux_amd64/go"
GOTMPDIR=""
GOTOOLDIR="/local/go1.12.1.linux_amd64/go /pkg/tool/linux_amd64"
GCCGO="gccgo"
CC="gcc"
CXX="g++"
CGO_ENABLED="1"
GOMOD=""
CGO_CFLAGS="-g -O2"
CGO_CPPFLAGS=""
CGO_CXXFLAGS="-g -O2"
CGO_FFLAGS="-g -O2"
CGO_LDFLAGS="-g -O2"
PKG_CONFIG="pkg-config"
GOGCCFLAGS="-fPIC -m64 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build865790926=/tmp/go-build"
</pre></details>
### What did you do?
```
$grep -A12 "func isVendoredPackage" $GOROOT/src/cmd/go/internal/modfetch/coderepo.go
```
https://github.com/golang/go/blob/c8aaec2f70c5ccbca1ec2152c57d19981ac09133/src/cmd/go/internal/modfetch/coderepo.go#L632-L642
### What did you expect to see?
```
func isVendoredPackage(name string) bool {
var i int
if strings.HasPrefix(name, "vendor/") {
i += len("vendor/")
} else if j := strings.Index(name, "/vendor/"); j >= 0 {
i += j + len("/vendor/")
} else {
return false
}
return strings.Contains(name[i:], "/")
}
```
### What did you see instead?
The function is not documented, but it is obvious that it is not doing what the intention is.
It seems that the intention is to match against this regex: `(^vendor/|/vendor/).*/`
But in fact it does random match by skipping first 8 bytes in case the input contains `/vendor/`.
The fix is to add the j to the i as states above.
It is quite important to fix it early on, because it may cause go-modules with different hashes and a late fix can cause more hash mismatch of the same content (defy the immutability logic).
|
non_test
|
cmd go internal modfetch document known bug in isvendoredpackage what version of go are you using go version go version go version linux does this issue reproduce with the latest release yes what operating system and processor architecture are you using go env go env output go env goarch gobin home gheibi go bin gocache home gheibi cache go build goexe goflags gohostarch gohostos linux goos linux gopath home gheibi go goproxy gorace goroot local linux go gotmpdir gotooldir local linux go pkg tool linux gccgo gccgo cc gcc cxx g cgo enabled gomod cgo cflags g cgo cppflags cgo cxxflags g cgo fflags g cgo ldflags g pkg config pkg config gogccflags fpic pthread fmessage length fdebug prefix map tmp go tmp go build what did you do grep func isvendoredpackage goroot src cmd go internal modfetch coderepo go what did you expect to see func isvendoredpackage name string bool var i int if strings hasprefix name vendor i len vendor else if j strings index name vendor j i j len vendor else return false return strings contains name what did you see instead the function is not documented but it is obvious that it is not doing what the intention is it seems that the intention is to match against this regex vendor vendor but in fact it does random match by skipping first bytes in case the input contains vendor the fix is to add the j to the i as states above it is quite important to fix it early on because it may cause go modules with different hashes and a late fix can cause more hash mismatch of the same content defy the immutability logic
| 0
|
93,703
| 8,442,029,726
|
IssuesEvent
|
2018-10-18 12:06:52
|
Kademi/kademi-dev
|
https://api.github.com/repos/Kademi/kademi-dev
|
closed
|
KCom2: browse by price
|
Ready to Test - Dev bug
|
Not quite understand how "Browse by price" section works
Why it dislpay only 5 products with such a high priceses and ignores those are with lower ones

Also, not quite understand why it show say 490 to 990 instead of 500-1000. Current user have no discounts.
http://vladtest54b.admin.kademi-ci.co/manageUsers/17908808/#summary-tab
|
1.0
|
KCom2: browse by price - Not quite understand how "Browse by price" section works
Why it dislpay only 5 products with such a high priceses and ignores those are with lower ones

Also, not quite understand why it show say 490 to 990 instead of 500-1000. Current user have no discounts.
http://vladtest54b.admin.kademi-ci.co/manageUsers/17908808/#summary-tab
|
test
|
browse by price not quite understand how browse by price section works why it dislpay only products with such a high priceses and ignores those are with lower ones also not quite understand why it show say to instead of current user have no discounts
| 1
|
143,697
| 5,521,861,213
|
IssuesEvent
|
2017-03-19 18:56:01
|
JosefAssad/SeMaWi
|
https://api.github.com/repos/JosefAssad/SeMaWi
|
closed
|
Analyser
|
enhancement high priority balk
|
Vi laver ofte analyser for andre centre med personfølsomme data . Det bliver større krav til dokumentationen til sådanne analyser i fremtiden, så vi overvejer at oprette en analyse kategori.
Props:
Modtager af analyse
|
1.0
|
Analyser - Vi laver ofte analyser for andre centre med personfølsomme data . Det bliver større krav til dokumentationen til sådanne analyser i fremtiden, så vi overvejer at oprette en analyse kategori.
Props:
Modtager af analyse
|
non_test
|
analyser vi laver ofte analyser for andre centre med personfølsomme data det bliver større krav til dokumentationen til sådanne analyser i fremtiden så vi overvejer at oprette en analyse kategori props modtager af analyse
| 0
|
4,604
| 6,725,362,957
|
IssuesEvent
|
2017-10-17 04:54:42
|
jhipster/generator-jhipster
|
https://api.github.com/repos/jhipster/generator-jhipster
|
closed
|
UAA: can't access to administration pages
|
microservice uaa
|
##### **Overview of the issue**
When using UAA option, I can't access to the administration pages
That's why the build for UAA + Protractor tests failed since more than 1 week:
- https://travis-ci.org/hipster-labs/jhipster-travis-build/builds/284099556
- https://travis-ci.org/hipster-labs/jhipster-travis-build/jobs/284099559
It works for 4.9.0, but it's broken in current master.
##### **Reproduce the error**
- generate a UAA project
- generate a gateway
- start the registry
- start UAA and wait
- start the gateway
##### **Related issues**
Similar to this ticket: https://github.com/jhipster/generator-jhipster/issues/6443
##### **Suggest a Fix**
Probably related to Angular part, so I have no idea for the moment, sorry :(
But some solution have been discuted here https://github.com/jhipster/generator-jhipster/issues/6443
##### **JHipster Version(s)**
Master only
Not in the last release 4.9.0
##### **JHipster configuration**
UAA project:
```yaml
{
"generator-jhipster": {
"promptValues": {
"packageName": "com.mycompany.myapp"
},
"jhipsterVersion": "4.9.0",
"baseName": "uaa",
"packageName": "com.mycompany.myapp",
"packageFolder": "com/mycompany/myapp",
"serverPort": "9999",
"authenticationType": "uaa",
"hibernateCache": "hazelcast",
"clusteredHttpSession": false,
"websocket": false,
"databaseType": "sql",
"devDatabaseType": "h2Disk",
"prodDatabaseType": "mysql",
"searchEngine": false,
"messageBroker": false,
"serviceDiscoveryType": "eureka",
"buildTool": "maven",
"enableSocialSignIn": false,
"enableSwaggerCodegen": false,
"jwtSecretKey": "bed31ce04f0071dd3700faff9e94b2131a0588d1",
"enableTranslation": false,
"applicationType": "uaa",
"testFrameworks": [],
"jhiPrefix": "jhi",
"skipClient": true,
"clientPackageManager": "yarn"
}
}
```
Gateway project:
```yaml
{
"generator-jhipster": {
"promptValues": {
"packageName": "com.mycompany.myapp"
},
"jhipsterVersion": "4.9.0",
"baseName": "gateway",
"packageName": "com.mycompany.myapp",
"packageFolder": "com/mycompany/myapp",
"serverPort": "8080",
"authenticationType": "uaa",
"uaaBaseName": "uaa",
"hibernateCache": "hazelcast",
"clusteredHttpSession": false,
"websocket": false,
"databaseType": "sql",
"devDatabaseType": "h2Disk",
"prodDatabaseType": "mysql",
"searchEngine": false,
"messageBroker": false,
"serviceDiscoveryType": "eureka",
"buildTool": "maven",
"enableSocialSignIn": false,
"enableSwaggerCodegen": false,
"clientFramework": "angularX",
"useSass": false,
"clientPackageManager": "yarn",
"applicationType": "gateway",
"testFrameworks": [
"protractor"
],
"jhiPrefix": "jhi",
"enableTranslation": false
}
}
```
##### **Entity configuration(s) `entityName.json` files generated in the `.jhipster` directory**
<!--
If the error is during an entity creation or associated with a specific entity.
If you are using JDL, please share that configuration as well.
-->
##### **Environment and Tools**
java version "1.8.0_131"
Java(TM) SE Runtime Environment (build 1.8.0_131-b11)
Java HotSpot(TM) 64-Bit Server VM (build 25.131-b11, mixed mode)
git version 2.13.0
node: v6.11.1
npm: 5.4.2
bower: 1.8.0
gulp:
[01:22:05] CLI version 1.2.2
yeoman: 1.8.5
yarn: 1.1.0
Docker version 17.06.0-ce, build 02c1d87
docker-compose version 1.16.1, build 6d1ac21
Execution complete
- [x] Checking this box is mandatory (this is just to show you read everything)
|
1.0
|
UAA: can't access to administration pages - ##### **Overview of the issue**
When using UAA option, I can't access to the administration pages
That's why the build for UAA + Protractor tests failed since more than 1 week:
- https://travis-ci.org/hipster-labs/jhipster-travis-build/builds/284099556
- https://travis-ci.org/hipster-labs/jhipster-travis-build/jobs/284099559
It works for 4.9.0, but it's broken in current master.
##### **Reproduce the error**
- generate a UAA project
- generate a gateway
- start the registry
- start UAA and wait
- start the gateway
##### **Related issues**
Similar to this ticket: https://github.com/jhipster/generator-jhipster/issues/6443
##### **Suggest a Fix**
Probably related to Angular part, so I have no idea for the moment, sorry :(
But some solution have been discuted here https://github.com/jhipster/generator-jhipster/issues/6443
##### **JHipster Version(s)**
Master only
Not in the last release 4.9.0
##### **JHipster configuration**
UAA project:
```yaml
{
"generator-jhipster": {
"promptValues": {
"packageName": "com.mycompany.myapp"
},
"jhipsterVersion": "4.9.0",
"baseName": "uaa",
"packageName": "com.mycompany.myapp",
"packageFolder": "com/mycompany/myapp",
"serverPort": "9999",
"authenticationType": "uaa",
"hibernateCache": "hazelcast",
"clusteredHttpSession": false,
"websocket": false,
"databaseType": "sql",
"devDatabaseType": "h2Disk",
"prodDatabaseType": "mysql",
"searchEngine": false,
"messageBroker": false,
"serviceDiscoveryType": "eureka",
"buildTool": "maven",
"enableSocialSignIn": false,
"enableSwaggerCodegen": false,
"jwtSecretKey": "bed31ce04f0071dd3700faff9e94b2131a0588d1",
"enableTranslation": false,
"applicationType": "uaa",
"testFrameworks": [],
"jhiPrefix": "jhi",
"skipClient": true,
"clientPackageManager": "yarn"
}
}
```
Gateway project:
```yaml
{
"generator-jhipster": {
"promptValues": {
"packageName": "com.mycompany.myapp"
},
"jhipsterVersion": "4.9.0",
"baseName": "gateway",
"packageName": "com.mycompany.myapp",
"packageFolder": "com/mycompany/myapp",
"serverPort": "8080",
"authenticationType": "uaa",
"uaaBaseName": "uaa",
"hibernateCache": "hazelcast",
"clusteredHttpSession": false,
"websocket": false,
"databaseType": "sql",
"devDatabaseType": "h2Disk",
"prodDatabaseType": "mysql",
"searchEngine": false,
"messageBroker": false,
"serviceDiscoveryType": "eureka",
"buildTool": "maven",
"enableSocialSignIn": false,
"enableSwaggerCodegen": false,
"clientFramework": "angularX",
"useSass": false,
"clientPackageManager": "yarn",
"applicationType": "gateway",
"testFrameworks": [
"protractor"
],
"jhiPrefix": "jhi",
"enableTranslation": false
}
}
```
##### **Entity configuration(s) `entityName.json` files generated in the `.jhipster` directory**
<!--
If the error is during an entity creation or associated with a specific entity.
If you are using JDL, please share that configuration as well.
-->
##### **Environment and Tools**
java version "1.8.0_131"
Java(TM) SE Runtime Environment (build 1.8.0_131-b11)
Java HotSpot(TM) 64-Bit Server VM (build 25.131-b11, mixed mode)
git version 2.13.0
node: v6.11.1
npm: 5.4.2
bower: 1.8.0
gulp:
[01:22:05] CLI version 1.2.2
yeoman: 1.8.5
yarn: 1.1.0
Docker version 17.06.0-ce, build 02c1d87
docker-compose version 1.16.1, build 6d1ac21
Execution complete
- [x] Checking this box is mandatory (this is just to show you read everything)
|
non_test
|
uaa can t access to administration pages overview of the issue when using uaa option i can t access to the administration pages that s why the build for uaa protractor tests failed since more than week it works for but it s broken in current master reproduce the error generate a uaa project generate a gateway start the registry start uaa and wait start the gateway related issues similar to this ticket suggest a fix probably related to angular part so i have no idea for the moment sorry but some solution have been discuted here jhipster version s master only not in the last release jhipster configuration uaa project yaml generator jhipster promptvalues packagename com mycompany myapp jhipsterversion basename uaa packagename com mycompany myapp packagefolder com mycompany myapp serverport authenticationtype uaa hibernatecache hazelcast clusteredhttpsession false websocket false databasetype sql devdatabasetype proddatabasetype mysql searchengine false messagebroker false servicediscoverytype eureka buildtool maven enablesocialsignin false enableswaggercodegen false jwtsecretkey enabletranslation false applicationtype uaa testframeworks jhiprefix jhi skipclient true clientpackagemanager yarn gateway project yaml generator jhipster promptvalues packagename com mycompany myapp jhipsterversion basename gateway packagename com mycompany myapp packagefolder com mycompany myapp serverport authenticationtype uaa uaabasename uaa hibernatecache hazelcast clusteredhttpsession false websocket false databasetype sql devdatabasetype proddatabasetype mysql searchengine false messagebroker false servicediscoverytype eureka buildtool maven enablesocialsignin false enableswaggercodegen false clientframework angularx usesass false clientpackagemanager yarn applicationtype gateway testframeworks protractor jhiprefix jhi enabletranslation false entity configuration s entityname json files generated in the jhipster directory if the error is during an entity creation or associated with a specific entity if you are using jdl please share that configuration as well environment and tools java version java tm se runtime environment build java hotspot tm bit server vm build mixed mode git version node npm bower gulp cli version yeoman yarn docker version ce build docker compose version build execution complete checking this box is mandatory this is just to show you read everything
| 0
|
288,451
| 21,705,753,106
|
IssuesEvent
|
2022-05-10 09:26:15
|
meugenom/markdown-ts-compiler
|
https://api.github.com/repos/meugenom/markdown-ts-compiler
|
closed
|
Design as an external module
|
documentation
|
It's very important to write code as an external module with index.html as an entry point and src.js
We need research about external modules
|
1.0
|
Design as an external module - It's very important to write code as an external module with index.html as an entry point and src.js
We need research about external modules
|
non_test
|
design as an external module it s very important to write code as an external module with index html as an entry point and src js we need research about external modules
| 0
|
111,385
| 9,529,560,002
|
IssuesEvent
|
2019-04-29 11:40:26
|
rust-lang/rust
|
https://api.github.com/repos/rust-lang/rust
|
closed
|
the test(s) from issue 25579 should be ported to NLL run-pass tests
|
E-needstest NLL-fixed-by-NLL
|
The work on reviewing the diagnostic differences between AST-borrowck and NLL leads me to conclude that NLL is (probably) correctly accepting the regression test we added for #25579
Namely this:
https://github.com/rust-lang/rust/blob/653da4fd006c97625247acd7e076d0782cdc149b/src/test/ui/issues/issue-25579.rs#L16-L28
We should make the original test be AST-borrowck only (via `-Z borrowck=ast`) and make a new version of the test (and perhaps the original code from the description on #25579) that is NLL-only and goes in run-pass.
|
1.0
|
the test(s) from issue 25579 should be ported to NLL run-pass tests - The work on reviewing the diagnostic differences between AST-borrowck and NLL leads me to conclude that NLL is (probably) correctly accepting the regression test we added for #25579
Namely this:
https://github.com/rust-lang/rust/blob/653da4fd006c97625247acd7e076d0782cdc149b/src/test/ui/issues/issue-25579.rs#L16-L28
We should make the original test be AST-borrowck only (via `-Z borrowck=ast`) and make a new version of the test (and perhaps the original code from the description on #25579) that is NLL-only and goes in run-pass.
|
test
|
the test s from issue should be ported to nll run pass tests the work on reviewing the diagnostic differences between ast borrowck and nll leads me to conclude that nll is probably correctly accepting the regression test we added for namely this we should make the original test be ast borrowck only via z borrowck ast and make a new version of the test and perhaps the original code from the description on that is nll only and goes in run pass
| 1
|
55,525
| 23,493,560,261
|
IssuesEvent
|
2022-08-17 21:25:53
|
quesst-technologies/qst-admin-status-all
|
https://api.github.com/repos/quesst-technologies/qst-admin-status-all
|
opened
|
🛑 Email Service is down
|
status email-service
|
In [`2284d85`](https://github.com/quesst-technologies/qst-admin-status-all/commit/2284d8519da01714a5674a59c3a31d83f33e0c48
), Email Service (https://quessttechnologies.com/email/healthcheck) was **down**:
- HTTP code: 0
- Response time: 0 ms
|
1.0
|
🛑 Email Service is down - In [`2284d85`](https://github.com/quesst-technologies/qst-admin-status-all/commit/2284d8519da01714a5674a59c3a31d83f33e0c48
), Email Service (https://quessttechnologies.com/email/healthcheck) was **down**:
- HTTP code: 0
- Response time: 0 ms
|
non_test
|
🛑 email service is down in email service was down http code response time ms
| 0
|
178,010
| 29,481,382,356
|
IssuesEvent
|
2023-06-02 06:05:49
|
CryptKeeperZK/crypt-keeper-extension
|
https://api.github.com/repos/CryptKeeperZK/crypt-keeper-extension
|
closed
|
Support multiple root keys for different identities
|
🖼️ ui logic 🖌️ ui design 🌱 new feature ⚙️ scripts
|
As a user I'd like to create CK accounts and create identities using different addresses.
- [x] Add array of root keys to key storage
- [x] Support backup of multiple keys
- [x] Specify ck account for identity creation and list view
- [x] Integrate wallet connection with ck (web3-react)
- [x] Rewrite connect to multiple wallets
- [x] Allow user to choose account to sign with cryptkeeper (move selected account to zero index)
- [x] Check address loading
- [x] Check address changing
- [x] Check identity creation with address
- [x] Check provider vs customProvider
|
1.0
|
Support multiple root keys for different identities - As a user I'd like to create CK accounts and create identities using different addresses.
- [x] Add array of root keys to key storage
- [x] Support backup of multiple keys
- [x] Specify ck account for identity creation and list view
- [x] Integrate wallet connection with ck (web3-react)
- [x] Rewrite connect to multiple wallets
- [x] Allow user to choose account to sign with cryptkeeper (move selected account to zero index)
- [x] Check address loading
- [x] Check address changing
- [x] Check identity creation with address
- [x] Check provider vs customProvider
|
non_test
|
support multiple root keys for different identities as a user i d like to create ck accounts and create identities using different addresses add array of root keys to key storage support backup of multiple keys specify ck account for identity creation and list view integrate wallet connection with ck react rewrite connect to multiple wallets allow user to choose account to sign with cryptkeeper move selected account to zero index check address loading check address changing check identity creation with address check provider vs customprovider
| 0
|
335,042
| 30,006,802,817
|
IssuesEvent
|
2023-06-26 12:58:51
|
enthought/traits
|
https://api.github.com/repos/enthought/traits
|
opened
|
Tests fail with latest Traits and TraitsUI from PyPI
|
type: bug component: test suite
|
I'm seeing a test failure with the latest Traits and TraitsUI from PyPI, on Python 3.11. (I haven't tested with other Python versions, but it likely affects those, too.)
Steps to reproduce:
- Create a new Python 3.11 venv, and activate it.
- Run `pip install traits traitsui`
- Run `python -m unittest discover -v traits`
Expected results: the test suite runs to completion.
Actual results: we get the following test failure:
```
======================================================================
FAIL: test_edit_not_given (traits.tests.test_configure_traits.TestConfigureTraits.test_edit_not_given)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/Users/mdickinson/.venvs/traits/lib/python3.11/site-packages/traits/tests/test_configure_traits.py", line 158, in test_edit_not_given
self.assertEqual(len(captured_warnings), 0)
AssertionError: 2 != 0
----------------------------------------------------------------------
```
This is likely to be an overzealous test, but as well as tightening this test, we should investigate what the warnings are, where they're coming from, and how we can fix them.
|
1.0
|
Tests fail with latest Traits and TraitsUI from PyPI - I'm seeing a test failure with the latest Traits and TraitsUI from PyPI, on Python 3.11. (I haven't tested with other Python versions, but it likely affects those, too.)
Steps to reproduce:
- Create a new Python 3.11 venv, and activate it.
- Run `pip install traits traitsui`
- Run `python -m unittest discover -v traits`
Expected results: the test suite runs to completion.
Actual results: we get the following test failure:
```
======================================================================
FAIL: test_edit_not_given (traits.tests.test_configure_traits.TestConfigureTraits.test_edit_not_given)
----------------------------------------------------------------------
Traceback (most recent call last):
File "/Users/mdickinson/.venvs/traits/lib/python3.11/site-packages/traits/tests/test_configure_traits.py", line 158, in test_edit_not_given
self.assertEqual(len(captured_warnings), 0)
AssertionError: 2 != 0
----------------------------------------------------------------------
```
This is likely to be an overzealous test, but as well as tightening this test, we should investigate what the warnings are, where they're coming from, and how we can fix them.
|
test
|
tests fail with latest traits and traitsui from pypi i m seeing a test failure with the latest traits and traitsui from pypi on python i haven t tested with other python versions but it likely affects those too steps to reproduce create a new python venv and activate it run pip install traits traitsui run python m unittest discover v traits expected results the test suite runs to completion actual results we get the following test failure fail test edit not given traits tests test configure traits testconfiguretraits test edit not given traceback most recent call last file users mdickinson venvs traits lib site packages traits tests test configure traits py line in test edit not given self assertequal len captured warnings assertionerror this is likely to be an overzealous test but as well as tightening this test we should investigate what the warnings are where they re coming from and how we can fix them
| 1
|
452,700
| 13,058,295,177
|
IssuesEvent
|
2020-07-30 08:48:28
|
siteorigin/so-widgets-bundle
|
https://api.github.com/repos/siteorigin/so-widgets-bundle
|
closed
|
Contact Form: Ensure compatibility with Akismet
|
bug priority-2
|
Via Twitter:
@SiteOrigin
: looks like your Contact Form plugin has a bug that impairs its spam-fighting abilities: when sending data to the Akismet API, content should be sent as `comment_content`, not `comment_text`. We've handled it, for now, so your users get better results.
|
1.0
|
Contact Form: Ensure compatibility with Akismet - Via Twitter:
@SiteOrigin
: looks like your Contact Form plugin has a bug that impairs its spam-fighting abilities: when sending data to the Akismet API, content should be sent as `comment_content`, not `comment_text`. We've handled it, for now, so your users get better results.
|
non_test
|
contact form ensure compatibility with akismet via twitter siteorigin looks like your contact form plugin has a bug that impairs its spam fighting abilities when sending data to the akismet api content should be sent as comment content not comment text we ve handled it for now so your users get better results
| 0
|
5,354
| 8,181,944,071
|
IssuesEvent
|
2018-08-29 02:01:21
|
GoogleCloudPlatform/google-cloud-python
|
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python
|
closed
|
[Monitoring] cannot be used with Python 3.7 (Pandas)
|
api: monitoring testing type: process
|
Pandas/Cython doesn't currently support 3.7. Once it does we need to enable 3.7 testing on Monitoring.
Collecting pandas>=0.17.1 (from google-cloud-monitoring==0.29.0)
Using cached https://files.pythonhosted.org/packages/08/01/803834bc8a4e708aedebb133095a88a4dad9f45bbaf5ad777d2bea543c7e/pandas-0.22.0.tar.gz
Could not find a version that satisfies the requirement Cython (from versions: )
No matching distribution found for Cython
|
1.0
|
[Monitoring] cannot be used with Python 3.7 (Pandas) - Pandas/Cython doesn't currently support 3.7. Once it does we need to enable 3.7 testing on Monitoring.
Collecting pandas>=0.17.1 (from google-cloud-monitoring==0.29.0)
Using cached https://files.pythonhosted.org/packages/08/01/803834bc8a4e708aedebb133095a88a4dad9f45bbaf5ad777d2bea543c7e/pandas-0.22.0.tar.gz
Could not find a version that satisfies the requirement Cython (from versions: )
No matching distribution found for Cython
|
non_test
|
cannot be used with python pandas pandas cython doesn t currently support once it does we need to enable testing on monitoring collecting pandas from google cloud monitoring using cached could not find a version that satisfies the requirement cython from versions no matching distribution found for cython
| 0
|
124,086
| 12,224,507,308
|
IssuesEvent
|
2020-05-02 23:01:02
|
czammar/MNO_finalproject
|
https://api.github.com/repos/czammar/MNO_finalproject
|
closed
|
1.a Extracción de los datos desde Yahoo Finance
|
documentation enhancement
|
Crear una función que extraiga los precios de las acciones de 50 empresas, para los últimos 5 años:
Industria *Energy*- SASE:2222, NYSE:XOM, NYSE:CVX,ENXTAM:RDSA, NSEI:RELIANCE
Industria *Real Estate* -NYSE:AMT, NYSE:CCI, NYSE:PLD, NasdaqGS:QEIX, NYSE:DLR
Industria *Materiales*- NYSE:LIN, ASX:BHP, LSE:RIO, ENXTPA:AI, SASE:2010
Industrials Company- NSYE:LMT, NYSE:UNIP, NYSE:HON, NYSE:RTX, NYSE:UPS
Industria *Discretionary consumer* -NasdaqGS:AMZN,NYSE:BABA, NYSE:HD,ENXTPA:MC, TSE:7203
Industria *Staples*- NYSE:WMT, SWX:NESN, NYSE:PG, NYSE:KO, NasdaqGS:PEP
Industria *Health Care*- NYSE:JNJ, SWX:ROG, NYSE:UNH,NYSE:PFE, SYSE:MRK
Industria *Financials*- ARCA:VTI, ARCA:VOO, NYSE:BRK.A, NYSE:JPM, SEHK:1398
Industria *Information Technology* :NasdaqGS:MSTF, NasdaqGS:APPL, NYSE:V, KOSE:A005930, NYSE:MA
Industria *Communication* - NasdaqGS:GOOG.L, NasdaqGS:FB, SEHK:700,NYSE:VZ, NYSE:T
|
1.0
|
1.a Extracción de los datos desde Yahoo Finance - Crear una función que extraiga los precios de las acciones de 50 empresas, para los últimos 5 años:
Industria *Energy*- SASE:2222, NYSE:XOM, NYSE:CVX,ENXTAM:RDSA, NSEI:RELIANCE
Industria *Real Estate* -NYSE:AMT, NYSE:CCI, NYSE:PLD, NasdaqGS:QEIX, NYSE:DLR
Industria *Materiales*- NYSE:LIN, ASX:BHP, LSE:RIO, ENXTPA:AI, SASE:2010
Industrials Company- NSYE:LMT, NYSE:UNIP, NYSE:HON, NYSE:RTX, NYSE:UPS
Industria *Discretionary consumer* -NasdaqGS:AMZN,NYSE:BABA, NYSE:HD,ENXTPA:MC, TSE:7203
Industria *Staples*- NYSE:WMT, SWX:NESN, NYSE:PG, NYSE:KO, NasdaqGS:PEP
Industria *Health Care*- NYSE:JNJ, SWX:ROG, NYSE:UNH,NYSE:PFE, SYSE:MRK
Industria *Financials*- ARCA:VTI, ARCA:VOO, NYSE:BRK.A, NYSE:JPM, SEHK:1398
Industria *Information Technology* :NasdaqGS:MSTF, NasdaqGS:APPL, NYSE:V, KOSE:A005930, NYSE:MA
Industria *Communication* - NasdaqGS:GOOG.L, NasdaqGS:FB, SEHK:700,NYSE:VZ, NYSE:T
|
non_test
|
a extracción de los datos desde yahoo finance crear una función que extraiga los precios de las acciones de empresas para los últimos años industria energy sase nyse xom nyse cvx enxtam rdsa nsei reliance industria real estate nyse amt nyse cci nyse pld nasdaqgs qeix nyse dlr industria materiales nyse lin asx bhp lse rio enxtpa ai sase industrials company nsye lmt nyse unip nyse hon nyse rtx nyse ups industria discretionary consumer nasdaqgs amzn nyse baba nyse hd enxtpa mc tse industria staples nyse wmt swx nesn nyse pg nyse ko nasdaqgs pep industria health care nyse jnj swx rog nyse unh nyse pfe syse mrk industria financials arca vti arca voo nyse brk a nyse jpm sehk industria information technology nasdaqgs mstf nasdaqgs appl nyse v kose nyse ma industria communication nasdaqgs goog l nasdaqgs fb sehk nyse vz nyse t
| 0
|
114,163
| 9,691,256,215
|
IssuesEvent
|
2019-05-24 10:41:01
|
pybamm-team/PyBaMM
|
https://api.github.com/repos/pybamm-team/PyBaMM
|
closed
|
Shape test
|
testing
|
Add a method `test_shape` to the Finite Volume class that checks, when discretising, that we can evaluate the shape of a discretised object without raising any errors.
Should help to catch bugs earlier.
|
1.0
|
Shape test - Add a method `test_shape` to the Finite Volume class that checks, when discretising, that we can evaluate the shape of a discretised object without raising any errors.
Should help to catch bugs earlier.
|
test
|
shape test add a method test shape to the finite volume class that checks when discretising that we can evaluate the shape of a discretised object without raising any errors should help to catch bugs earlier
| 1
|
823,261
| 30,962,746,335
|
IssuesEvent
|
2023-08-08 06:04:58
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
9to5google.com - see bug description
|
browser-firefox priority-normal engine-gecko
|
<!-- @browser: Firefox 117.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/117.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/125512 -->
**URL**: https://9to5google.com/
**Browser / Version**: Firefox 117.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Dark mode doesn't work and Poles doesn't work in the new'ish design. It works fine in for example Chrome and Brave.
**Steps to Reproduce**:
Dark mode doesn't work and Poles doesn't work in the new'ish design. It works fine in for example Chrome and Brave.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/8/0a7025cd-aaf7-46b9-8bbe-a4031167e7f8.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>buildID: 20230803180221</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/8/6d9bdac7-ce53-47e5-a16c-df9aa4678001)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
9to5google.com - see bug description - <!-- @browser: Firefox 117.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:109.0) Gecko/20100101 Firefox/117.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/125512 -->
**URL**: https://9to5google.com/
**Browser / Version**: Firefox 117.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Dark mode doesn't work and Poles doesn't work in the new'ish design. It works fine in for example Chrome and Brave.
**Steps to Reproduce**:
Dark mode doesn't work and Poles doesn't work in the new'ish design. It works fine in for example Chrome and Brave.
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/8/0a7025cd-aaf7-46b9-8bbe-a4031167e7f8.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>buildID: 20230803180221</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2023/8/6d9bdac7-ce53-47e5-a16c-df9aa4678001)
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_test
|
com see bug description url browser version firefox operating system windows tested another browser yes chrome problem type something else description dark mode doesn t work and poles doesn t work in the new ish design it works fine in for example chrome and brave steps to reproduce dark mode doesn t work and poles doesn t work in the new ish design it works fine in for example chrome and brave view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
314,042
| 26,972,252,869
|
IssuesEvent
|
2023-02-09 06:24:55
|
ClickHouse/ClickHouse
|
https://api.github.com/repos/ClickHouse/ClickHouse
|
closed
|
Add a technique similar to GWP-ASan
|
feature testing
|
**Describe the solution you'd like**
Small random subset of memory allocations should be protected by guard pages.
GWP-ASan (is in fact almost unrelated to ASan) is not a specific tool but a technique that requires a small change in memory allocator. https://llvm.org/docs/GwpAsan.html
The only point is to enable it by default in production.
|
1.0
|
Add a technique similar to GWP-ASan - **Describe the solution you'd like**
Small random subset of memory allocations should be protected by guard pages.
GWP-ASan (is in fact almost unrelated to ASan) is not a specific tool but a technique that requires a small change in memory allocator. https://llvm.org/docs/GwpAsan.html
The only point is to enable it by default in production.
|
test
|
add a technique similar to gwp asan describe the solution you d like small random subset of memory allocations should be protected by guard pages gwp asan is in fact almost unrelated to asan is not a specific tool but a technique that requires a small change in memory allocator the only point is to enable it by default in production
| 1
|
64,565
| 15,949,182,673
|
IssuesEvent
|
2021-04-15 07:04:15
|
nextcloud/desktop
|
https://api.github.com/repos/nextcloud/desktop
|
closed
|
CMake Can’t Find Included SQLite3 on macOS
|
bug building os: :apple: macOS
|
## How to use GitHub
* Please use the 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to show that you are affected by the same issue.
* Please don't comment if you have no relevant information to add. It's just extra noise for everyone subscribed to this issue.
* Subscribe to receive notifications on status change and new comments.
## Expected behaviour
CMake should use included SQLite3 on macOS (like it usually does).
## Actual behaviour
CMake can’t find included SQLite3 on macOS:
```bash
elsiehupp@Elsies-MacBook build % cmake ..
-- Build of crashreporter disabled.
-- GIT_SHA1 71658b09741fd6e49e6b4fb0913ce88142f01f8b
Compiling without updater
-- Using Qt 5.15.2 (/usr/local/opt/qt/bin/qmake)
-- Using own sqlite3 from /Users/elsiehupp/Repositories/nextcloud-desktop/src/3rdparty/sqlite3
-- Using own sqlite3 from /Users/elsiehupp/Repositories/nextcloud-desktop/src/3rdparty/sqlite3
-- OWNCLOUD_SIDEBAR_ICONS: Nextcloud: /Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/128-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/16-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/256-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/32-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/64-Nextcloud-sidebar.png
-- Using own sqlite3 from /Users/elsiehupp/Repositories/nextcloud-desktop/src/3rdparty/sqlite3
-- The following OPTIONAL packages have been found:
* Sphinx
* PdfLatex
* Qt5LinguistTools
* Qt5MacExtras
* Doxygen
* PkgConfig
* CMocka
-- The following REQUIRED packages have been found:
* OpenSSL (required version >= 1.1)
* ZLIB
* Qt5Concurrent
* Qt5WebChannel (required version >= 5.15.2)
* Qt5Positioning (required version >= 5.15.2)
* Qt5WebEngineCore (required version >= 5.15.2)
* Qt5PrintSupport (required version >= 5.15.2)
* Qt5WebEngineWidgets
* Qt5WebEngine
* Qt5Keychain
* Qt5WebSockets
* Qt5Widgets
* Qt5Svg
* Qt5QuickControls2
* Qt5Core
* Qt5Test
* Qt5Xml
* Qt5Network
* Qt5Qml
* Qt5Gui (required version >= 5.15.2)
* Qt5QmlModels (required version >= 5.15.2)
* Qt5Quick
* Qt5
-- The following REQUIRED packages have not been found:
* SQLite3 (required version >= 3.8.0)
CMake Error at /usr/local/Cellar/cmake/3.19.4/share/cmake/Modules/FeatureSummary.cmake:457 (message):
feature_summary() Error: REQUIRED package(s) are missing, aborting CMake
run.
Call Stack (most recent call first):
CMakeLists.txt:280 (feature_summary)
-- Configuring incomplete, errors occurred!
```
## Steps to reproduce
```bash
% git checkout b1fc7d4b52fe61cd35159674985033ae1fb0543b
% cd build
% cmake ..
```
## Environment
Operating system: macOS Big Sur 11.2.1
OS language: English (US)
Git commit where error is present: [`b1fc7d4`
](https://github.com/nextcloud/desktop/commit/b1fc7d4b52fe61cd35159674985033ae1fb0543b) (i.e. latest on `master`)
|
1.0
|
CMake Can’t Find Included SQLite3 on macOS - ## How to use GitHub
* Please use the 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to show that you are affected by the same issue.
* Please don't comment if you have no relevant information to add. It's just extra noise for everyone subscribed to this issue.
* Subscribe to receive notifications on status change and new comments.
## Expected behaviour
CMake should use included SQLite3 on macOS (like it usually does).
## Actual behaviour
CMake can’t find included SQLite3 on macOS:
```bash
elsiehupp@Elsies-MacBook build % cmake ..
-- Build of crashreporter disabled.
-- GIT_SHA1 71658b09741fd6e49e6b4fb0913ce88142f01f8b
Compiling without updater
-- Using Qt 5.15.2 (/usr/local/opt/qt/bin/qmake)
-- Using own sqlite3 from /Users/elsiehupp/Repositories/nextcloud-desktop/src/3rdparty/sqlite3
-- Using own sqlite3 from /Users/elsiehupp/Repositories/nextcloud-desktop/src/3rdparty/sqlite3
-- OWNCLOUD_SIDEBAR_ICONS: Nextcloud: /Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/128-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/16-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/256-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/32-Nextcloud-sidebar.png;/Users/elsiehupp/Repositories/nextcloud-desktop/theme/colored/64-Nextcloud-sidebar.png
-- Using own sqlite3 from /Users/elsiehupp/Repositories/nextcloud-desktop/src/3rdparty/sqlite3
-- The following OPTIONAL packages have been found:
* Sphinx
* PdfLatex
* Qt5LinguistTools
* Qt5MacExtras
* Doxygen
* PkgConfig
* CMocka
-- The following REQUIRED packages have been found:
* OpenSSL (required version >= 1.1)
* ZLIB
* Qt5Concurrent
* Qt5WebChannel (required version >= 5.15.2)
* Qt5Positioning (required version >= 5.15.2)
* Qt5WebEngineCore (required version >= 5.15.2)
* Qt5PrintSupport (required version >= 5.15.2)
* Qt5WebEngineWidgets
* Qt5WebEngine
* Qt5Keychain
* Qt5WebSockets
* Qt5Widgets
* Qt5Svg
* Qt5QuickControls2
* Qt5Core
* Qt5Test
* Qt5Xml
* Qt5Network
* Qt5Qml
* Qt5Gui (required version >= 5.15.2)
* Qt5QmlModels (required version >= 5.15.2)
* Qt5Quick
* Qt5
-- The following REQUIRED packages have not been found:
* SQLite3 (required version >= 3.8.0)
CMake Error at /usr/local/Cellar/cmake/3.19.4/share/cmake/Modules/FeatureSummary.cmake:457 (message):
feature_summary() Error: REQUIRED package(s) are missing, aborting CMake
run.
Call Stack (most recent call first):
CMakeLists.txt:280 (feature_summary)
-- Configuring incomplete, errors occurred!
```
## Steps to reproduce
```bash
% git checkout b1fc7d4b52fe61cd35159674985033ae1fb0543b
% cd build
% cmake ..
```
## Environment
Operating system: macOS Big Sur 11.2.1
OS language: English (US)
Git commit where error is present: [`b1fc7d4`
](https://github.com/nextcloud/desktop/commit/b1fc7d4b52fe61cd35159674985033ae1fb0543b) (i.e. latest on `master`)
|
non_test
|
cmake can’t find included on macos how to use github please use the 👍 to show that you are affected by the same issue please don t comment if you have no relevant information to add it s just extra noise for everyone subscribed to this issue subscribe to receive notifications on status change and new comments expected behaviour cmake should use included on macos like it usually does actual behaviour cmake can’t find included on macos bash elsiehupp elsies macbook build cmake build of crashreporter disabled git compiling without updater using qt usr local opt qt bin qmake using own from users elsiehupp repositories nextcloud desktop src using own from users elsiehupp repositories nextcloud desktop src owncloud sidebar icons nextcloud users elsiehupp repositories nextcloud desktop theme colored nextcloud sidebar png users elsiehupp repositories nextcloud desktop theme colored nextcloud sidebar png users elsiehupp repositories nextcloud desktop theme colored nextcloud sidebar png users elsiehupp repositories nextcloud desktop theme colored nextcloud sidebar png users elsiehupp repositories nextcloud desktop theme colored nextcloud sidebar png using own from users elsiehupp repositories nextcloud desktop src the following optional packages have been found sphinx pdflatex doxygen pkgconfig cmocka the following required packages have been found openssl required version zlib required version required version required version required version required version required version the following required packages have not been found required version cmake error at usr local cellar cmake share cmake modules featuresummary cmake message feature summary error required package s are missing aborting cmake run call stack most recent call first cmakelists txt feature summary configuring incomplete errors occurred steps to reproduce bash git checkout cd build cmake environment operating system macos big sur os language english us git commit where error is present i e latest on master
| 0
|
59,901
| 6,666,153,595
|
IssuesEvent
|
2017-10-03 06:49:35
|
RaRe-Technologies/gensim
|
https://api.github.com/repos/RaRe-Technologies/gensim
|
closed
|
use AppVeyor to test on Windows and upload wheels
|
medium testing wishlist
|
AppVeyor is free for open source projects:
http://www.appveyor.com/
|
1.0
|
use AppVeyor to test on Windows and upload wheels - AppVeyor is free for open source projects:
http://www.appveyor.com/
|
test
|
use appveyor to test on windows and upload wheels appveyor is free for open source projects
| 1
|
176,649
| 28,130,863,341
|
IssuesEvent
|
2023-03-31 22:53:06
|
pulumi/pulumi-cloudflare
|
https://api.github.com/repos/pulumi/pulumi-cloudflare
|
closed
|
Importing zone doesn't work propertly
|
kind/bug resolution/by-design
|
### What happened?
I've imported a zone that already exists, but during preview, the resource is not shown at all (with the `import` info, for example).
### Steps to reproduce
e.g.
```
new cloudflare.Zone("myzone", {
zone: "example.com",
plan: "free",
},
{
id: "<the zone id>",
}
```
Run pulumi up, let it go, and then check the state. The resource will be there.
### Expected Behavior
1. resource is shown during pulumi preview
2. once imported, the options.id can be removed without causing any changes to the stack
### Actual Behavior
1. the resource should be shown in during preview as to be `imported`
2. removing the id causes the resource to be replaced
This is the output after removing the `id` from the resource options:
```
pulumi:pulumi:Stack: (same)
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:pulumi:Stack::cloudflare-provisioning-example-cloudflare-provisioning]
++cloudflare:index/zone:Zone: (create-replacement) 🔒
[id=9f9441adc91869e2557ef2e67add339e]
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::cloudflare:index/zone:Zone::example-com.zone]
[provider=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:providers:cloudflare::default_4_6_0::f4cb6084-3d93-4eff-9caa-c96f06d63858]
jumpStart : false
plan : "enterprise"
type : "full"
zone : "example.com"
+-cloudflare:index/zone:Zone: (replace) 🔒
[id=9f9441adc91869e2557ef2e67add339e]
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::cloudflare:index/zone:Zone::example-com.zone]
[provider=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:providers:cloudflare::default_4_6_0::f4cb6084-3d93-4eff-9caa-c96f06d63858]
- id : "9f9441adc91869e2817ef2e67add339e"
- meta : {
- phishing_detected : false
- wildcard_proxiable: false
}
- nameServers : [
- [0]: "daisy.ns.cloudflare.com"
- [1]: "owen.ns.cloudflare.com"
]
- paused : false
- status : "active"
- vanityNameServers: []
- verificationKey : ""
<<cloudflare:index/zone:Zone: (discard-replaced)
[id=9f9441adc91869e2557ef2e67add339e]
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::cloudflare:index/zone:Zone::example-com.zone]
[provider=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:providers:cloudflare::default_4_6_0::f4cb6084-3d93-4eff-9caa-c96f06d63858]
```
### Versions used
```
CLI
Version 3.32.0
Go Version go1.17.9
Go Compiler gc
Plugins
NAME VERSION
aws 4.38.1
cloudflare 4.6.0
cloudflare 3.6.0
nodejs unknown
tls 4.3.0
Host
OS debian
Version 11.3
Arch x86_64
This project is written in nodejs (/home/renann/.asdf/shims/node v16.0.0
```
### Additional context
~I do have intentionally, at the moment, two versions of the Cloudflare provider, but I believe that should not be the cause of the issue. Though as a followup I'll try this out alone in a separate stack to see whether the behavior changes.~
I could reproduce it with a more common implementation and the behavior is the same.
### Contributing
Vote on this issue by adding a 👍 reaction.
To contribute a fix for this issue, leave a comment (and link to your pull request, if you've opened one already).
|
1.0
|
Importing zone doesn't work propertly - ### What happened?
I've imported a zone that already exists, but during preview, the resource is not shown at all (with the `import` info, for example).
### Steps to reproduce
e.g.
```
new cloudflare.Zone("myzone", {
zone: "example.com",
plan: "free",
},
{
id: "<the zone id>",
}
```
Run pulumi up, let it go, and then check the state. The resource will be there.
### Expected Behavior
1. resource is shown during pulumi preview
2. once imported, the options.id can be removed without causing any changes to the stack
### Actual Behavior
1. the resource should be shown in during preview as to be `imported`
2. removing the id causes the resource to be replaced
This is the output after removing the `id` from the resource options:
```
pulumi:pulumi:Stack: (same)
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:pulumi:Stack::cloudflare-provisioning-example-cloudflare-provisioning]
++cloudflare:index/zone:Zone: (create-replacement) 🔒
[id=9f9441adc91869e2557ef2e67add339e]
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::cloudflare:index/zone:Zone::example-com.zone]
[provider=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:providers:cloudflare::default_4_6_0::f4cb6084-3d93-4eff-9caa-c96f06d63858]
jumpStart : false
plan : "enterprise"
type : "full"
zone : "example.com"
+-cloudflare:index/zone:Zone: (replace) 🔒
[id=9f9441adc91869e2557ef2e67add339e]
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::cloudflare:index/zone:Zone::example-com.zone]
[provider=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:providers:cloudflare::default_4_6_0::f4cb6084-3d93-4eff-9caa-c96f06d63858]
- id : "9f9441adc91869e2817ef2e67add339e"
- meta : {
- phishing_detected : false
- wildcard_proxiable: false
}
- nameServers : [
- [0]: "daisy.ns.cloudflare.com"
- [1]: "owen.ns.cloudflare.com"
]
- paused : false
- status : "active"
- vanityNameServers: []
- verificationKey : ""
<<cloudflare:index/zone:Zone: (discard-replaced)
[id=9f9441adc91869e2557ef2e67add339e]
[urn=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::cloudflare:index/zone:Zone::example-com.zone]
[provider=urn:pulumi:example-cloudflare-provisioning::cloudflare-provisioning::pulumi:providers:cloudflare::default_4_6_0::f4cb6084-3d93-4eff-9caa-c96f06d63858]
```
### Versions used
```
CLI
Version 3.32.0
Go Version go1.17.9
Go Compiler gc
Plugins
NAME VERSION
aws 4.38.1
cloudflare 4.6.0
cloudflare 3.6.0
nodejs unknown
tls 4.3.0
Host
OS debian
Version 11.3
Arch x86_64
This project is written in nodejs (/home/renann/.asdf/shims/node v16.0.0
```
### Additional context
~I do have intentionally, at the moment, two versions of the Cloudflare provider, but I believe that should not be the cause of the issue. Though as a followup I'll try this out alone in a separate stack to see whether the behavior changes.~
I could reproduce it with a more common implementation and the behavior is the same.
### Contributing
Vote on this issue by adding a 👍 reaction.
To contribute a fix for this issue, leave a comment (and link to your pull request, if you've opened one already).
|
non_test
|
importing zone doesn t work propertly what happened i ve imported a zone that already exists but during preview the resource is not shown at all with the import info for example steps to reproduce e g new cloudflare zone myzone zone example com plan free id run pulumi up let it go and then check the state the resource will be there expected behavior resource is shown during pulumi preview once imported the options id can be removed without causing any changes to the stack actual behavior the resource should be shown in during preview as to be imported removing the id causes the resource to be replaced this is the output after removing the id from the resource options pulumi pulumi stack same cloudflare index zone zone create replacement 🔒 jumpstart false plan enterprise type full zone example com cloudflare index zone zone replace 🔒 id meta phishing detected false wildcard proxiable false nameservers daisy ns cloudflare com owen ns cloudflare com paused false status active vanitynameservers verificationkey cloudflare index zone zone discard replaced versions used cli version go version go compiler gc plugins name version aws cloudflare cloudflare nodejs unknown tls host os debian version arch this project is written in nodejs home renann asdf shims node additional context i do have intentionally at the moment two versions of the cloudflare provider but i believe that should not be the cause of the issue though as a followup i ll try this out alone in a separate stack to see whether the behavior changes i could reproduce it with a more common implementation and the behavior is the same contributing vote on this issue by adding a 👍 reaction to contribute a fix for this issue leave a comment and link to your pull request if you ve opened one already
| 0
|
96,939
| 12,193,394,297
|
IssuesEvent
|
2020-04-29 14:21:00
|
USDA-FSA/fsa-style
|
https://api.github.com/repos/USDA-FSA/fsa-style
|
closed
|
Apply Level to relevant existing components
|
P3 source: internal FBCSS type: design type: feature request
|
The new "Level" component (#385) will naturally lend itself to many other "composite-like" components.
* Evaluate which components can be updated/amended with it
* Deprecate variations/examples as necessary, etc.
|
1.0
|
Apply Level to relevant existing components - The new "Level" component (#385) will naturally lend itself to many other "composite-like" components.
* Evaluate which components can be updated/amended with it
* Deprecate variations/examples as necessary, etc.
|
non_test
|
apply level to relevant existing components the new level component will naturally lend itself to many other composite like components evaluate which components can be updated amended with it deprecate variations examples as necessary etc
| 0
|
557,979
| 16,523,688,862
|
IssuesEvent
|
2021-05-26 17:14:10
|
chef/chef
|
https://api.github.com/repos/chef/chef
|
closed
|
Chef::DataBagItem.save does not properly validate names with periods
|
Focus: knife Priority: Low Status: Good First Issue Type: Bug
|
BUG:
Databag item ID can not contain period " . " when created through `knife databag` commands or via `chef manage` UI.
However, they can be created from inside the recipe via code with `Chef::DataBagItem.new`; `raw_data` & `.save`
Once the Databag Item is created via the recipe, the item can be edited via the GUI but can't be refreshed via `knife databag` command
Workaround:
This issue manifested itself when trying to save server names and state data in a data_bag. Since FQDN contains periods, we had to write gsub() code to convert dots ' . ' to underscores ' _ '.
|
1.0
|
Chef::DataBagItem.save does not properly validate names with periods - BUG:
Databag item ID can not contain period " . " when created through `knife databag` commands or via `chef manage` UI.
However, they can be created from inside the recipe via code with `Chef::DataBagItem.new`; `raw_data` & `.save`
Once the Databag Item is created via the recipe, the item can be edited via the GUI but can't be refreshed via `knife databag` command
Workaround:
This issue manifested itself when trying to save server names and state data in a data_bag. Since FQDN contains periods, we had to write gsub() code to convert dots ' . ' to underscores ' _ '.
|
non_test
|
chef databagitem save does not properly validate names with periods bug databag item id can not contain period when created through knife databag commands or via chef manage ui however they can be created from inside the recipe via code with chef databagitem new raw data save once the databag item is created via the recipe the item can be edited via the gui but can t be refreshed via knife databag command workaround this issue manifested itself when trying to save server names and state data in a data bag since fqdn contains periods we had to write gsub code to convert dots to underscores
| 0
|
26,583
| 4,234,992,459
|
IssuesEvent
|
2016-07-05 13:55:39
|
brave/browser-laptop
|
https://api.github.com/repos/brave/browser-laptop
|
closed
|
Manual tests for Ubuntu 0.11.0 RC3
|
tests
|
1. [x] Check that installer is close to the size of last release.
2. [x] Check Brave, electron, and libchromiumcontent version in About and make sure it is EXACTLY as expected.
## About pages
1. [x] Test that about:bookmarks loads bookmarks
2. [x] Test that about:downloads loads downloads
3. [x] Test that about:preferences changing a preference takes effect right away
4. [x] Test that about:preferences language change takes effect on re-start
5. [x] Test that about:passwords loads
## Context menus
1. [x] Make sure context menu items in the URL bar work
2. [x] Make sure context menu items on content work with no selected text.
3. [x] Make sure context menu items on content work with selected text.
4. [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable).
## Find on page
1. [x] Ensure search box is shown with shortcut
2. [x] Test successful find
3. [x] Test forward and backward find navigation
4. [x] Test failed find shows 0 results
5. [x] Test match case find
## Site hacks
1. [x] Test twitch.tv sub-page loads a video and you can play it
## Downloads
1. [x] Test downloading a file works and that all actions on the download item works.
## Fullscreen
1. [x] Test that entering full screen window works View -> Toggle Full Screen. And exit back (Not Esc).
2. [x] Test that entering HTML5 full screen works. And Esc to go back. (youtube.com)
## Tabs and Pinning
1. [x] Test that tabs are pinnable
2. [x] Test that tabs are unpinnable
3. [x] Test that tabs are draggable to same tabset
4. [x] Test that tabs are draggable to alternate tabset
## Zoom
1. [x] Test zoom in / out shortcut works
2. [x] Test hamburger menu zooms.
3. [x] Test zoom saved when you close the browser and restore on a single site.
4. [x] Test zoom saved when you navigate within a single origin site.
5. [x] Test that navigating to a different origin resets the zoom
## Bookmarks
1. [x] Test that creating a bookmark on the bookmarks toolbar works
2. [x] Test that creating a bookmark folder on the bookmarks toolbar works
3. [x] Test that moving a bookmark into a folder by drag and drop on the bookmarks folder works
4. [x] Test that clicking a bookmark in the toolbar loads the bookmark.
5. [x] Test that clicking a bookmark in a bookmark toolbar folder loads the bookmark.
## Bravery settings
1. [x] Check that HTTPS Everywhere works by loading http://www.apple.com
2. [x] Turning HTTPS Everywhere off and shields off both disable the redirect to apple.com
3. [x] Check that ad replacement works on http://slashdot.org
4. [x] Check that toggling to blocking and allow ads works as expected.
5. [x] Test that clicking through a cert error in badssl.com works.
6. [x] Test that Safe Browsing works (excellentmovies.net)
7. [x] Turning Safe Browsing off and shields off both disable safe browsing for excellentmovies.net.
8. [x] Visit brianbondy.com and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work.
9. [x] Test that about:preferences default Bravery settings take effect on pages with no site settings.
10. [x] Test that turning on fingerprinting protection in about:preferences blocks fingerprint at http://browserleaks.com/canvas. Test that turning it off in the Bravery menu shows the fingerprint again.
11. [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/7/ when 3rd party cookies are blocked.
12. [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ when fingerprinting protection is on.
## Content tests
1. [x] Load twitter and click on a tweet so the popup div shows. Click to dismiss and repeat with another div. Make sure it shows.
2. [x] Go to brianbondy.com and click on the twitter icon on the top right. Test that context menus work in the new twitter tab.
3. [ ] Go to http://www.bennish.net/web-notifications.html and test that clicking on 'Show' pops up a notification asking for permission. Make sure that clicking 'Deny' leads to no notifications being shown.
4. [x] Go to https://trac.torproject.org/projects/tor/login and make sure that the password can be saved. Make sure the saved password shows up in `about:passwords`.
5. [x] Open a github issue and type some misspellings, make sure they are underlined.
6. [x] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text.
7. [x] Make sure that Command + Click (Control + Click on Windows, Control + Click on Ubuntu) on a link opens a new tab but does NOT switch to it. Click on it and make sure it is already loaded.
8. [x] Make sure that clicking links from gmail or inbox.google.com works.
## Session storage
1. [x] Temporarily move away your `~/Library/Application\ Support/Brave/session-store-1` and test that clean session storage works. (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu)
2. [x] Make sure that data from the last version appears in the new version OK.
3. [x] Test that windows and tabs restore when closed, including active tab.
4. [x] Test that the previous version's cookies are preserved in the next version.
5. [x] Move away your entire `~/Library/Application\ Support/Brave` folder (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu)
## Update tests
1. [x] Test that update works correctly
|
1.0
|
Manual tests for Ubuntu 0.11.0 RC3 - 1. [x] Check that installer is close to the size of last release.
2. [x] Check Brave, electron, and libchromiumcontent version in About and make sure it is EXACTLY as expected.
## About pages
1. [x] Test that about:bookmarks loads bookmarks
2. [x] Test that about:downloads loads downloads
3. [x] Test that about:preferences changing a preference takes effect right away
4. [x] Test that about:preferences language change takes effect on re-start
5. [x] Test that about:passwords loads
## Context menus
1. [x] Make sure context menu items in the URL bar work
2. [x] Make sure context menu items on content work with no selected text.
3. [x] Make sure context menu items on content work with selected text.
4. [x] Make sure context menu items on content work inside an editable control (input, textarea, or contenteditable).
## Find on page
1. [x] Ensure search box is shown with shortcut
2. [x] Test successful find
3. [x] Test forward and backward find navigation
4. [x] Test failed find shows 0 results
5. [x] Test match case find
## Site hacks
1. [x] Test twitch.tv sub-page loads a video and you can play it
## Downloads
1. [x] Test downloading a file works and that all actions on the download item works.
## Fullscreen
1. [x] Test that entering full screen window works View -> Toggle Full Screen. And exit back (Not Esc).
2. [x] Test that entering HTML5 full screen works. And Esc to go back. (youtube.com)
## Tabs and Pinning
1. [x] Test that tabs are pinnable
2. [x] Test that tabs are unpinnable
3. [x] Test that tabs are draggable to same tabset
4. [x] Test that tabs are draggable to alternate tabset
## Zoom
1. [x] Test zoom in / out shortcut works
2. [x] Test hamburger menu zooms.
3. [x] Test zoom saved when you close the browser and restore on a single site.
4. [x] Test zoom saved when you navigate within a single origin site.
5. [x] Test that navigating to a different origin resets the zoom
## Bookmarks
1. [x] Test that creating a bookmark on the bookmarks toolbar works
2. [x] Test that creating a bookmark folder on the bookmarks toolbar works
3. [x] Test that moving a bookmark into a folder by drag and drop on the bookmarks folder works
4. [x] Test that clicking a bookmark in the toolbar loads the bookmark.
5. [x] Test that clicking a bookmark in a bookmark toolbar folder loads the bookmark.
## Bravery settings
1. [x] Check that HTTPS Everywhere works by loading http://www.apple.com
2. [x] Turning HTTPS Everywhere off and shields off both disable the redirect to apple.com
3. [x] Check that ad replacement works on http://slashdot.org
4. [x] Check that toggling to blocking and allow ads works as expected.
5. [x] Test that clicking through a cert error in badssl.com works.
6. [x] Test that Safe Browsing works (excellentmovies.net)
7. [x] Turning Safe Browsing off and shields off both disable safe browsing for excellentmovies.net.
8. [x] Visit brianbondy.com and then turn on script blocking, nothing should load. Allow it from the script blocking UI in the URL bar and it should work.
9. [x] Test that about:preferences default Bravery settings take effect on pages with no site settings.
10. [x] Test that turning on fingerprinting protection in about:preferences blocks fingerprint at http://browserleaks.com/canvas. Test that turning it off in the Bravery menu shows the fingerprint again.
11. [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/7/ when 3rd party cookies are blocked.
12. [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ when fingerprinting protection is on.
## Content tests
1. [x] Load twitter and click on a tweet so the popup div shows. Click to dismiss and repeat with another div. Make sure it shows.
2. [x] Go to brianbondy.com and click on the twitter icon on the top right. Test that context menus work in the new twitter tab.
3. [ ] Go to http://www.bennish.net/web-notifications.html and test that clicking on 'Show' pops up a notification asking for permission. Make sure that clicking 'Deny' leads to no notifications being shown.
4. [x] Go to https://trac.torproject.org/projects/tor/login and make sure that the password can be saved. Make sure the saved password shows up in `about:passwords`.
5. [x] Open a github issue and type some misspellings, make sure they are underlined.
6. [x] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text.
7. [x] Make sure that Command + Click (Control + Click on Windows, Control + Click on Ubuntu) on a link opens a new tab but does NOT switch to it. Click on it and make sure it is already loaded.
8. [x] Make sure that clicking links from gmail or inbox.google.com works.
## Session storage
1. [x] Temporarily move away your `~/Library/Application\ Support/Brave/session-store-1` and test that clean session storage works. (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu)
2. [x] Make sure that data from the last version appears in the new version OK.
3. [x] Test that windows and tabs restore when closed, including active tab.
4. [x] Test that the previous version's cookies are preserved in the next version.
5. [x] Move away your entire `~/Library/Application\ Support/Brave` folder (`%appdata%\Brave in Windows`, `./config/brave` in Ubuntu)
## Update tests
1. [x] Test that update works correctly
|
test
|
manual tests for ubuntu check that installer is close to the size of last release check brave electron and libchromiumcontent version in about and make sure it is exactly as expected about pages test that about bookmarks loads bookmarks test that about downloads loads downloads test that about preferences changing a preference takes effect right away test that about preferences language change takes effect on re start test that about passwords loads context menus make sure context menu items in the url bar work make sure context menu items on content work with no selected text make sure context menu items on content work with selected text make sure context menu items on content work inside an editable control input textarea or contenteditable find on page ensure search box is shown with shortcut test successful find test forward and backward find navigation test failed find shows results test match case find site hacks test twitch tv sub page loads a video and you can play it downloads test downloading a file works and that all actions on the download item works fullscreen test that entering full screen window works view toggle full screen and exit back not esc test that entering full screen works and esc to go back youtube com tabs and pinning test that tabs are pinnable test that tabs are unpinnable test that tabs are draggable to same tabset test that tabs are draggable to alternate tabset zoom test zoom in out shortcut works test hamburger menu zooms test zoom saved when you close the browser and restore on a single site test zoom saved when you navigate within a single origin site test that navigating to a different origin resets the zoom bookmarks test that creating a bookmark on the bookmarks toolbar works test that creating a bookmark folder on the bookmarks toolbar works test that moving a bookmark into a folder by drag and drop on the bookmarks folder works test that clicking a bookmark in the toolbar loads the bookmark test that clicking a bookmark in a bookmark toolbar folder loads the bookmark bravery settings check that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to apple com check that ad replacement works on check that toggling to blocking and allow ads works as expected test that clicking through a cert error in badssl com works test that safe browsing works excellentmovies net turning safe browsing off and shields off both disable safe browsing for excellentmovies net visit brianbondy com and then turn on script blocking nothing should load allow it from the script blocking ui in the url bar and it should work test that about preferences default bravery settings take effect on pages with no site settings test that turning on fingerprinting protection in about preferences blocks fingerprint at test that turning it off in the bravery menu shows the fingerprint again test that party storage results are blank at when party cookies are blocked test that audio fingerprint is blocked at when fingerprinting protection is on content tests load twitter and click on a tweet so the popup div shows click to dismiss and repeat with another div make sure it shows go to brianbondy com and click on the twitter icon on the top right test that context menus work in the new twitter tab go to and test that clicking on show pops up a notification asking for permission make sure that clicking deny leads to no notifications being shown go to and make sure that the password can be saved make sure the saved password shows up in about passwords open a github issue and type some misspellings make sure they are underlined make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text make sure that command click control click on windows control click on ubuntu on a link opens a new tab but does not switch to it click on it and make sure it is already loaded make sure that clicking links from gmail or inbox google com works session storage temporarily move away your library application support brave session store and test that clean session storage works appdata brave in windows config brave in ubuntu make sure that data from the last version appears in the new version ok test that windows and tabs restore when closed including active tab test that the previous version s cookies are preserved in the next version move away your entire library application support brave folder appdata brave in windows config brave in ubuntu update tests test that update works correctly
| 1
|
557,848
| 16,520,787,876
|
IssuesEvent
|
2021-05-26 14:22:09
|
ARMmbed/mbed-os
|
https://api.github.com/repos/ARMmbed/mbed-os
|
closed
|
High-ish SPI clock speed breaks STM async SPI
|
devices: st priority: untriaged
|
<!--
************************************** WARNING **************************************
The ciarcom bot parses this header automatically. Any deviation from the
template may cause the bot to automatically correct this header or may result in a
warning message, requesting updates.
Please ensure all sections of the template below are filled in and no changes
are made to the template format. Only bugs should be raised here as issues.
Questions or enhancements should instead be raised on our forums:
https://forums.mbed.com/ .
*************************************************************************************
-->
### Description of defect
In my testing, I found that setting a high-ish SPI clock speed (at least 1.5MHz) on my NUCLEO_F429ZI board causes all SPI transactions to fail with `SPI_EVENT_ERROR`. However, if I set the clock speed to 1.4MHz or lower, my driver works as expected.
After some investigation from the debugger, I found that it's entering [this error block here](https://github.com/ARMmbed/mbed-os/blob/f2278567d09b9ae9f4843e1d9d393526b9462783/targets/TARGET_STM/TARGET_STM32F4/STM32Cube_FW/STM32F4xx_HAL_Driver/stm32f4xx_hal_spi.c#L2363) in the HAL's SPI interrupt function.
I believe I understand why this error is happening -- it comes from the way that STMicro has implemented async SPI using interrupts. Their setup relies on the processor being interrupted by the SPI peripheral each time a byte has been transmitted. The interrupt handler (`HAL_SPI_IRQHandler`) enqueues the next byte in software. However, it seems that once you go above a certain (not actually that high) SPI frequency, the interrupt simply can't keep up anymore. This then causes the overflow error to be generated by the hardware, since the first byte of the message doesn't get transferred by the time the second byte is read.
As far as what can be done about this, I recognize that it's to some extent a design limitation. However, at minimum, it would be very nice if it were documented as an issue somewhere, such as the docstring for SPI::transfer(). Maybe just a note that, "Some processors have a limit on the maximum bus clock speed that can be reached when using asynchronous SPI, especially if other interrupts are triggered during the transfer. If the SPI transfer fails, try reducing the clock speed." Even better would be to add more detailed error codes including something like `SPI_EVENT_OVERRUN` that indicate to the user what has happened.
There also ways you could try to address the issue directly. Maybe there's some way to optimize the SPI IRQ handler so it requires less instructions to copy the byte and get out? Currently the SPI IRQ seems to take about 100 instructions based on the max achievable speed, which seems like more than would strictly be needed. Even better, however, would be to enable the use of DMA for high speed SPI transactions, which wouldn't tax the CPU at all. I've heard that it's supposed to be on the roadmap soon, right? So fingers crossed.
#### Target(s) affected by this defect ?
Tested on STM32F429ZI. I assume it affects all STMicro chips, though the max usable frequency depends on the processor clock.
#### Toolchain(s) (name and version) displaying this defect ?
Tested with GCC_ARM, using the standard debug profile
#### What version of Mbed-os are you using (tag or sha) ?
mbed-os-6.1.0
#### What version(s) of tools are you using. List all that apply (E.g. mbed-cli)
I'm using mbed-cmake 1.4.1 as my build system, but this should be identical to mbed-cli.
#### How is this defect reproduced ?
- Create an SPI object and set its clock speed to a value of 1500000 Hz or more.
- Start an asynchronous transfer to read bytes from any sensor or other peripheral
- Observe `SPI_EVENT_ERROR` being passed to the callback
|
1.0
|
High-ish SPI clock speed breaks STM async SPI - <!--
************************************** WARNING **************************************
The ciarcom bot parses this header automatically. Any deviation from the
template may cause the bot to automatically correct this header or may result in a
warning message, requesting updates.
Please ensure all sections of the template below are filled in and no changes
are made to the template format. Only bugs should be raised here as issues.
Questions or enhancements should instead be raised on our forums:
https://forums.mbed.com/ .
*************************************************************************************
-->
### Description of defect
In my testing, I found that setting a high-ish SPI clock speed (at least 1.5MHz) on my NUCLEO_F429ZI board causes all SPI transactions to fail with `SPI_EVENT_ERROR`. However, if I set the clock speed to 1.4MHz or lower, my driver works as expected.
After some investigation from the debugger, I found that it's entering [this error block here](https://github.com/ARMmbed/mbed-os/blob/f2278567d09b9ae9f4843e1d9d393526b9462783/targets/TARGET_STM/TARGET_STM32F4/STM32Cube_FW/STM32F4xx_HAL_Driver/stm32f4xx_hal_spi.c#L2363) in the HAL's SPI interrupt function.
I believe I understand why this error is happening -- it comes from the way that STMicro has implemented async SPI using interrupts. Their setup relies on the processor being interrupted by the SPI peripheral each time a byte has been transmitted. The interrupt handler (`HAL_SPI_IRQHandler`) enqueues the next byte in software. However, it seems that once you go above a certain (not actually that high) SPI frequency, the interrupt simply can't keep up anymore. This then causes the overflow error to be generated by the hardware, since the first byte of the message doesn't get transferred by the time the second byte is read.
As far as what can be done about this, I recognize that it's to some extent a design limitation. However, at minimum, it would be very nice if it were documented as an issue somewhere, such as the docstring for SPI::transfer(). Maybe just a note that, "Some processors have a limit on the maximum bus clock speed that can be reached when using asynchronous SPI, especially if other interrupts are triggered during the transfer. If the SPI transfer fails, try reducing the clock speed." Even better would be to add more detailed error codes including something like `SPI_EVENT_OVERRUN` that indicate to the user what has happened.
There also ways you could try to address the issue directly. Maybe there's some way to optimize the SPI IRQ handler so it requires less instructions to copy the byte and get out? Currently the SPI IRQ seems to take about 100 instructions based on the max achievable speed, which seems like more than would strictly be needed. Even better, however, would be to enable the use of DMA for high speed SPI transactions, which wouldn't tax the CPU at all. I've heard that it's supposed to be on the roadmap soon, right? So fingers crossed.
#### Target(s) affected by this defect ?
Tested on STM32F429ZI. I assume it affects all STMicro chips, though the max usable frequency depends on the processor clock.
#### Toolchain(s) (name and version) displaying this defect ?
Tested with GCC_ARM, using the standard debug profile
#### What version of Mbed-os are you using (tag or sha) ?
mbed-os-6.1.0
#### What version(s) of tools are you using. List all that apply (E.g. mbed-cli)
I'm using mbed-cmake 1.4.1 as my build system, but this should be identical to mbed-cli.
#### How is this defect reproduced ?
- Create an SPI object and set its clock speed to a value of 1500000 Hz or more.
- Start an asynchronous transfer to read bytes from any sensor or other peripheral
- Observe `SPI_EVENT_ERROR` being passed to the callback
|
non_test
|
high ish spi clock speed breaks stm async spi warning the ciarcom bot parses this header automatically any deviation from the template may cause the bot to automatically correct this header or may result in a warning message requesting updates please ensure all sections of the template below are filled in and no changes are made to the template format only bugs should be raised here as issues questions or enhancements should instead be raised on our forums description of defect in my testing i found that setting a high ish spi clock speed at least on my nucleo board causes all spi transactions to fail with spi event error however if i set the clock speed to or lower my driver works as expected after some investigation from the debugger i found that it s entering in the hal s spi interrupt function i believe i understand why this error is happening it comes from the way that stmicro has implemented async spi using interrupts their setup relies on the processor being interrupted by the spi peripheral each time a byte has been transmitted the interrupt handler hal spi irqhandler enqueues the next byte in software however it seems that once you go above a certain not actually that high spi frequency the interrupt simply can t keep up anymore this then causes the overflow error to be generated by the hardware since the first byte of the message doesn t get transferred by the time the second byte is read as far as what can be done about this i recognize that it s to some extent a design limitation however at minimum it would be very nice if it were documented as an issue somewhere such as the docstring for spi transfer maybe just a note that some processors have a limit on the maximum bus clock speed that can be reached when using asynchronous spi especially if other interrupts are triggered during the transfer if the spi transfer fails try reducing the clock speed even better would be to add more detailed error codes including something like spi event overrun that indicate to the user what has happened there also ways you could try to address the issue directly maybe there s some way to optimize the spi irq handler so it requires less instructions to copy the byte and get out currently the spi irq seems to take about instructions based on the max achievable speed which seems like more than would strictly be needed even better however would be to enable the use of dma for high speed spi transactions which wouldn t tax the cpu at all i ve heard that it s supposed to be on the roadmap soon right so fingers crossed target s affected by this defect tested on i assume it affects all stmicro chips though the max usable frequency depends on the processor clock toolchain s name and version displaying this defect tested with gcc arm using the standard debug profile what version of mbed os are you using tag or sha mbed os what version s of tools are you using list all that apply e g mbed cli i m using mbed cmake as my build system but this should be identical to mbed cli how is this defect reproduced create an spi object and set its clock speed to a value of hz or more start an asynchronous transfer to read bytes from any sensor or other peripheral observe spi event error being passed to the callback
| 0
|
438,639
| 30,653,611,853
|
IssuesEvent
|
2023-07-25 10:32:11
|
josdem/status-catcher
|
https://api.github.com/repos/josdem/status-catcher
|
closed
|
Add project documentation
|
documentation enhancement in progress
|
As a **user**, I want to add a `readme` file **so that** I can have documentation about this project.
**Acceptance Criteria**
Documentation with this information:
- Project software requirements
- Build instructions
- Execute instructions
- Run tests instructions
- Run
- References
|
1.0
|
Add project documentation - As a **user**, I want to add a `readme` file **so that** I can have documentation about this project.
**Acceptance Criteria**
Documentation with this information:
- Project software requirements
- Build instructions
- Execute instructions
- Run tests instructions
- Run
- References
|
non_test
|
add project documentation as a user i want to add a readme file so that i can have documentation about this project acceptance criteria documentation with this information project software requirements build instructions execute instructions run tests instructions run references
| 0
|
158,965
| 6,038,188,488
|
IssuesEvent
|
2017-06-09 20:43:52
|
DCLP/dclpxsltbox
|
https://api.github.com/repos/DCLP/dclpxsltbox
|
closed
|
unmerged commits on deprecated "dclp" branch of DCLP/idp.data repository
|
priority: high task
|
While working on #177, I discovered that @kirchnerf and @anagnosis have made 5 commits to [the "dclp" branch of DCLP/idp.data](https://github.com/DCLP/idp.data/tree/dclp) since 31 March that have not been merged to [the "master" branch](https://github.com/DCLP/idp.data/tree/master). The "dclp" branch was deprecated in February. This change was coordinated with Heidelberg, but we failed to inform Würzburg. Please accept my apologies for that oversight and for the additional work outlined below. I will forward the email trail to you for record.
I would have merged these changes into DCLP/idp.data "master" myself, but [there are merge conflicts. Eighty-nine files are affected.](https://github.com/DCLP/idp.data/compare/master...DCLP:dclp)
Therefore, I request that @HolgerEssler, @anagnosis, and @kirchnerf:
- review these commits,
- merge FROM DCLP/idp.data "master" into DCLP/idp.data "dclp" (i.e., don't change content in "master"),
- resolve the conflicts appropriately, and then
- issue a pull request against DCLP/idp.data "master" (not against papyri.idp.data) for their review and incorporation, assigning the pull request to me.
|
1.0
|
unmerged commits on deprecated "dclp" branch of DCLP/idp.data repository - While working on #177, I discovered that @kirchnerf and @anagnosis have made 5 commits to [the "dclp" branch of DCLP/idp.data](https://github.com/DCLP/idp.data/tree/dclp) since 31 March that have not been merged to [the "master" branch](https://github.com/DCLP/idp.data/tree/master). The "dclp" branch was deprecated in February. This change was coordinated with Heidelberg, but we failed to inform Würzburg. Please accept my apologies for that oversight and for the additional work outlined below. I will forward the email trail to you for record.
I would have merged these changes into DCLP/idp.data "master" myself, but [there are merge conflicts. Eighty-nine files are affected.](https://github.com/DCLP/idp.data/compare/master...DCLP:dclp)
Therefore, I request that @HolgerEssler, @anagnosis, and @kirchnerf:
- review these commits,
- merge FROM DCLP/idp.data "master" into DCLP/idp.data "dclp" (i.e., don't change content in "master"),
- resolve the conflicts appropriately, and then
- issue a pull request against DCLP/idp.data "master" (not against papyri.idp.data) for their review and incorporation, assigning the pull request to me.
|
non_test
|
unmerged commits on deprecated dclp branch of dclp idp data repository while working on i discovered that kirchnerf and anagnosis have made commits to since march that have not been merged to the dclp branch was deprecated in february this change was coordinated with heidelberg but we failed to inform würzburg please accept my apologies for that oversight and for the additional work outlined below i will forward the email trail to you for record i would have merged these changes into dclp idp data master myself but therefore i request that holgeressler anagnosis and kirchnerf review these commits merge from dclp idp data master into dclp idp data dclp i e don t change content in master resolve the conflicts appropriately and then issue a pull request against dclp idp data master not against papyri idp data for their review and incorporation assigning the pull request to me
| 0
|
134,211
| 12,577,570,645
|
IssuesEvent
|
2020-06-09 09:45:21
|
GoogleCloudPlatform/serverless-photosharing-workshop
|
https://api.github.com/repos/GoogleCloudPlatform/serverless-photosharing-workshop
|
closed
|
Add cleanup steps at the end of each codelabs
|
documentation enhancement
|
Once users have gone through each codelab, we could add an extra section where we mention steps to get rid of all the resources, to avoid recurring costs.
|
1.0
|
Add cleanup steps at the end of each codelabs - Once users have gone through each codelab, we could add an extra section where we mention steps to get rid of all the resources, to avoid recurring costs.
|
non_test
|
add cleanup steps at the end of each codelabs once users have gone through each codelab we could add an extra section where we mention steps to get rid of all the resources to avoid recurring costs
| 0
|
359,094
| 25,219,324,297
|
IssuesEvent
|
2022-11-14 11:32:55
|
nicolasdaudin/pulpito
|
https://api.github.com/repos/nicolasdaudin/pulpito
|
closed
|
Nettoyer le code
|
documentation enhancement api ux
|
- [x] Chercher si il y a un indicateur "automatisé" de clean code
- [x] Relire les règles de Clean Code https://medium.com/swlh/clean-code-4-rules-of-simple-design-f86b066ee43d et https://gist.github.com/wojteklu/73c6914cc446146b8b533c0988cf8d29
- [x] Nettoyer le code et le rendre plus clean.
- [x] dans le matchingItem.addEventListener('click', function (e) { on pourrait peut-être nettoyer utiliser une closure non?
- [x] passer à import au lieu de require?
- [x] vérifier le JS du front-end, est-ce DRY?
- [x] rajouter un validateur pour cheapestWeekend (voir destinationsRoute et validatorService
|
1.0
|
Nettoyer le code -
- [x] Chercher si il y a un indicateur "automatisé" de clean code
- [x] Relire les règles de Clean Code https://medium.com/swlh/clean-code-4-rules-of-simple-design-f86b066ee43d et https://gist.github.com/wojteklu/73c6914cc446146b8b533c0988cf8d29
- [x] Nettoyer le code et le rendre plus clean.
- [x] dans le matchingItem.addEventListener('click', function (e) { on pourrait peut-être nettoyer utiliser une closure non?
- [x] passer à import au lieu de require?
- [x] vérifier le JS du front-end, est-ce DRY?
- [x] rajouter un validateur pour cheapestWeekend (voir destinationsRoute et validatorService
|
non_test
|
nettoyer le code chercher si il y a un indicateur automatisé de clean code relire les règles de clean code et nettoyer le code et le rendre plus clean dans le matchingitem addeventlistener click function e on pourrait peut être nettoyer utiliser une closure non passer à import au lieu de require vérifier le js du front end est ce dry rajouter un validateur pour cheapestweekend voir destinationsroute et validatorservice
| 0
|
200,909
| 15,164,531,226
|
IssuesEvent
|
2021-02-12 13:52:27
|
arturo-lang/arturo
|
https://api.github.com/repos/arturo-lang/arturo
|
closed
|
[Collections\extend] verify functionality
|
library todo unit-test
|
[Collections\extend] verify functionality
https://github.com/arturo-lang/arturo/blob/aba9af1045a5008bf4669dbe3203164d79d82963/src/library/Collections.nim#L240
```text
else: discard
# TODO(Collections\extend) verify functionality
# labels: library, unit-test
builtin "extend",
alias = unaliased,
rule = PrefixPrecedence,
```
8b6be74e1eb0c9a69d0adcf4442b20d958791b53
|
1.0
|
[Collections\extend] verify functionality - [Collections\extend] verify functionality
https://github.com/arturo-lang/arturo/blob/aba9af1045a5008bf4669dbe3203164d79d82963/src/library/Collections.nim#L240
```text
else: discard
# TODO(Collections\extend) verify functionality
# labels: library, unit-test
builtin "extend",
alias = unaliased,
rule = PrefixPrecedence,
```
8b6be74e1eb0c9a69d0adcf4442b20d958791b53
|
test
|
verify functionality verify functionality text else discard todo collections extend verify functionality labels library unit test builtin extend alias unaliased rule prefixprecedence
| 1
|
292,897
| 25,249,252,294
|
IssuesEvent
|
2022-11-15 13:32:29
|
rancher/dashboard
|
https://api.github.com/repos/rancher/dashboard
|
closed
|
In namespace selection dropdown, only "Create a New Namespace" should be blue
|
kind/bug [zube]: To Test
|
I opened this issue based on Kenneth's feedback from a demo.
Currently, all list items in the namespace dropdown are blue:
<img width="780" alt="Screen Shot 2022-07-19 at 6 32 41 PM" src="https://user-images.githubusercontent.com/20599230/179876449-b59782ff-8491-41df-87e7-15cc6731dadd.png">
But if only the "Create a New Namespace" was blue and the others were a normal text color, the creation action would stand out more.
|
1.0
|
In namespace selection dropdown, only "Create a New Namespace" should be blue - I opened this issue based on Kenneth's feedback from a demo.
Currently, all list items in the namespace dropdown are blue:
<img width="780" alt="Screen Shot 2022-07-19 at 6 32 41 PM" src="https://user-images.githubusercontent.com/20599230/179876449-b59782ff-8491-41df-87e7-15cc6731dadd.png">
But if only the "Create a New Namespace" was blue and the others were a normal text color, the creation action would stand out more.
|
test
|
in namespace selection dropdown only create a new namespace should be blue i opened this issue based on kenneth s feedback from a demo currently all list items in the namespace dropdown are blue img width alt screen shot at pm src but if only the create a new namespace was blue and the others were a normal text color the creation action would stand out more
| 1
|
20,288
| 13,792,660,530
|
IssuesEvent
|
2020-10-09 13:54:53
|
bkochuna/ners570f20-Lab06
|
https://api.github.com/repos/bkochuna/ners570f20-Lab06
|
closed
|
Choose a programming language(s)
|
infrastructure
|
Prior to getting too far into the project, the team should decide on using one or more of C, C++, or Fortran for implementation of the code.
This issue is about having that discussion to decide which language(s) to use.
|
1.0
|
Choose a programming language(s) - Prior to getting too far into the project, the team should decide on using one or more of C, C++, or Fortran for implementation of the code.
This issue is about having that discussion to decide which language(s) to use.
|
non_test
|
choose a programming language s prior to getting too far into the project the team should decide on using one or more of c c or fortran for implementation of the code this issue is about having that discussion to decide which language s to use
| 0
|
196,252
| 6,926,210,583
|
IssuesEvent
|
2017-11-30 18:22:10
|
minio/minio
|
https://api.github.com/repos/minio/minio
|
closed
|
Minio behind nginx
|
priority: medium
|
I have set up minio using [these instructions](https://github.com/minio/minio-service/tree/master/linux-systemd) and have proxied it behind nginx as per [this guide](https://github.com/minio/cookbook/blob/master/docs/setup-nginx-proxy-with-minio.md#non-root-configuration) using this block:
```
location ~^/ix35-s3 {
proxy_buffering off;
proxy_set_header Host $http_host;
proxy_pass http://127.0.0.1:9199;
}
```
but when I access `https://<external hostname>/ix35-s3` I get a `NoSuchBucket` error.
What dd I do wrong, and/or where can I find the logs that will help me find out what I did wrong? The logile I pointed minio to from the config.json is created but remains empty, and `journalctl` just shows
```
okt 31 09:32:07 37-97-179-209.colo.transip.net minio[16439]: Created minio configuration file successfully at <home>/.minio
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: Endpoint: http://127.0.0.1:9199
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: AccessKey: <key>
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: SecretKey: <secret>
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: Browser Access:
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: http://127.0.0.1:9199
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: Command-line Access: https://docs.minio.io/docs/minio-client-quickstart-guide
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: $ mc config host add myminio http://127.0.0.1:9199 <key> <secret>
```
I'm running CentOS Linux release 7.3.1611 (Core)
|
1.0
|
Minio behind nginx - I have set up minio using [these instructions](https://github.com/minio/minio-service/tree/master/linux-systemd) and have proxied it behind nginx as per [this guide](https://github.com/minio/cookbook/blob/master/docs/setup-nginx-proxy-with-minio.md#non-root-configuration) using this block:
```
location ~^/ix35-s3 {
proxy_buffering off;
proxy_set_header Host $http_host;
proxy_pass http://127.0.0.1:9199;
}
```
but when I access `https://<external hostname>/ix35-s3` I get a `NoSuchBucket` error.
What dd I do wrong, and/or where can I find the logs that will help me find out what I did wrong? The logile I pointed minio to from the config.json is created but remains empty, and `journalctl` just shows
```
okt 31 09:32:07 37-97-179-209.colo.transip.net minio[16439]: Created minio configuration file successfully at <home>/.minio
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: Endpoint: http://127.0.0.1:9199
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: AccessKey: <key>
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: SecretKey: <secret>
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: Browser Access:
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: http://127.0.0.1:9199
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: Command-line Access: https://docs.minio.io/docs/minio-client-quickstart-guide
okt 31 09:32:08 37-97-179-209.colo.transip.net minio[16439]: $ mc config host add myminio http://127.0.0.1:9199 <key> <secret>
```
I'm running CentOS Linux release 7.3.1611 (Core)
|
non_test
|
minio behind nginx i have set up minio using and have proxied it behind nginx as per using this block location proxy buffering off proxy set header host http host proxy pass but when i access hostname i get a nosuchbucket error what dd i do wrong and or where can i find the logs that will help me find out what i did wrong the logile i pointed minio to from the config json is created but remains empty and journalctl just shows okt colo transip net minio created minio configuration file successfully at minio okt colo transip net minio endpoint okt colo transip net minio accesskey okt colo transip net minio secretkey okt colo transip net minio browser access okt colo transip net minio okt colo transip net minio command line access okt colo transip net minio mc config host add myminio i m running centos linux release core
| 0
|
102,427
| 8,827,020,758
|
IssuesEvent
|
2019-01-03 06:44:31
|
humera987/FXLabs-Test-Automation
|
https://api.github.com/repos/humera987/FXLabs-Test-Automation
|
closed
|
Testing 3/1 : ApiV1BotClustersGetNoParams
|
Testing 3/1
|
Project : Testing 3/1
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NThjNjI1NmEtMzFkNC00ZTFkLTg4MjktZDQyMmM5ZTM1NDQw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 03 Jan 2019 06:16:56 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/bot-clusters
Request :
Response :
{
"timestamp" : "2019-01-03T06:16:57.201+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/bot-clusters"
}
Logs :
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : URL [http://13.56.210.25/api/v1/api/v1/bot-clusters]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Method [GET]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Request []
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Response [{
"timestamp" : "2019-01-03T06:16:02.656+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/bot-clusters"
}]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Response-Headers [{X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=ZmIxYzdlOTUtYzNhMy00MjRlLWI1NzItNGY4OGQyZjhmYjE2; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 03 Jan 2019 06:16:02 GMT]}]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : StatusCode [404]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Time [374]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Size [147]
2019-01-03 06:16:02 ERROR [ApiV1BotClustersGetNoParams] : Assertion [@StatusCode == 200] resolved-to [404 == 200] result [Failed]
--- FX Bot ---
|
1.0
|
Testing 3/1 : ApiV1BotClustersGetNoParams - Project : Testing 3/1
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NThjNjI1NmEtMzFkNC00ZTFkLTg4MjktZDQyMmM5ZTM1NDQw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 03 Jan 2019 06:16:56 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/bot-clusters
Request :
Response :
{
"timestamp" : "2019-01-03T06:16:57.201+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/bot-clusters"
}
Logs :
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : URL [http://13.56.210.25/api/v1/api/v1/bot-clusters]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Method [GET]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Request []
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Response [{
"timestamp" : "2019-01-03T06:16:02.656+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/bot-clusters"
}]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Response-Headers [{X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=ZmIxYzdlOTUtYzNhMy00MjRlLWI1NzItNGY4OGQyZjhmYjE2; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 03 Jan 2019 06:16:02 GMT]}]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : StatusCode [404]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Time [374]
2019-01-03 06:16:02 DEBUG [ApiV1BotClustersGetNoParams] : Size [147]
2019-01-03 06:16:02 ERROR [ApiV1BotClustersGetNoParams] : Assertion [@StatusCode == 200] resolved-to [404 == 200] result [Failed]
--- FX Bot ---
|
test
|
testing project testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api bot clusters logs debug url debug method debug request debug request headers accept authorization debug response timestamp status error not found message no message available path api api bot clusters debug response headers x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date debug statuscode debug time debug size error assertion resolved to result fx bot
| 1
|
19,329
| 10,346,983,304
|
IssuesEvent
|
2019-09-04 16:20:19
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Manual installation powershell command not working for linux
|
Pri2 cxp product-question security-center/svc triaged
|
Hi,
I have tried to install the OMS agent on Linux VM and the powershell command not working I got Error Message: "Set-AzVMExtension : Unterminated string. Expected delimiter: ". Path 'workspaceKey', line 3, position 2.
At line:9 char:2
+ Set-AzVMExtension -ResourceGroupName $vm1.ResourceGroupName -VMName ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : CloseError: (:) [Set-AzVMExtension], JsonReaderException
+ FullyQualifiedErrorId : Microsoft.Azure.Commands.Compute.SetAzureVMExtensionCommand"
I have fixed the issue based on the the Set-AzVMExtension documentation.
Changing the variable and the command to that.
$PublicConf = @{"workspaceId"="WorkspaceID value"}
$PrivateConf = @{"workspaceKey"="<Primary key value>”}
Set-AzVMExtension -ResourceGroupName $vm1.ResourceGroupName -VMName $vm1.Name -Name "OmsAgentForLinux" -Publisher "Microsoft.EnterpriseCloud.Monitoring" -ExtensionType "OmsAgentForLinux" -TypeHandlerVersion '1.0' -Location $vm1.Location -Settings $PublicConf -ProtectedSettings $PrivateConf -ForceRerun True
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8267e218-3748-2820-3eee-80479705eec2
* Version Independent ID: 8c6d5fce-6383-a0f8-7df3-f5e0a77bbad6
* Content: [Data Collection in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-enable-data-collection#feedback)
* Content Source: [articles/security-center/security-center-enable-data-collection.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-enable-data-collection.md)
* Service: **security-center**
* GitHub Login: @monhaber
* Microsoft Alias: **v-mohabe**
|
True
|
Manual installation powershell command not working for linux - Hi,
I have tried to install the OMS agent on Linux VM and the powershell command not working I got Error Message: "Set-AzVMExtension : Unterminated string. Expected delimiter: ". Path 'workspaceKey', line 3, position 2.
At line:9 char:2
+ Set-AzVMExtension -ResourceGroupName $vm1.ResourceGroupName -VMName ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : CloseError: (:) [Set-AzVMExtension], JsonReaderException
+ FullyQualifiedErrorId : Microsoft.Azure.Commands.Compute.SetAzureVMExtensionCommand"
I have fixed the issue based on the the Set-AzVMExtension documentation.
Changing the variable and the command to that.
$PublicConf = @{"workspaceId"="WorkspaceID value"}
$PrivateConf = @{"workspaceKey"="<Primary key value>”}
Set-AzVMExtension -ResourceGroupName $vm1.ResourceGroupName -VMName $vm1.Name -Name "OmsAgentForLinux" -Publisher "Microsoft.EnterpriseCloud.Monitoring" -ExtensionType "OmsAgentForLinux" -TypeHandlerVersion '1.0' -Location $vm1.Location -Settings $PublicConf -ProtectedSettings $PrivateConf -ForceRerun True
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 8267e218-3748-2820-3eee-80479705eec2
* Version Independent ID: 8c6d5fce-6383-a0f8-7df3-f5e0a77bbad6
* Content: [Data Collection in Azure Security Center](https://docs.microsoft.com/en-us/azure/security-center/security-center-enable-data-collection#feedback)
* Content Source: [articles/security-center/security-center-enable-data-collection.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-enable-data-collection.md)
* Service: **security-center**
* GitHub Login: @monhaber
* Microsoft Alias: **v-mohabe**
|
non_test
|
manual installation powershell command not working for linux hi i have tried to install the oms agent on linux vm and the powershell command not working i got error message set azvmextension unterminated string expected delimiter path workspacekey line position at line char set azvmextension resourcegroupname resourcegroupname vmname categoryinfo closeerror jsonreaderexception fullyqualifiederrorid microsoft azure commands compute setazurevmextensioncommand i have fixed the issue based on the the set azvmextension documentation changing the variable and the command to that publicconf workspaceid workspaceid value privateconf workspacekey lt primary key value gt ” set azvmextension resourcegroupname resourcegroupname vmname name name omsagentforlinux publisher microsoft enterprisecloud monitoring extensiontype omsagentforlinux typehandlerversion location location settings publicconf protectedsettings privateconf forcererun true document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service security center github login monhaber microsoft alias v mohabe
| 0
|
79,121
| 15,147,778,084
|
IssuesEvent
|
2021-02-11 09:40:09
|
Regalis11/Barotrauma
|
https://api.github.com/repos/Regalis11/Barotrauma
|
opened
|
No stack restrictions on older saves
|
Bug Code
|
If you are running a older save you are able to place full stacks on items which otherwise has a restriction in place, for example 8 oxygen tanks in to a diving mask.
|
1.0
|
No stack restrictions on older saves - If you are running a older save you are able to place full stacks on items which otherwise has a restriction in place, for example 8 oxygen tanks in to a diving mask.
|
non_test
|
no stack restrictions on older saves if you are running a older save you are able to place full stacks on items which otherwise has a restriction in place for example oxygen tanks in to a diving mask
| 0
|
794,502
| 28,038,457,250
|
IssuesEvent
|
2023-03-28 16:38:26
|
ImagingDataCommons/slim
|
https://api.github.com/repos/ImagingDataCommons/slim
|
reopened
|
Tile grid shifts upon zooming
|
bug idc:priority
|
From Steve: as Bill reported there's a shift to the down-right when high res tiles come in which may be a problem with the coordinate systems or maybe the data but probably the viewer.
|
1.0
|
Tile grid shifts upon zooming - From Steve: as Bill reported there's a shift to the down-right when high res tiles come in which may be a problem with the coordinate systems or maybe the data but probably the viewer.
|
non_test
|
tile grid shifts upon zooming from steve as bill reported there s a shift to the down right when high res tiles come in which may be a problem with the coordinate systems or maybe the data but probably the viewer
| 0
|
222,013
| 17,382,353,583
|
IssuesEvent
|
2021-08-01 00:54:06
|
LuanMiranda77/agility-e-commerce-api
|
https://api.github.com/repos/LuanMiranda77/agility-e-commerce-api
|
opened
|
#01 Criatela de login
|
Disponivel para QA Testando
|
Tela de login de media e alta fidelidade, axulia o usuario para fazer login
┆Issue is synchronized with this [Trello card](https://trello.com/c/cLa65JHu) by [Unito](https://www.unito.io)
|
1.0
|
#01 Criatela de login - Tela de login de media e alta fidelidade, axulia o usuario para fazer login
┆Issue is synchronized with this [Trello card](https://trello.com/c/cLa65JHu) by [Unito](https://www.unito.io)
|
test
|
criatela de login tela de login de media e alta fidelidade axulia o usuario para fazer login ┆issue is synchronized with this by
| 1
|
237,481
| 19,649,288,620
|
IssuesEvent
|
2022-01-10 03:40:15
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
No completions for qualified name reference within a union in return type descriptor
|
Type/Bug Team/LanguageServer Area/Completion GA-Test-Hackathon
|
**Description:**
Consider the following code:
```ballerina
function temp() returns io:<cursor>|io:Error {
}
```
At `<cursor>` we should provide completions from module `io`.
**Steps to reproduce:**
See description
**Affected Versions:**
Swan Lake GA Snapshot Build 1
|
1.0
|
No completions for qualified name reference within a union in return type descriptor - **Description:**
Consider the following code:
```ballerina
function temp() returns io:<cursor>|io:Error {
}
```
At `<cursor>` we should provide completions from module `io`.
**Steps to reproduce:**
See description
**Affected Versions:**
Swan Lake GA Snapshot Build 1
|
test
|
no completions for qualified name reference within a union in return type descriptor description consider the following code ballerina function temp returns io io error at we should provide completions from module io steps to reproduce see description affected versions swan lake ga snapshot build
| 1
|
275,548
| 23,921,742,050
|
IssuesEvent
|
2022-09-09 17:37:10
|
wazuh/wazuh-qa
|
https://api.github.com/repos/wazuh/wazuh-qa
|
closed
|
Nightly report: 2022-08-22
|
team/qa test/integration test/nightly type/nightly-test-failure subteam/qa-rainbow
|
## Data
| Date| Commit | Commit title | Build | Notes |
|--|--|--|--|--|
| 2022-08-22| [6cce5b8](6cce5b8465ecdc8008950e6fc0a2bfd0ebb251c1) | Merge branch '4.4' | [#1175](https://ci.wazuh.info/job/test_nightly/1175/) | 4.5 |
## Description
- [x] Fix Logcollector IT: test_ignore_works_over_restrict.py not received expected message [#3195](https://github.com/wazuh/wazuh-qa/issues/3195)
- [ ] Reliability tests: Generic Environent launcher error [#3465](https://github.com/wazuh/wazuh-jenkins/issues/3465) (Reported to wazuh-jenkins)
- [ ] IT pipeline: Unexpected Windows test launch error [#3487](https://github.com/wazuh/wazuh-jenkins/issues/3487) (Reported to wazuh-jenkins)
**Note**: This issue also has information about Nightly launched previous than 2022-08-09
|
3.0
|
Nightly report: 2022-08-22 - ## Data
| Date| Commit | Commit title | Build | Notes |
|--|--|--|--|--|
| 2022-08-22| [6cce5b8](6cce5b8465ecdc8008950e6fc0a2bfd0ebb251c1) | Merge branch '4.4' | [#1175](https://ci.wazuh.info/job/test_nightly/1175/) | 4.5 |
## Description
- [x] Fix Logcollector IT: test_ignore_works_over_restrict.py not received expected message [#3195](https://github.com/wazuh/wazuh-qa/issues/3195)
- [ ] Reliability tests: Generic Environent launcher error [#3465](https://github.com/wazuh/wazuh-jenkins/issues/3465) (Reported to wazuh-jenkins)
- [ ] IT pipeline: Unexpected Windows test launch error [#3487](https://github.com/wazuh/wazuh-jenkins/issues/3487) (Reported to wazuh-jenkins)
**Note**: This issue also has information about Nightly launched previous than 2022-08-09
|
test
|
nightly report data date commit commit title build notes merge branch description fix logcollector it test ignore works over restrict py not received expected message reliability tests generic environent launcher error reported to wazuh jenkins it pipeline unexpected windows test launch error reported to wazuh jenkins note this issue also has information about nightly launched previous than
| 1
|
319,169
| 27,353,822,450
|
IssuesEvent
|
2023-02-27 11:25:57
|
oras-project/oras
|
https://api.github.com/repos/oras-project/oras
|
closed
|
Building up E2E Tests for ORAS
|
E2E testing
|
### Discussed in https://github.com/oras-project/oras/discussions/485
<div type='discussions-op-text'>
<sup>Originally posted by **qweeah** August 9, 2022</sup>
## What is Needed
End-to-end tests that continuously validate the functional correctness of oras features.
## Why it's Needed
- As the number of supported commands grows, it's impossible to manually validate all the features before releasing.
- Although we have unit tests for CLI, things might still go wrong after integration. Having E2E tests would at least cover high-value use cases and reduce regression.
## How to
### Setup Test Environment
The E2E tests should run against oras CLI binary with [oras-distribution](https://github.com/oras-project/distribution) as the backend.
### Choose Test Framework
Technically, there are two different directions of tools we use to build E2E test suite:
#### 1) Use golang-native test framework to run test based on Cobra command execution calls
[ginkgo](https://github.com/onsi/ginkgo) + [gomega](https://github.com/onsi/gomega)
- Pros:
- Based on golang standard testing package, which matches tech-stack of most active contributors.
- No change in github actions.
- Applied by many state-of-the-art CLI projects like [kubectl](https://github.com/kubernetes/kubernetes) and [frp](https://github.com/fatedier/frp) so we have a lot of existing good practices to learn from.
- Cons:
- Use Cobra command execution call to mock user input, not 100% E2E.
- Brings new dependencies in only for testing.
- Won't be able to be reused by oras-py or other registry CLI tools sharing same UX as oras.
#### 2) Use common Behaviour-Driven Development (BDD) tools to test the built out binary
Gherkin + [godog](https://github.com/cucumber/godog)
- Pros:
- Testing oras as an CLI application, 100% E2E.
- Idomatic BDD framework.
- The latest cucumber binary can be downloaded during test time, which brings no depedency into oras.
- Test cases(specs) can be reused by oras-py or other registry CLI tools sharing same UX as oras.
- Cons:
- Extra one-time setup effort in Github actions.
- Not Golang-native. (As far as I know) No expert in the current active contributors.
Both options provides expressive assertion that can meet oras' testing need.
**Future Works**
Below items are valuable but not in scope of this discussion:
- Benchmark Test & Profiling.
- E2E tests based on registry services from different cloud providers.
If you have any other ideas or suggestions on building the E2E test suite for oras, feel free to response below. Any comments are highly appreciated 🍻 </div>
|
1.0
|
Building up E2E Tests for ORAS - ### Discussed in https://github.com/oras-project/oras/discussions/485
<div type='discussions-op-text'>
<sup>Originally posted by **qweeah** August 9, 2022</sup>
## What is Needed
End-to-end tests that continuously validate the functional correctness of oras features.
## Why it's Needed
- As the number of supported commands grows, it's impossible to manually validate all the features before releasing.
- Although we have unit tests for CLI, things might still go wrong after integration. Having E2E tests would at least cover high-value use cases and reduce regression.
## How to
### Setup Test Environment
The E2E tests should run against oras CLI binary with [oras-distribution](https://github.com/oras-project/distribution) as the backend.
### Choose Test Framework
Technically, there are two different directions of tools we use to build E2E test suite:
#### 1) Use golang-native test framework to run test based on Cobra command execution calls
[ginkgo](https://github.com/onsi/ginkgo) + [gomega](https://github.com/onsi/gomega)
- Pros:
- Based on golang standard testing package, which matches tech-stack of most active contributors.
- No change in github actions.
- Applied by many state-of-the-art CLI projects like [kubectl](https://github.com/kubernetes/kubernetes) and [frp](https://github.com/fatedier/frp) so we have a lot of existing good practices to learn from.
- Cons:
- Use Cobra command execution call to mock user input, not 100% E2E.
- Brings new dependencies in only for testing.
- Won't be able to be reused by oras-py or other registry CLI tools sharing same UX as oras.
#### 2) Use common Behaviour-Driven Development (BDD) tools to test the built out binary
Gherkin + [godog](https://github.com/cucumber/godog)
- Pros:
- Testing oras as an CLI application, 100% E2E.
- Idomatic BDD framework.
- The latest cucumber binary can be downloaded during test time, which brings no depedency into oras.
- Test cases(specs) can be reused by oras-py or other registry CLI tools sharing same UX as oras.
- Cons:
- Extra one-time setup effort in Github actions.
- Not Golang-native. (As far as I know) No expert in the current active contributors.
Both options provides expressive assertion that can meet oras' testing need.
**Future Works**
Below items are valuable but not in scope of this discussion:
- Benchmark Test & Profiling.
- E2E tests based on registry services from different cloud providers.
If you have any other ideas or suggestions on building the E2E test suite for oras, feel free to response below. Any comments are highly appreciated 🍻 </div>
|
test
|
building up tests for oras discussed in originally posted by qweeah august what is needed end to end tests that continuously validate the functional correctness of oras features why it s needed as the number of supported commands grows it s impossible to manually validate all the features before releasing although we have unit tests for cli things might still go wrong after integration having tests would at least cover high value use cases and reduce regression how to setup test environment the tests should run against oras cli binary with as the backend choose test framework technically there are two different directions of tools we use to build test suite use golang native test framework to run test based on cobra command execution calls pros based on golang standard testing package which matches tech stack of most active contributors no change in github actions applied by many state of the art cli projects like and so we have a lot of existing good practices to learn from cons use cobra command execution call to mock user input not brings new dependencies in only for testing won t be able to be reused by oras py or other registry cli tools sharing same ux as oras use common behaviour driven development bdd tools to test the built out binary gherkin pros testing oras as an cli application idomatic bdd framework the latest cucumber binary can be downloaded during test time which brings no depedency into oras test cases specs can be reused by oras py or other registry cli tools sharing same ux as oras cons extra one time setup effort in github actions not golang native as far as i know no expert in the current active contributors both options provides expressive assertion that can meet oras testing need future works below items are valuable but not in scope of this discussion benchmark test profiling tests based on registry services from different cloud providers if you have any other ideas or suggestions on building the test suite for oras feel free to response below any comments are highly appreciated 🍻
| 1
|
15,220
| 26,718,761,875
|
IssuesEvent
|
2023-01-28 21:46:45
|
remlse/buenzlimarks
|
https://api.github.com/repos/remlse/buenzlimarks
|
closed
|
Persist bookmarks across browser sessions
|
Requirement
|
As a user, I want to persist bookmarks across browser sessions, so that my added bookmarks from an earlier session will show up in future sessions.
|
1.0
|
Persist bookmarks across browser sessions - As a user, I want to persist bookmarks across browser sessions, so that my added bookmarks from an earlier session will show up in future sessions.
|
non_test
|
persist bookmarks across browser sessions as a user i want to persist bookmarks across browser sessions so that my added bookmarks from an earlier session will show up in future sessions
| 0
|
278,386
| 24,150,677,208
|
IssuesEvent
|
2022-09-22 00:09:10
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
Failing test: Jest Tests.x-pack/plugins/security_solution/public/management/components/endpoint_responder - When using processes action from response actions console should call `running-procs` api when command is entered
|
failed-test
|
A test failed on a tracked branch
```
Error: thrown: "Exceeded timeout of 5000 ms for a test.
Use jest.setTimeout(newTimeout) to increase the timeout value, if this is a long-running test."
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/x-pack/plugins/security_solution/public/management/components/endpoint_responder/get_processes_action.test.tsx:56:3
at _dispatchDescribe (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/index.js:67:26)
at describe (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/index.js:30:5)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/x-pack/plugins/security_solution/public/management/components/endpoint_responder/get_processes_action.test.tsx:20:1)
at Runtime._execModule (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runtime/build/index.js:1299:24)
at Runtime._loadModule (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runtime/build/index.js:898:12)
at Runtime.requireModule (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runtime/build/index.js:746:10)
at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:106:13)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/testWorker.js:133:12)
```
First failure: [CI Build - 8.4](https://buildkite.com/elastic/kibana-on-merge/builds/21225#01836262-2ac7-457b-b072-232a56fac77c)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/security_solution/public/management/components/endpoint_responder","test.name":"When using processes action from response actions console should call `running-procs` api when command is entered","test.failCount":1}} -->
|
1.0
|
Failing test: Jest Tests.x-pack/plugins/security_solution/public/management/components/endpoint_responder - When using processes action from response actions console should call `running-procs` api when command is entered - A test failed on a tracked branch
```
Error: thrown: "Exceeded timeout of 5000 ms for a test.
Use jest.setTimeout(newTimeout) to increase the timeout value, if this is a long-running test."
at /var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/x-pack/plugins/security_solution/public/management/components/endpoint_responder/get_processes_action.test.tsx:56:3
at _dispatchDescribe (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/index.js:67:26)
at describe (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/index.js:30:5)
at Object.<anonymous> (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/x-pack/plugins/security_solution/public/management/components/endpoint_responder/get_processes_action.test.tsx:20:1)
at Runtime._execModule (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runtime/build/index.js:1299:24)
at Runtime._loadModule (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runtime/build/index.js:898:12)
at Runtime.requireModule (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runtime/build/index.js:746:10)
at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:106:13)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (node:internal/process/task_queues:96:5)
at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:380:16)
at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:472:34)
at Object.worker (/var/lib/buildkite-agent/builds/kb-n2-4-spot-4ea9c5b308cc4f7a/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/testWorker.js:133:12)
```
First failure: [CI Build - 8.4](https://buildkite.com/elastic/kibana-on-merge/builds/21225#01836262-2ac7-457b-b072-232a56fac77c)
<!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/security_solution/public/management/components/endpoint_responder","test.name":"When using processes action from response actions console should call `running-procs` api when command is entered","test.failCount":1}} -->
|
test
|
failing test jest tests x pack plugins security solution public management components endpoint responder when using processes action from response actions console should call running procs api when command is entered a test failed on a tracked branch error thrown exceeded timeout of ms for a test use jest settimeout newtimeout to increase the timeout value if this is a long running test at var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins security solution public management components endpoint responder get processes action test tsx at dispatchdescribe var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build index js at describe var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build index js at object var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins security solution public management components endpoint responder get processes action test tsx at runtime execmodule var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runtime build index js at runtime loadmodule var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runtime build index js at runtime requiremodule var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runtime build index js at jestadapter var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapter js at runmicrotasks at processticksandrejections node internal process task queues at runtestinternal var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js at object worker var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build testworker js first failure
| 1
|
328,593
| 28,126,752,612
|
IssuesEvent
|
2023-03-31 18:26:25
|
urban-stack/OpenZoning
|
https://api.github.com/repos/urban-stack/OpenZoning
|
closed
|
what are our steps for getting to a releasable v0.1.0 geo standard?
|
question geo-standard Minneapolis test
|
steps
- ...
sub-question
- ...
|
1.0
|
what are our steps for getting to a releasable v0.1.0 geo standard? - steps
- ...
sub-question
- ...
|
test
|
what are our steps for getting to a releasable geo standard steps sub question
| 1
|
165,889
| 6,287,761,804
|
IssuesEvent
|
2017-07-19 15:33:55
|
edcarroll/ng2-semantic-ui
|
https://api.github.com/repos/edcarroll/ng2-semantic-ui
|
closed
|
bug(popup): view destroyed error on popups
|
comp: popup priority: important type: bug
|
There are a few cases I'm getting the below error when navigating away from a page containing a popup. A simple replication for Chrome is in this plunkr
http://plnkr.co/edit/4IrG2eK8igIubS2bXN4u
Keep clicking the green button quickly, at some point the popup message will flash out then back in, and the cursor will duplicate so there are two cursors side by side by a few pixels. Navigating away by clicking the page2 link will now give the following error
```
App.html:4 ERROR Error: ViewDestroyedError: Attempt to use a destroyed view: detectChanges
at viewDestroyedError (VM884 core.umd.js:8475)
at Object.debugUpdateDirectives [as updateDirectives] (VM884 core.umd.js:13085)
at checkAndUpdateView (VM884 core.umd.js:12268)
at callWithDebugContext (VM884 core.umd.js:13488)
at Object.debugCheckAndUpdateView [as checkAndUpdateView] (VM884 core.umd.js:13028)
at ViewRef_.detectChanges (VM884 core.umd.js:10199)
at eval (VM884 core.umd.js:4837)
at Array.forEach (<anonymous>)
at ApplicationRef_.tick (VM884 core.umd.js:4837)
at eval (VM884 core.umd.js:4715)
```
|
1.0
|
bug(popup): view destroyed error on popups - There are a few cases I'm getting the below error when navigating away from a page containing a popup. A simple replication for Chrome is in this plunkr
http://plnkr.co/edit/4IrG2eK8igIubS2bXN4u
Keep clicking the green button quickly, at some point the popup message will flash out then back in, and the cursor will duplicate so there are two cursors side by side by a few pixels. Navigating away by clicking the page2 link will now give the following error
```
App.html:4 ERROR Error: ViewDestroyedError: Attempt to use a destroyed view: detectChanges
at viewDestroyedError (VM884 core.umd.js:8475)
at Object.debugUpdateDirectives [as updateDirectives] (VM884 core.umd.js:13085)
at checkAndUpdateView (VM884 core.umd.js:12268)
at callWithDebugContext (VM884 core.umd.js:13488)
at Object.debugCheckAndUpdateView [as checkAndUpdateView] (VM884 core.umd.js:13028)
at ViewRef_.detectChanges (VM884 core.umd.js:10199)
at eval (VM884 core.umd.js:4837)
at Array.forEach (<anonymous>)
at ApplicationRef_.tick (VM884 core.umd.js:4837)
at eval (VM884 core.umd.js:4715)
```
|
non_test
|
bug popup view destroyed error on popups there are a few cases i m getting the below error when navigating away from a page containing a popup a simple replication for chrome is in this plunkr keep clicking the green button quickly at some point the popup message will flash out then back in and the cursor will duplicate so there are two cursors side by side by a few pixels navigating away by clicking the link will now give the following error app html error error viewdestroyederror attempt to use a destroyed view detectchanges at viewdestroyederror core umd js at object debugupdatedirectives core umd js at checkandupdateview core umd js at callwithdebugcontext core umd js at object debugcheckandupdateview core umd js at viewref detectchanges core umd js at eval core umd js at array foreach at applicationref tick core umd js at eval core umd js
| 0
|
8,665
| 27,172,058,755
|
IssuesEvent
|
2023-02-17 20:25:06
|
OneDrive/onedrive-api-docs
|
https://api.github.com/repos/OneDrive/onedrive-api-docs
|
closed
|
Ability to get ShareId when listing children?
|
automation:Closed
|
I have several folders that have been shared using createLink - https://docs.microsoft.com/en-us/onedrive/developer/rest-api/api/driveitem_createlink
That works fine, but I'm trying to find a way to list all folders that have been shared, and I also need their shareId. Is there any way to make such query?
`https://graph.microsoft.com/v1.0/me/drive/root/children` doesn't seem to return the shareId for the shared folders, and there's no expand for permissions, shares, etc.
Any help is much appreciated. Thanks.
|
1.0
|
Ability to get ShareId when listing children? - I have several folders that have been shared using createLink - https://docs.microsoft.com/en-us/onedrive/developer/rest-api/api/driveitem_createlink
That works fine, but I'm trying to find a way to list all folders that have been shared, and I also need their shareId. Is there any way to make such query?
`https://graph.microsoft.com/v1.0/me/drive/root/children` doesn't seem to return the shareId for the shared folders, and there's no expand for permissions, shares, etc.
Any help is much appreciated. Thanks.
|
non_test
|
ability to get shareid when listing children i have several folders that have been shared using createlink that works fine but i m trying to find a way to list all folders that have been shared and i also need their shareid is there any way to make such query doesn t seem to return the shareid for the shared folders and there s no expand for permissions shares etc any help is much appreciated thanks
| 0
|
130,982
| 10,677,076,401
|
IssuesEvent
|
2019-10-21 14:51:10
|
stevenschader/kabanero-foundation
|
https://api.github.com/repos/stevenschader/kabanero-foundation
|
opened
|
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-20361-1
|
SVT Kabanero Test Execution Test Execution Record
|
running setup!
Using label defaults
SVT: TER: Kabanero Automation test execution master branch -
https://api.github.com/repos/stevenschader/kabanero-foundation/issues
Successfully created Issue "SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-20361-1"
Successfully created Issue "SVT: TER: Kabanero Automation test execution master branch - Verify: svtcrc-20361-1"
----------------------------------------
Script completed - exiting
|
2.0
|
SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-20361-1 - running setup!
Using label defaults
SVT: TER: Kabanero Automation test execution master branch -
https://api.github.com/repos/stevenschader/kabanero-foundation/issues
Successfully created Issue "SVT: TER: Kabanero Automation test execution master branch - Setup: svtcrc-20361-1"
Successfully created Issue "SVT: TER: Kabanero Automation test execution master branch - Verify: svtcrc-20361-1"
----------------------------------------
Script completed - exiting
|
test
|
svt ter kabanero automation test execution master branch setup svtcrc running setup using label defaults svt ter kabanero automation test execution master branch successfully created issue svt ter kabanero automation test execution master branch setup svtcrc successfully created issue svt ter kabanero automation test execution master branch verify svtcrc script completed exiting
| 1
|
263,973
| 23,093,982,548
|
IssuesEvent
|
2022-07-26 17:38:56
|
projectcontour/contour
|
https://api.github.com/repos/projectcontour/contour
|
closed
|
run Gateway API conformance from our own Go test, using the GWAPI conformance module
|
area/testing area/gateway-api
|
Currently we're running the Gateway API conformance tests by [cloning the Gateway API repo and running the tests from there](https://github.com/projectcontour/contour/blob/main/test/scripts/run-gateway-conformance.sh#L44-L48), however as I understand it the intended way of running them is to have our own Go test that mirrors the one [in the Gateway API repo](https://github.com/kubernetes-sigs/gateway-api/blob/main/conformance/conformance_test.go#L33-L57), and uses the `conformance` Go module. Importantly, this will allow us to customize the [SupportedFeatures](https://github.com/kubernetes-sigs/gateway-api/blob/main/conformance/conformance_test.go#L51-L53); we support HTTPRoute query param matching and should [opt into the conformance test for it](https://github.com/kubernetes-sigs/gateway-api/blob/main/conformance/utils/suite/suite.go#L49-L50).
|
1.0
|
run Gateway API conformance from our own Go test, using the GWAPI conformance module - Currently we're running the Gateway API conformance tests by [cloning the Gateway API repo and running the tests from there](https://github.com/projectcontour/contour/blob/main/test/scripts/run-gateway-conformance.sh#L44-L48), however as I understand it the intended way of running them is to have our own Go test that mirrors the one [in the Gateway API repo](https://github.com/kubernetes-sigs/gateway-api/blob/main/conformance/conformance_test.go#L33-L57), and uses the `conformance` Go module. Importantly, this will allow us to customize the [SupportedFeatures](https://github.com/kubernetes-sigs/gateway-api/blob/main/conformance/conformance_test.go#L51-L53); we support HTTPRoute query param matching and should [opt into the conformance test for it](https://github.com/kubernetes-sigs/gateway-api/blob/main/conformance/utils/suite/suite.go#L49-L50).
|
test
|
run gateway api conformance from our own go test using the gwapi conformance module currently we re running the gateway api conformance tests by however as i understand it the intended way of running them is to have our own go test that mirrors the one and uses the conformance go module importantly this will allow us to customize the we support httproute query param matching and should
| 1
|
186,228
| 14,394,659,720
|
IssuesEvent
|
2020-12-03 01:49:21
|
github-vet/rangeclosure-findings
|
https://api.github.com/repos/github-vet/rangeclosure-findings
|
closed
|
nhdms/guggmeta: Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go; 41 LoC
|
fresh small test
|
Found a possible issue in [nhdms/guggmeta](https://www.github.com/nhdms/guggmeta) at [Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go](https://github.com/nhdms/guggmeta/blob/bb2178df85e7ace01e5cef69ce26abac601ab767/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/nhdms/guggmeta/blob/bb2178df85e7ace01e5cef69ce26abac601ab767/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
<details>
<summary>Click here to show the 41 line(s) of Go which triggered the analyzer.</summary>
```go
for _, count := range []int{10000, 50, 432} {
recs := make(chan *log.Record)
done := make(chan int)
spec := SpeculativeHandler(100, log.ChannelHandler(recs))
go func() {
defer close(done)
expectedCount := int(math.Min(float64(count), float64(100)))
expectedIdx := count - expectedCount
for r := range recs {
if r.Ctx[1] != expectedIdx {
t.Errorf("Bad ctx 'i', got %d expected %d", r.Ctx[1], expectedIdx)
return
}
expectedIdx++
expectedCount--
if expectedCount == 0 {
// got everything we expected
break
}
}
select {
case <-recs:
t.Errorf("got an extra record we shouldn't have!")
default:
}
}()
lg := log.New()
lg.SetHandler(spec)
for i := 0; i < count; i++ {
lg.Debug("test speculative", "i", i)
}
go spec.Flush()
// wait for the go routine to finish
<-done
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable count used in defer or goroutine at line 52
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: bb2178df85e7ace01e5cef69ce26abac601ab767
|
1.0
|
nhdms/guggmeta: Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go; 41 LoC -
Found a possible issue in [nhdms/guggmeta](https://www.github.com/nhdms/guggmeta) at [Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go](https://github.com/nhdms/guggmeta/blob/bb2178df85e7ace01e5cef69ce26abac601ab767/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements
which capture loop variables.
[Click here to see the code in its original context.](https://github.com/nhdms/guggmeta/blob/bb2178df85e7ace01e5cef69ce26abac601ab767/Godeps/_workspace/src/gopkg.in/inconshreveable/log15.v2/ext/ext_test.go#L45-L85)
<details>
<summary>Click here to show the 41 line(s) of Go which triggered the analyzer.</summary>
```go
for _, count := range []int{10000, 50, 432} {
recs := make(chan *log.Record)
done := make(chan int)
spec := SpeculativeHandler(100, log.ChannelHandler(recs))
go func() {
defer close(done)
expectedCount := int(math.Min(float64(count), float64(100)))
expectedIdx := count - expectedCount
for r := range recs {
if r.Ctx[1] != expectedIdx {
t.Errorf("Bad ctx 'i', got %d expected %d", r.Ctx[1], expectedIdx)
return
}
expectedIdx++
expectedCount--
if expectedCount == 0 {
// got everything we expected
break
}
}
select {
case <-recs:
t.Errorf("got an extra record we shouldn't have!")
default:
}
}()
lg := log.New()
lg.SetHandler(spec)
for i := 0; i < count; i++ {
lg.Debug("test speculative", "i", i)
}
go spec.Flush()
// wait for the go routine to finish
<-done
}
```
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable count used in defer or goroutine at line 52
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: bb2178df85e7ace01e5cef69ce26abac601ab767
|
test
|
nhdms guggmeta godeps workspace src gopkg in inconshreveable ext ext test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for count range int recs make chan log record done make chan int spec speculativehandler log channelhandler recs go func defer close done expectedcount int math min count expectedidx count expectedcount for r range recs if r ctx expectedidx t errorf bad ctx i got d expected d r ctx expectedidx return expectedidx expectedcount if expectedcount got everything we expected break select case recs t errorf got an extra record we shouldn t have default lg log new lg sethandler spec for i i count i lg debug test speculative i i go spec flush wait for the go routine to finish done below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable count used in defer or goroutine at line leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 1
|
1,844
| 4,647,114,224
|
IssuesEvent
|
2016-10-01 09:09:48
|
AllenFang/react-bootstrap-table
|
https://api.github.com/repos/AllenFang/react-bootstrap-table
|
closed
|
caretRender needs more than just direction
|
inprocess
|
It would be really nice to add row and column parameters besides direction to caretRender, similar to the signature that sortFunc has.
I need to initially set the caret direction based on the last sort that user did to their table.
So for example if the list was sorted by name field with desc direction last time user sorted the list, I need to preserve that when they come back.
Cheers
|
1.0
|
caretRender needs more than just direction - It would be really nice to add row and column parameters besides direction to caretRender, similar to the signature that sortFunc has.
I need to initially set the caret direction based on the last sort that user did to their table.
So for example if the list was sorted by name field with desc direction last time user sorted the list, I need to preserve that when they come back.
Cheers
|
non_test
|
caretrender needs more than just direction it would be really nice to add row and column parameters besides direction to caretrender similar to the signature that sortfunc has i need to initially set the caret direction based on the last sort that user did to their table so for example if the list was sorted by name field with desc direction last time user sorted the list i need to preserve that when they come back cheers
| 0
|
247,365
| 20,974,817,734
|
IssuesEvent
|
2022-03-28 14:24:53
|
gakonst/foundry
|
https://api.github.com/repos/gakonst/foundry
|
closed
|
forge test --debug panics with "no ABI for contract"
|
T-bug Cmd-forge-test C-forge P-high Cmd-forge-run
|
### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
forge 0.2.0 (865bc46 2022-03-28T00:08:15.310912+00:00)
### What command(s) is the bug in?
forge test --debug
### Operating System
macOS (M1)
### Describe the bug
running `forge test --debug` seems to panic with "no ABI for contract" for many tests i try it on
to reproduce, run `forge test --debug testWithdrawZero` in solmate
```
λ forge test --debug testWithdrawZero
[⠔] Compiling...
No files changed, compilation skipped
[⠒] Compiling...
[⠊] Compiling 12 files with 0.8.10
Compiler run successful (with warnings)
The application panicked (crashed).
Message: no ABI for contract
Location: cli/src/cmd/forge/run.rs:96
Backtrace omitted.
Run with RUST_BACKTRACE=1 environment variable to display it.
Run with RUST_BACKTRACE=full to include source snippets.
zsh: abort forge test --debug testWithdrawZero
```
|
1.0
|
forge test --debug panics with "no ABI for contract" - ### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
forge 0.2.0 (865bc46 2022-03-28T00:08:15.310912+00:00)
### What command(s) is the bug in?
forge test --debug
### Operating System
macOS (M1)
### Describe the bug
running `forge test --debug` seems to panic with "no ABI for contract" for many tests i try it on
to reproduce, run `forge test --debug testWithdrawZero` in solmate
```
λ forge test --debug testWithdrawZero
[⠔] Compiling...
No files changed, compilation skipped
[⠒] Compiling...
[⠊] Compiling 12 files with 0.8.10
Compiler run successful (with warnings)
The application panicked (crashed).
Message: no ABI for contract
Location: cli/src/cmd/forge/run.rs:96
Backtrace omitted.
Run with RUST_BACKTRACE=1 environment variable to display it.
Run with RUST_BACKTRACE=full to include source snippets.
zsh: abort forge test --debug testWithdrawZero
```
|
test
|
forge test debug panics with no abi for contract component forge have you ensured that all of these are up to date foundry foundryup what version of foundry are you on forge what command s is the bug in forge test debug operating system macos describe the bug running forge test debug seems to panic with no abi for contract for many tests i try it on to reproduce run forge test debug testwithdrawzero in solmate λ forge test debug testwithdrawzero compiling no files changed compilation skipped compiling compiling files with compiler run successful with warnings the application panicked crashed message no abi for contract location cli src cmd forge run rs backtrace omitted run with rust backtrace environment variable to display it run with rust backtrace full to include source snippets zsh abort forge test debug testwithdrawzero
| 1
|
289,540
| 24,996,999,897
|
IssuesEvent
|
2022-11-03 02:01:31
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
There is an extra action 'Search From Here' for one service
|
:heavy_check_mark: merged 🧪 testing :beetle: regression 🌳 new-tree
|
**Storage Explorer Version**: 1.26.0-dev
**Build Number**: 20221006.3
**Branch**: feature branch
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.6 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: AD-hoc testing
**Regression From**: Previous release (1.26.0)
## Steps to Reproduce ##
1. Expand one storage account -> Blob Containers.
2. Right click one blob container -> Observe its context menu and Actions panel.
3. Check there is no action 'Search From Here'.
## Expected Experience ##
There is no action 'Search From Here'.

## Actual Experience ##
There is an extra action 'Search From Here'.

|
1.0
|
There is an extra action 'Search From Here' for one service - **Storage Explorer Version**: 1.26.0-dev
**Build Number**: 20221006.3
**Branch**: feature branch
**Platform/OS**: Windows 10/Linux Ubuntu 22.04/MacOS Monterey 12.6 (Apple M1 Pro)
**Architecture**: ia32/x64
**How Found**: AD-hoc testing
**Regression From**: Previous release (1.26.0)
## Steps to Reproduce ##
1. Expand one storage account -> Blob Containers.
2. Right click one blob container -> Observe its context menu and Actions panel.
3. Check there is no action 'Search From Here'.
## Expected Experience ##
There is no action 'Search From Here'.

## Actual Experience ##
There is an extra action 'Search From Here'.

|
test
|
there is an extra action search from here for one service storage explorer version dev build number branch feature branch platform os windows linux ubuntu macos monterey apple pro architecture how found ad hoc testing regression from previous release steps to reproduce expand one storage account blob containers right click one blob container observe its context menu and actions panel check there is no action search from here expected experience there is no action search from here actual experience there is an extra action search from here
| 1
|
1,634
| 3,296,163,627
|
IssuesEvent
|
2015-11-01 16:38:20
|
eslint/eslint
|
https://api.github.com/repos/eslint/eslint
|
closed
|
Add node 5 to build
|
accepted enhancement infrastructure
|
Add node 5.x version to travis build.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/27873308-add-node-5-to-build?utm_campaign=plugin&utm_content=tracker%2F282608&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F282608&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
Add node 5 to build - Add node 5.x version to travis build.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/27873308-add-node-5-to-build?utm_campaign=plugin&utm_content=tracker%2F282608&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F282608&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_test
|
add node to build add node x version to travis build want to back this issue we accept bounties via
| 0
|
217,972
| 16,892,197,876
|
IssuesEvent
|
2021-06-23 10:36:34
|
ita-social-projects/GreenCityTests
|
https://api.github.com/repos/ita-social-projects/GreenCityTests
|
closed
|
Fix maximizeWindow
|
Test_Broken
|
unknown error: failed to change window state to 'normal', current state is 'maximized'
(Session info: chrome=87.0.4280.88)
Build info: version: '3.141.59', revision: 'e82be7d358', time: '2018-11-14T08:17:03'
System info: host: 'COMPUTER', ip: '192.168.0.103', os.name: 'Windows 10', os.arch: 'amd64', os.version: '10.0', java.version: '1.8.0_271'
Driver info: org.openqa.selenium.remote.RemoteWebDriver
Capabilities {acceptInsecureCerts: false, browserName: chrome, browserVersion: 87.0.4280.88, chrome: {chromedriverVersion: 87.0.4280.88 (89e2380a3e36c..., userDataDir: C:\Users\Mariya\AppData\Loc...}, goog:chromeOptions: {debuggerAddress: localhost:51790}, javascriptEnabled: true, networkConnectionEnabled: false, pageLoadStrategy: normal, platform: WINDOWS, platformName: WINDOWS, proxy: Proxy(), setWindowRect: true, strictFileInteractability: false, timeouts: {implicit: 0, pageLoad: 300000, script: 30000}, unhandledPromptBehavior: dismiss and notify, webauthn:virtualAuthenticators: true, webdriver.remote.sessionid: bbdd1bdf5ae652ba34c403a8888...}
Session ID: bbdd1bdf5ae652ba34c403a888852f10
|
1.0
|
Fix maximizeWindow - unknown error: failed to change window state to 'normal', current state is 'maximized'
(Session info: chrome=87.0.4280.88)
Build info: version: '3.141.59', revision: 'e82be7d358', time: '2018-11-14T08:17:03'
System info: host: 'COMPUTER', ip: '192.168.0.103', os.name: 'Windows 10', os.arch: 'amd64', os.version: '10.0', java.version: '1.8.0_271'
Driver info: org.openqa.selenium.remote.RemoteWebDriver
Capabilities {acceptInsecureCerts: false, browserName: chrome, browserVersion: 87.0.4280.88, chrome: {chromedriverVersion: 87.0.4280.88 (89e2380a3e36c..., userDataDir: C:\Users\Mariya\AppData\Loc...}, goog:chromeOptions: {debuggerAddress: localhost:51790}, javascriptEnabled: true, networkConnectionEnabled: false, pageLoadStrategy: normal, platform: WINDOWS, platformName: WINDOWS, proxy: Proxy(), setWindowRect: true, strictFileInteractability: false, timeouts: {implicit: 0, pageLoad: 300000, script: 30000}, unhandledPromptBehavior: dismiss and notify, webauthn:virtualAuthenticators: true, webdriver.remote.sessionid: bbdd1bdf5ae652ba34c403a8888...}
Session ID: bbdd1bdf5ae652ba34c403a888852f10
|
test
|
fix maximizewindow unknown error failed to change window state to normal current state is maximized session info chrome build info version revision time system info host computer ip os name windows os arch os version java version driver info org openqa selenium remote remotewebdriver capabilities acceptinsecurecerts false browsername chrome browserversion chrome chromedriverversion userdatadir c users mariya appdata loc goog chromeoptions debuggeraddress localhost javascriptenabled true networkconnectionenabled false pageloadstrategy normal platform windows platformname windows proxy proxy setwindowrect true strictfileinteractability false timeouts implicit pageload script unhandledpromptbehavior dismiss and notify webauthn virtualauthenticators true webdriver remote sessionid session id
| 1
|
586,469
| 17,578,606,317
|
IssuesEvent
|
2021-08-16 02:09:40
|
SNDST00M/material-dynmap
|
https://api.github.com/repos/SNDST00M/material-dynmap
|
closed
|
Script elements injected with the same ID
|
type-bug context-script priority-medium scope-stability status-tracking
|
## 🧱 Bug Report
- [x] I'm using the latest version of `material-dynmap` available.
- [x] I searched [existing issues][material-dynmap-issues], open & closed. Yes, my issue is new.
#### Describe the bug
<!-- A clear and concise description of what the bug is. -->
The main app and site entrypoint script tags are both using the `material-dynmap-script` ID.
Lets rename the main app ID to `material-dynmap-app` to prevent undefined behaviour.
#### Steps to reproduce
1. Install the userscript.
2. Launch the map [this Brazil survival server][brazil-survival-server]
3. Open the console and type:
document.querySelectorAll('[id="material-dynmap-script"]').length
4. The result is `2`.
#### Screenshots
<!-- Add screenshots to help explaining. If a screenshot is not provided, the issue will be closed. -->
<img alt="image" src="https://user-images.githubusercontent.com/82655227/129267637-8ae179da-cdfb-44c7-b1aa-d920095cae9b.png" />
#### Possible solution
<!-- Only if you have suggestions on a fix for the bug. If not, please remove the section. -->
Simply switch the main app script's ID to the `-app` suffix.
#### Additional context
<!-- Add any other context about the problem here (, , ...). -->
- Browser version: Edge 92
- Extensions: Tampermonkey, Refined Github, Notifier for Github, PiP Extension by Google, uBlock Origin
- OS version: Windows 10 Home 20H2
<!-- Checklist -->
[material-dynmap-issues]: https://github.com/SNDST00M/material-dynmap/issues?q=is%3Aissue+is%3Aopen+sort%3Aupdated-desc
[brazil-survival-server]: http://170.81.41.79.ipv4.reishosting.com.br:26033
|
1.0
|
Script elements injected with the same ID - ## 🧱 Bug Report
- [x] I'm using the latest version of `material-dynmap` available.
- [x] I searched [existing issues][material-dynmap-issues], open & closed. Yes, my issue is new.
#### Describe the bug
<!-- A clear and concise description of what the bug is. -->
The main app and site entrypoint script tags are both using the `material-dynmap-script` ID.
Lets rename the main app ID to `material-dynmap-app` to prevent undefined behaviour.
#### Steps to reproduce
1. Install the userscript.
2. Launch the map [this Brazil survival server][brazil-survival-server]
3. Open the console and type:
document.querySelectorAll('[id="material-dynmap-script"]').length
4. The result is `2`.
#### Screenshots
<!-- Add screenshots to help explaining. If a screenshot is not provided, the issue will be closed. -->
<img alt="image" src="https://user-images.githubusercontent.com/82655227/129267637-8ae179da-cdfb-44c7-b1aa-d920095cae9b.png" />
#### Possible solution
<!-- Only if you have suggestions on a fix for the bug. If not, please remove the section. -->
Simply switch the main app script's ID to the `-app` suffix.
#### Additional context
<!-- Add any other context about the problem here (, , ...). -->
- Browser version: Edge 92
- Extensions: Tampermonkey, Refined Github, Notifier for Github, PiP Extension by Google, uBlock Origin
- OS version: Windows 10 Home 20H2
<!-- Checklist -->
[material-dynmap-issues]: https://github.com/SNDST00M/material-dynmap/issues?q=is%3Aissue+is%3Aopen+sort%3Aupdated-desc
[brazil-survival-server]: http://170.81.41.79.ipv4.reishosting.com.br:26033
|
non_test
|
script elements injected with the same id 🧱 bug report i m using the latest version of material dynmap available i searched open closed yes my issue is new describe the bug the main app and site entrypoint script tags are both using the material dynmap script id lets rename the main app id to material dynmap app to prevent undefined behaviour steps to reproduce install the userscript launch the map open the console and type document queryselectorall length the result is screenshots possible solution simply switch the main app script s id to the app suffix additional context browser version edge extensions tampermonkey refined github notifier for github pip extension by google ublock origin os version windows home
| 0
|
153,030
| 12,132,850,042
|
IssuesEvent
|
2020-04-23 08:02:21
|
WoWManiaUK/Blackwing-Lair
|
https://api.github.com/repos/WoWManiaUK/Blackwing-Lair
|
closed
|
[Quest/item] Natural Oil (not dropping) - Ashenvale
|
Confirmed By Tester Fixed Confirmed Fixed in Dev Starting Zone
|
**Links:**
item http://cata.cavernoftime.com/item=45071
quest http://cata.cavernoftime.com/quest=13651
**What is happening:**
- quest item not dropping
Could be due to low spawns (should be 81) #4360
**What should happen:**
_Gorka at Warsong Labor Camp wants you to collect 5 Natural Oil from the Rotting Slimes in Ashenvale._
|
1.0
|
[Quest/item] Natural Oil (not dropping) - Ashenvale - **Links:**
item http://cata.cavernoftime.com/item=45071
quest http://cata.cavernoftime.com/quest=13651
**What is happening:**
- quest item not dropping
Could be due to low spawns (should be 81) #4360
**What should happen:**
_Gorka at Warsong Labor Camp wants you to collect 5 Natural Oil from the Rotting Slimes in Ashenvale._
|
test
|
natural oil not dropping ashenvale links item quest what is happening quest item not dropping could be due to low spawns should be what should happen gorka at warsong labor camp wants you to collect natural oil from the rotting slimes in ashenvale
| 1
|
52,339
| 6,228,022,366
|
IssuesEvent
|
2017-07-10 22:07:02
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Health check instance stuck in "Initilaizing" state
|
area/container-refactoring kind/bug status/to-test
|
Rancher Versions: Build from wagyu
Steps to reproduce the problem:
Add 2 hosts to "Default" environment.
Health check instance stuck in "Initializing" state on 1 of the host and there is no health check instance that gets created on the other host
```
mysql> select id,name,state,created,removed,health_state from instance where name like "health%";
+----+---------------------------+---------+---------------------+---------+--------------+
| id | name | state | created | removed | health_state |
+----+---------------------------+---------+---------------------+---------+--------------+
| 12 | healthcheck-healthcheck-1 | running | 2017-04-03 22:42:23 | NULL | initializing |
+----+---------------------------+---------+---------------------+---------+--------------+
1 row in set (0.00 sec)
mysql>
```
Service activation fails with error - ```Failed to acquire lock [DEPLOYMENT.UNIT.5])```
|
1.0
|
Health check instance stuck in "Initilaizing" state - Rancher Versions: Build from wagyu
Steps to reproduce the problem:
Add 2 hosts to "Default" environment.
Health check instance stuck in "Initializing" state on 1 of the host and there is no health check instance that gets created on the other host
```
mysql> select id,name,state,created,removed,health_state from instance where name like "health%";
+----+---------------------------+---------+---------------------+---------+--------------+
| id | name | state | created | removed | health_state |
+----+---------------------------+---------+---------------------+---------+--------------+
| 12 | healthcheck-healthcheck-1 | running | 2017-04-03 22:42:23 | NULL | initializing |
+----+---------------------------+---------+---------------------+---------+--------------+
1 row in set (0.00 sec)
mysql>
```
Service activation fails with error - ```Failed to acquire lock [DEPLOYMENT.UNIT.5])```
|
test
|
health check instance stuck in initilaizing state rancher versions build from wagyu steps to reproduce the problem add hosts to default environment health check instance stuck in initializing state on of the host and there is no health check instance that gets created on the other host mysql select id name state created removed health state from instance where name like health id name state created removed health state healthcheck healthcheck running null initializing row in set sec mysql service activation fails with error failed to acquire lock
| 1
|
155,003
| 19,765,642,132
|
IssuesEvent
|
2022-01-17 01:38:18
|
tuanducdesign/infomation-covid19
|
https://api.github.com/repos/tuanducdesign/infomation-covid19
|
closed
|
CVE-2021-23382 (Medium) detected in postcss-7.0.34.tgz, postcss-7.0.21.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.34.tgz</b>, <b>postcss-7.0.21.tgz</b></p></summary>
<p>
<details><summary><b>postcss-7.0.34.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.34.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.34.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- postcss-safe-parser-4.0.1.tgz
- :x: **postcss-7.0.34.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.21.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/resolve-url-loader/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- resolve-url-loader-3.1.1.tgz
- :x: **postcss-7.0.21.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/infomation-covid19/commit/1a8934856afd37b89c3529b13ed31bfd40c5eb72">1a8934856afd37b89c3529b13ed31bfd40c5eb72</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23382 (Medium) detected in postcss-7.0.34.tgz, postcss-7.0.21.tgz - autoclosed - ## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.34.tgz</b>, <b>postcss-7.0.21.tgz</b></p></summary>
<p>
<details><summary><b>postcss-7.0.34.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.34.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.34.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- postcss-safe-parser-4.0.1.tgz
- :x: **postcss-7.0.34.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.21.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.21.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/resolve-url-loader/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- resolve-url-loader-3.1.1.tgz
- :x: **postcss-7.0.21.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/infomation-covid19/commit/1a8934856afd37b89c3529b13ed31bfd40c5eb72">1a8934856afd37b89c3529b13ed31bfd40c5eb72</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_test
|
cve medium detected in postcss tgz postcss tgz autoclosed cve medium severity vulnerability vulnerable libraries postcss tgz postcss tgz postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules postcss package json dependency hierarchy react scripts tgz root library postcss safe parser tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file package json path to vulnerable library node modules resolve url loader node modules postcss package json dependency hierarchy react scripts tgz root library resolve url loader tgz x postcss tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss step up your open source security game with whitesource
| 0
|
202,522
| 15,286,842,784
|
IssuesEvent
|
2021-02-23 15:08:07
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: schemachange/during/kv failed
|
C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker
|
[(roachtest).schemachange/during/kv failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661563&tab=buildLog) on [release-20.1@90f78268f3b5b08ba838ac3ad164821d2f5a5362](https://github.com/cockroachdb/cockroach/commits/90f78268f3b5b08ba838ac3ad164821d2f5a5362):
```
The test failed on branch=release-20.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/schemachange/during/kv/run_1
cluster.go:2198,schemachange.go:48,test_runner.go:749: output in run_092457.520_n1_workload_init_kv: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661563-1612941348-108-n5cpu4:1 -- ./workload init kv --drop --db=test returned: exit status 20
(1) attached stack trace
| main.(*cluster).RunE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2276
| main.(*cluster).Run
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2196
| main.registerSchemaChangeDuringKV.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/schemachange.go:48
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) 2 safe details enclosed
Wraps: (3) output in run_092457.520_n1_workload_init_kv
Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661563-1612941348-108-n5cpu4:1 -- ./workload init kv --drop --db=test returned
| stderr:
| ./workload: error while loading shared libraries: libncurses.so.6: cannot open shared object file: No such file or directory
| Error: COMMAND_PROBLEM: exit status 127
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 1. Command with error:
| | ```
| | ./workload init kv --drop --db=test
| | ```
| Wraps: (3) exit status 127
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (5) exit status 20
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/schemachange/during/kv](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661563&tab=artifacts#/schemachange/during/kv)
Related:
- #60246 roachtest: schemachange/during/kv failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-60149](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-60149) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aschemachange%2Fduring%2Fkv.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
2.0
|
roachtest: schemachange/during/kv failed - [(roachtest).schemachange/during/kv failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661563&tab=buildLog) on [release-20.1@90f78268f3b5b08ba838ac3ad164821d2f5a5362](https://github.com/cockroachdb/cockroach/commits/90f78268f3b5b08ba838ac3ad164821d2f5a5362):
```
The test failed on branch=release-20.1, cloud=gce:
test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/schemachange/during/kv/run_1
cluster.go:2198,schemachange.go:48,test_runner.go:749: output in run_092457.520_n1_workload_init_kv: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661563-1612941348-108-n5cpu4:1 -- ./workload init kv --drop --db=test returned: exit status 20
(1) attached stack trace
| main.(*cluster).RunE
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2276
| main.(*cluster).Run
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2196
| main.registerSchemaChangeDuringKV.func1
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/schemachange.go:48
| main.(*testRunner).runTest.func2
| /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749
| runtime.goexit
| /usr/local/go/src/runtime/asm_amd64.s:1374
Wraps: (2) 2 safe details enclosed
Wraps: (3) output in run_092457.520_n1_workload_init_kv
Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2661563-1612941348-108-n5cpu4:1 -- ./workload init kv --drop --db=test returned
| stderr:
| ./workload: error while loading shared libraries: libncurses.so.6: cannot open shared object file: No such file or directory
| Error: COMMAND_PROBLEM: exit status 127
| (1) COMMAND_PROBLEM
| Wraps: (2) Node 1. Command with error:
| | ```
| | ./workload init kv --drop --db=test
| | ```
| Wraps: (3) exit status 127
| Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError
|
| stdout:
Wraps: (5) exit status 20
Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError
```
<details><summary>More</summary><p>
Artifacts: [/schemachange/during/kv](https://teamcity.cockroachdb.com/viewLog.html?buildId=2661563&tab=artifacts#/schemachange/during/kv)
Related:
- #60246 roachtest: schemachange/during/kv failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-60149](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-60149) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aschemachange%2Fduring%2Fkv.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
|
test
|
roachtest schemachange during kv failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts schemachange during kv run cluster go schemachange go test runner go output in run workload init kv home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload init kv drop db test returned exit status attached stack trace main cluster rune home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main cluster run home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main registerschemachangeduringkv home agent work go src github com cockroachdb cockroach pkg cmd roachtest schemachange go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps output in run workload init kv wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload init kv drop db test returned stderr workload error while loading shared libraries libncurses so cannot open shared object file no such file or directory error command problem exit status command problem wraps node command with error workload init kv drop db test wraps exit status error types errors cmd hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror more artifacts related roachtest schemachange during kv failed powered by
| 1
|
169
| 2,550,718,125
|
IssuesEvent
|
2015-02-01 21:13:46
|
Starcounter/Starcounter
|
https://api.github.com/repos/Starcounter/Starcounter
|
closed
|
Confusions with parameters in project Develop-Nightly-Debug
|
Infrastructure question
|
When I look to the parameters of the project Develop-Nightly-Debug, I am confused that the configuration parameter is set to `Release`, while I expect to see `Debug`. I also see parameter SC_RUN_DEBUG_BUILD, which sounds to have the same role as Configuration.
Can it be clarified?
What is the role of parameter Configuration for all the project in general? Does it have any meaning?
|
1.0
|
Confusions with parameters in project Develop-Nightly-Debug - When I look to the parameters of the project Develop-Nightly-Debug, I am confused that the configuration parameter is set to `Release`, while I expect to see `Debug`. I also see parameter SC_RUN_DEBUG_BUILD, which sounds to have the same role as Configuration.
Can it be clarified?
What is the role of parameter Configuration for all the project in general? Does it have any meaning?
|
non_test
|
confusions with parameters in project develop nightly debug when i look to the parameters of the project develop nightly debug i am confused that the configuration parameter is set to release while i expect to see debug i also see parameter sc run debug build which sounds to have the same role as configuration can it be clarified what is the role of parameter configuration for all the project in general does it have any meaning
| 0
|
118,738
| 10,001,493,786
|
IssuesEvent
|
2019-07-12 15:45:36
|
kyma-project/kyma
|
https://api.github.com/repos/kyma-project/kyma
|
closed
|
api-controller tests can be run in parallel with others
|
quality/testability
|
**Description**
It should be possible to run api-controller tests in parallel with other Kyma tests. We need to verify if it is possible now and change the test if necessary. Then its TestDefinition should be modified to enable concurrency.
**Reasons**
One way to make the Kyma test suite faster is to run tests in parallel.
**Acceptance Criteria**
- [ ] concurrency is enabled in api-controller test and it is stable on CI
|
1.0
|
api-controller tests can be run in parallel with others - **Description**
It should be possible to run api-controller tests in parallel with other Kyma tests. We need to verify if it is possible now and change the test if necessary. Then its TestDefinition should be modified to enable concurrency.
**Reasons**
One way to make the Kyma test suite faster is to run tests in parallel.
**Acceptance Criteria**
- [ ] concurrency is enabled in api-controller test and it is stable on CI
|
test
|
api controller tests can be run in parallel with others description it should be possible to run api controller tests in parallel with other kyma tests we need to verify if it is possible now and change the test if necessary then its testdefinition should be modified to enable concurrency reasons one way to make the kyma test suite faster is to run tests in parallel acceptance criteria concurrency is enabled in api controller test and it is stable on ci
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.