Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,441 | 8,638,204,691 | IssuesEvent | 2018-11-23 14:00:05 | invertase/react-native-firebase | https://api.github.com/repos/invertase/react-native-firebase | closed | Can`t receive credential-already-in-use from linkWithCredential on Android | await-maintainer-feedback await-react-native-pr ios π await-user-feedback π€ android | <!---
Hello there you awesome person;
Please note that the issue list of this repo is exclusively for bug reports;
1) For feature requests please visit our [Feature Request Board](https://boards.invertase.io/react-native-firebase).
2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase
3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed.
-->
<!-- NOTE: You can change any of the `[ ]` to `[x]` to mark an option(s) as selected -->
<!-- PLEASE DO NOT REMOVE ANY SECTIONS FROM THIS ISSUE TEMPLATE -->
<!-- Leave them as they are even if they're irrelevant to your issue -->
## Issue
I have code which should authtentificate user if it already exists or link to anonymous
```
export const authorizeSkippedUser = function* authorizeSkippedUser(
user,
credentials
) {
try {
let result = yield call(
[user, user.linkAndRetrieveDataWithCredential],
credentials
);
let userProfile = yield call(updateAndRetrieveUserProfile, user.uid);
yield put(AuthActions.authSuccess(userProfile, true));
} catch (ex) {
if (
ex.code &&
ex.code === 'auth/credential-already-in-use' &&
isSocialProvider(credentials.providerId)
) {
yield put(AuthActions.authSignOut());
yield call(
[firebaseAuth, firebaseAuth.signInAndRetrieveDataWithCredential],
credentials
);
} else if (ex.code && ex.code === 'auth/email-already-in-use') {
ex.message = yield* getProvideMessage(ex.email);
throw ex;
} else {
throw ex;
}
}
};
```
it work fine before but now i can`t receive error auth/credential-already-in-use and I don`t know why, I think maybe after somewhere update
<!-- Please describe your issue here --^ and provide as much detail as you can. -->
<!-- Include code snippets that show your usages of the library in the context of your project. -->
<!-- Snippets that also show how and where the library is imported in JS are useful to debug issues relating to importing or methods not found issues -->
---
## Project Files
<!-- Provide the contents of key project files which will help to debug -->
<!-- For Example: -->
<!-- - iOS: `Podfile` contents. -->
<!-- - Android: `android/build.gradle` contents. -->
<!-- - Android: `android/app/build.gradle` contents. -->
<!-- - Android: `AndroidManifest.xml` contents. -->
<!-- ADD THE CONTENTS OF THE FILES IN THE PROVIDED CODE BLOCKS BELOW -->
### Android
#### `android/build.gradle`:
```groovy
// Top-level build file where you can add configuration options common to all sub-projects/modules.
buildscript {
ext {
buildToolsVersion = "27.0.3"
minSdkVersion = 16
compileSdkVersion = 27
targetSdkVersion = 26
supportLibVersion = "27.1.1"
}
repositories {
jcenter()
maven { url 'https://maven.fabric.io/public' }
maven {
url 'https://maven.google.com/'
name 'Google'
}
google()
}
dependencies {
classpath 'com.android.tools.build:gradle:3.1.4'
classpath 'com.google.gms:google-services:3.2.1'
classpath 'io.fabric.tools:gradle:1.25.4'
classpath 'com.google.firebase:firebase-plugins:1.1.5'
// NOTE: Do not place your application dependencies here; they belong
// in the individual module build.gradle files
}
}
allprojects {
repositories {
google()
mavenLocal()
jcenter()
maven {
// All of React Native (JS, Obj-C sources, Android binaries) is installed from npm
url "$rootDir/../node_modules/react-native/android"
}
maven {
url 'https://maven.google.com/'
name 'Google'
}
configurations.all {
resolutionStrategy {
force 'com.facebook.android:facebook-android-sdk:4.22.1'
eachDependency { DependencyResolveDetails details ->
if (details.requested.group == 'com.facebook.react' && details.requested.name == 'react-native') {
def file = new File("$rootDir/../node_modules/react-native/package.json")
def version = new groovy.json.JsonSlurper().parseText(file.text).version
details.useVersion version
}
}
}
}
}
}
task wrapper(type: Wrapper) {
gradleVersion = '4.4'
distributionUrl = distributionUrl.replace("bin", "all")
}
```
#### `android/app/build.gradle`:
```groovy
dependencies {
compile project(':react-native-config')
implementation project(':react-native-fabric')
implementation(project(':react-native-firebase')) {
transitive = false
}
implementation project(':react-native-gesture-handler')
implementation project(':appcenter-crashes')
implementation project(':appcenter-analytics')
implementation project(':appcenter')
implementation project(':react-native-android-location-services-dialog-box')
implementation project(':react-native-background-timer')
implementation project(':react-native-code-push')
implementation(project(':react-native-device-info')) {
exclude group: "com.google.android.gms" // very important
}
implementation project(':react-native-fbsdk')
implementation project(':react-native-fetch-blob')
implementation(project(':react-native-google-signin')) {
exclude group: "com.google.android.gms" // very important
}
implementation project(':react-native-keep-awake')
implementation project(':react-native-linear-gradient')
implementation project(':react-native-shake-event')
implementation project(':react-native-splash-screen')
implementation project(':react-native-svg')
implementation project(':react-native-vector-icons')
implementation project(':react-native-view-shot')
implementation fileTree(dir: "libs", include: ["*.jar"])
//noinspection GradleCompatible
implementation "com.android.support:appcompat-v7:${rootProject.ext.supportLibVersion}"
implementation 'com.android.support:multidex:1.0.1'
//noinspection GradleDynamicVersion
implementation 'com.facebook.react:react-native:+' // From node_modules
implementation 'com.twitter.sdk.android:twitter:3.0.0'
implementation 'com.google.code.gson:gson:2.8.0'
implementation('com.google.firebase:firebase-core:16.0.1') {
force = true
}
implementation('com.google.firebase:firebase-messaging:17.1.0') {
force = true
}
implementation "com.google.firebase:firebase-ads:15.0.1"
implementation "com.google.firebase:firebase-storage:16.0.1"
implementation "com.google.firebase:firebase-perf:16.1.0"
implementation "com.google.firebase:firebase-database:16.0.1"
implementation "com.google.firebase:firebase-auth:16.0.3"
implementation 'me.leolin:ShortcutBadger:1.1.21@aar'
implementation("com.google.android.gms:play-services-base:15.0.1") {
force = true
}
implementation("com.google.android.gms:play-services-base:15.0.1") {
force = true
}
implementation('com.google.android.gms:play-services-ads:15.0.1') {
force = true
}
implementation('com.google.android.gms:play-services-auth:15.0.1') {
force = true
}
implementation('com.google.android.gms:play-services-gcm:15.0.1') {
force = true
}
implementation('com.crashlytics.sdk.android:crashlytics:2.9.3@aar') {
transitive = true
}
}
```
---
## Environment
<!-- change `[ ]` to `[x]` to select an option(s) -->
- **Platform that you're experiencing the issue on**:
- [ ] iOS
- [x] Android
- [ ] **iOS** but have not tested behavior on Android
- [ ] **Android** but have not tested behavior on iOS
- [ ] Both
- **Operating System:**
- [x] MacOS, version: `10.12.6`
- [ ] Windows, version: `N/A`
- [ ] Other, please specify: `N/A`
- **Build Tools:**
- `Android Studio 3.2`
- **`React Native` version:**
- `0.57.1`
- **`React Native Firebase` library version:**
- `4.3.8`
- **`Firebase` module(s) you're using that has the issue:**
- [] **N/A**
- [x] Authentication
- [ ] Analytics
- [ ] Cloud **Firestore**
- [ ] Cloud **Messaging** (FCM)
- [ ] Crashlytics
- [ ] Dynamic **Links**
- [ ] **Functions** Callable
- [ ] Invites
- [ ] Instance ID
- [ ] Notifications
- [ ] Performance Monitoring
- [ ] Realtime **Database**
- [ ] Remote **Config**
- [ ] Storage
- **Are you using `TypeScript`?**
- [x] No
- [ ] Yes, version: `N/A`
- **Are you using Expo, e.g. `ExpoKit`?**
- [x] No
- [ ] Yes, I've _not_ ejected
- [ ] Yes, but I **have** ejected to `ExpoKit`
- [ ] Yes, but I **have** ejected to vanilla React Native
- Expo version: `N/A`
<!-- Thanks for reading this far down β€οΈ -->
<!-- High quality, detailed issues are much easier and quicker to triage for maintainers -->
<!-- For bonus points, if you put a π₯ (:fire:) emojii at the start of the issue title we'll know -->
<!-- that you took the time to fill this out correctly, or, at least read this far -->
---
Think `react-native-firebase` is great? Please consider supporting the project with any of the below:
- π Donate via [Open Collective](https://opencollective.com/react-native-firebase/donate)
- π Follow [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter
- π Star this repo on GitHub βοΈ
- π Contribute; see our [contributing guide](./../../CONTRIBUTING.md)
| True | Can`t receive credential-already-in-use from linkWithCredential on Android - <!---
Hello there you awesome person;
Please note that the issue list of this repo is exclusively for bug reports;
1) For feature requests please visit our [Feature Request Board](https://boards.invertase.io/react-native-firebase).
2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase
3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed.
-->
<!-- NOTE: You can change any of the `[ ]` to `[x]` to mark an option(s) as selected -->
<!-- PLEASE DO NOT REMOVE ANY SECTIONS FROM THIS ISSUE TEMPLATE -->
<!-- Leave them as they are even if they're irrelevant to your issue -->
## Issue
I have code which should authtentificate user if it already exists or link to anonymous
```
export const authorizeSkippedUser = function* authorizeSkippedUser(
user,
credentials
) {
try {
let result = yield call(
[user, user.linkAndRetrieveDataWithCredential],
credentials
);
let userProfile = yield call(updateAndRetrieveUserProfile, user.uid);
yield put(AuthActions.authSuccess(userProfile, true));
} catch (ex) {
if (
ex.code &&
ex.code === 'auth/credential-already-in-use' &&
isSocialProvider(credentials.providerId)
) {
yield put(AuthActions.authSignOut());
yield call(
[firebaseAuth, firebaseAuth.signInAndRetrieveDataWithCredential],
credentials
);
} else if (ex.code && ex.code === 'auth/email-already-in-use') {
ex.message = yield* getProvideMessage(ex.email);
throw ex;
} else {
throw ex;
}
}
};
```
it work fine before but now i can`t receive error auth/credential-already-in-use and I don`t know why, I think maybe after somewhere update
<!-- Please describe your issue here --^ and provide as much detail as you can. -->
<!-- Include code snippets that show your usages of the library in the context of your project. -->
<!-- Snippets that also show how and where the library is imported in JS are useful to debug issues relating to importing or methods not found issues -->
---
## Project Files
<!-- Provide the contents of key project files which will help to debug -->
<!-- For Example: -->
<!-- - iOS: `Podfile` contents. -->
<!-- - Android: `android/build.gradle` contents. -->
<!-- - Android: `android/app/build.gradle` contents. -->
<!-- - Android: `AndroidManifest.xml` contents. -->
<!-- ADD THE CONTENTS OF THE FILES IN THE PROVIDED CODE BLOCKS BELOW -->
### Android
#### `android/build.gradle`:
```groovy
// Top-level build file where you can add configuration options common to all sub-projects/modules.
buildscript {
ext {
buildToolsVersion = "27.0.3"
minSdkVersion = 16
compileSdkVersion = 27
targetSdkVersion = 26
supportLibVersion = "27.1.1"
}
repositories {
jcenter()
maven { url 'https://maven.fabric.io/public' }
maven {
url 'https://maven.google.com/'
name 'Google'
}
google()
}
dependencies {
classpath 'com.android.tools.build:gradle:3.1.4'
classpath 'com.google.gms:google-services:3.2.1'
classpath 'io.fabric.tools:gradle:1.25.4'
classpath 'com.google.firebase:firebase-plugins:1.1.5'
// NOTE: Do not place your application dependencies here; they belong
// in the individual module build.gradle files
}
}
allprojects {
repositories {
google()
mavenLocal()
jcenter()
maven {
// All of React Native (JS, Obj-C sources, Android binaries) is installed from npm
url "$rootDir/../node_modules/react-native/android"
}
maven {
url 'https://maven.google.com/'
name 'Google'
}
configurations.all {
resolutionStrategy {
force 'com.facebook.android:facebook-android-sdk:4.22.1'
eachDependency { DependencyResolveDetails details ->
if (details.requested.group == 'com.facebook.react' && details.requested.name == 'react-native') {
def file = new File("$rootDir/../node_modules/react-native/package.json")
def version = new groovy.json.JsonSlurper().parseText(file.text).version
details.useVersion version
}
}
}
}
}
}
task wrapper(type: Wrapper) {
gradleVersion = '4.4'
distributionUrl = distributionUrl.replace("bin", "all")
}
```
#### `android/app/build.gradle`:
```groovy
dependencies {
compile project(':react-native-config')
implementation project(':react-native-fabric')
implementation(project(':react-native-firebase')) {
transitive = false
}
implementation project(':react-native-gesture-handler')
implementation project(':appcenter-crashes')
implementation project(':appcenter-analytics')
implementation project(':appcenter')
implementation project(':react-native-android-location-services-dialog-box')
implementation project(':react-native-background-timer')
implementation project(':react-native-code-push')
implementation(project(':react-native-device-info')) {
exclude group: "com.google.android.gms" // very important
}
implementation project(':react-native-fbsdk')
implementation project(':react-native-fetch-blob')
implementation(project(':react-native-google-signin')) {
exclude group: "com.google.android.gms" // very important
}
implementation project(':react-native-keep-awake')
implementation project(':react-native-linear-gradient')
implementation project(':react-native-shake-event')
implementation project(':react-native-splash-screen')
implementation project(':react-native-svg')
implementation project(':react-native-vector-icons')
implementation project(':react-native-view-shot')
implementation fileTree(dir: "libs", include: ["*.jar"])
//noinspection GradleCompatible
implementation "com.android.support:appcompat-v7:${rootProject.ext.supportLibVersion}"
implementation 'com.android.support:multidex:1.0.1'
//noinspection GradleDynamicVersion
implementation 'com.facebook.react:react-native:+' // From node_modules
implementation 'com.twitter.sdk.android:twitter:3.0.0'
implementation 'com.google.code.gson:gson:2.8.0'
implementation('com.google.firebase:firebase-core:16.0.1') {
force = true
}
implementation('com.google.firebase:firebase-messaging:17.1.0') {
force = true
}
implementation "com.google.firebase:firebase-ads:15.0.1"
implementation "com.google.firebase:firebase-storage:16.0.1"
implementation "com.google.firebase:firebase-perf:16.1.0"
implementation "com.google.firebase:firebase-database:16.0.1"
implementation "com.google.firebase:firebase-auth:16.0.3"
implementation 'me.leolin:ShortcutBadger:1.1.21@aar'
implementation("com.google.android.gms:play-services-base:15.0.1") {
force = true
}
implementation("com.google.android.gms:play-services-base:15.0.1") {
force = true
}
implementation('com.google.android.gms:play-services-ads:15.0.1') {
force = true
}
implementation('com.google.android.gms:play-services-auth:15.0.1') {
force = true
}
implementation('com.google.android.gms:play-services-gcm:15.0.1') {
force = true
}
implementation('com.crashlytics.sdk.android:crashlytics:2.9.3@aar') {
transitive = true
}
}
```
---
## Environment
<!-- change `[ ]` to `[x]` to select an option(s) -->
- **Platform that you're experiencing the issue on**:
- [ ] iOS
- [x] Android
- [ ] **iOS** but have not tested behavior on Android
- [ ] **Android** but have not tested behavior on iOS
- [ ] Both
- **Operating System:**
- [x] MacOS, version: `10.12.6`
- [ ] Windows, version: `N/A`
- [ ] Other, please specify: `N/A`
- **Build Tools:**
- `Android Studio 3.2`
- **`React Native` version:**
- `0.57.1`
- **`React Native Firebase` library version:**
- `4.3.8`
- **`Firebase` module(s) you're using that has the issue:**
- [] **N/A**
- [x] Authentication
- [ ] Analytics
- [ ] Cloud **Firestore**
- [ ] Cloud **Messaging** (FCM)
- [ ] Crashlytics
- [ ] Dynamic **Links**
- [ ] **Functions** Callable
- [ ] Invites
- [ ] Instance ID
- [ ] Notifications
- [ ] Performance Monitoring
- [ ] Realtime **Database**
- [ ] Remote **Config**
- [ ] Storage
- **Are you using `TypeScript`?**
- [x] No
- [ ] Yes, version: `N/A`
- **Are you using Expo, e.g. `ExpoKit`?**
- [x] No
- [ ] Yes, I've _not_ ejected
- [ ] Yes, but I **have** ejected to `ExpoKit`
- [ ] Yes, but I **have** ejected to vanilla React Native
- Expo version: `N/A`
<!-- Thanks for reading this far down β€οΈ -->
<!-- High quality, detailed issues are much easier and quicker to triage for maintainers -->
<!-- For bonus points, if you put a π₯ (:fire:) emojii at the start of the issue title we'll know -->
<!-- that you took the time to fill this out correctly, or, at least read this far -->
---
Think `react-native-firebase` is great? Please consider supporting the project with any of the below:
- π Donate via [Open Collective](https://opencollective.com/react-native-firebase/donate)
- π Follow [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter
- π Star this repo on GitHub βοΈ
- π Contribute; see our [contributing guide](./../../CONTRIBUTING.md)
| main | can t receive credential already in use from linkwithcredential on android hello there you awesome person please note that the issue list of this repo is exclusively for bug reports for feature requests please visit our for questions and support please use our discord chat or stack overflow if this is a setup issue then please make sure you ve correctly followed the setup guides most setup issues such as duplicate dex files default app has not been initialized etc are all down to an incorrect setup as the guides haven t been correctly followed issue i have code which should authtentificate user if it already exists or link to anonymous export const authorizeskippeduser function authorizeskippeduser user credentials try let result yield call credentials let userprofile yield call updateandretrieveuserprofile user uid yield put authactions authsuccess userprofile true catch ex if ex code ex code auth credential already in use issocialprovider credentials providerid yield put authactions authsignout yield call credentials else if ex code ex code auth email already in use ex message yield getprovidemessage ex email throw ex else throw ex it work fine before but now i can t receive error auth credential already in use and i don t know why i think maybe after somewhere update project files android android build gradle groovy top level build file where you can add configuration options common to all sub projects modules buildscript ext buildtoolsversion minsdkversion compilesdkversion targetsdkversion supportlibversion repositories jcenter maven url maven url name google google dependencies classpath com android tools build gradle classpath com google gms google services classpath io fabric tools gradle classpath com google firebase firebase plugins note do not place your application dependencies here they belong in the individual module build gradle files allprojects repositories google mavenlocal jcenter maven all of react native js obj c sources android binaries is installed from npm url rootdir node modules react native android maven url name google configurations all resolutionstrategy force com facebook android facebook android sdk eachdependency dependencyresolvedetails details if details requested group com facebook react details requested name react native def file new file rootdir node modules react native package json def version new groovy json jsonslurper parsetext file text version details useversion version task wrapper type wrapper gradleversion distributionurl distributionurl replace bin all android app build gradle groovy dependencies compile project react native config implementation project react native fabric implementation project react native firebase transitive false implementation project react native gesture handler implementation project appcenter crashes implementation project appcenter analytics implementation project appcenter implementation project react native android location services dialog box implementation project react native background timer implementation project react native code push implementation project react native device info exclude group com google android gms very important implementation project react native fbsdk implementation project react native fetch blob implementation project react native google signin exclude group com google android gms very important implementation project react native keep awake implementation project react native linear gradient implementation project react native shake event implementation project react native splash screen implementation project react native svg implementation project react native vector icons implementation project react native view shot implementation filetree dir libs include noinspection gradlecompatible implementation com android support appcompat rootproject ext supportlibversion implementation com android support multidex noinspection gradledynamicversion implementation com facebook react react native from node modules implementation com twitter sdk android twitter implementation com google code gson gson implementation com google firebase firebase core force true implementation com google firebase firebase messaging force true implementation com google firebase firebase ads implementation com google firebase firebase storage implementation com google firebase firebase perf implementation com google firebase firebase database implementation com google firebase firebase auth implementation me leolin shortcutbadger aar implementation com google android gms play services base force true implementation com google android gms play services base force true implementation com google android gms play services ads force true implementation com google android gms play services auth force true implementation com google android gms play services gcm force true implementation com crashlytics sdk android crashlytics aar transitive true environment platform that you re experiencing the issue on ios android ios but have not tested behavior on android android but have not tested behavior on ios both operating system macos version windows version n a other please specify n a build tools android studio react native version react native firebase library version firebase module s you re using that has the issue n a authentication analytics cloud firestore cloud messaging fcm crashlytics dynamic links functions callable invites instance id notifications performance monitoring realtime database remote config storage are you using typescript no yes version n a are you using expo e g expokit no yes i ve not ejected yes but i have ejected to expokit yes but i have ejected to vanilla react native expo version n a think react native firebase is great please consider supporting the project with any of the below π donate via π follow and on twitter π star this repo on github βοΈ π contribute see our contributing md | 1 |
192 | 2,814,796,774 | IssuesEvent | 2015-05-18 22:05:39 | DotNetAnalyzers/StyleCopAnalyzers | https://api.github.com/repos/DotNetAnalyzers/StyleCopAnalyzers | closed | Rule proposal - Readability (Maintainablility) rule - Report assignments in condition expressions | maintainability needs discussion new rule proposal | To carry on the side topic at http://stylecop.codeplex.com/discussions/252502...
StyleCop does not currently complain about either of the following code patterns. Where bb and cc are bools:
```
if (bb = cc)
{
// ...
}
while (bb = cc)
{
// ...
}
if ((a = b) > 3)
{
// ...
}
while ((a = b) < 4)
{
// ...
}
switch (bb = cc)
{
case true:
default:
}
```
I believe the pattern shown by the first two assignments-inside-conditionals can sometimes be a typo/mistake, and if not by a mistake the coder, could easily get mistaken by maintainers scanning over it as a comparison operator. Fortunately this isn't a very common problem since it generally only applies to bools. However, all of these examples have more straight-forward, logical implementations by pulling the assignments into their own lines of code. The "shortcuts" being employed by allowing the assignment to be within conditionals are unnecessary and I think we should provide a rule to call this out.
Originally proposed by Andy Reeves: http://stylecop.codeplex.com/workitem/6893
-----------------------------------------------------------
follow up discussion:
xanatos wrote Nov 7, 2011 at 11:38 AM
This is usable:
```
while ((a = b) < 4)
{
// ...
}
```
I often use it for:
```
int count;
while ((count = myStream.Read(buffer, 0, buffer.Length)) > 0)
{
mySecondStream.Write(buffer, 0, count);
}
```
The alternatives are all quite ugly:
This moves away from the while the exit logic:
```
while (true)
{
int count = myStream.Read(buffer, 0, buffer.Length);
if (count == 0)
{
break;
}
mySecondStream.Write(buffer, 0, count);
}
```
or
This is even worse, because it repeats twice the Read line (so the two Reads could become different)
```
int count = myStream.Read(buffer, 0, buffer.Length);
while (count > 0)
{
mySecondStream.Write(buffer, 0, count);
count = myStream.Read(buffer, 0, buffer.Length);
}
``` | True | Rule proposal - Readability (Maintainablility) rule - Report assignments in condition expressions - To carry on the side topic at http://stylecop.codeplex.com/discussions/252502...
StyleCop does not currently complain about either of the following code patterns. Where bb and cc are bools:
```
if (bb = cc)
{
// ...
}
while (bb = cc)
{
// ...
}
if ((a = b) > 3)
{
// ...
}
while ((a = b) < 4)
{
// ...
}
switch (bb = cc)
{
case true:
default:
}
```
I believe the pattern shown by the first two assignments-inside-conditionals can sometimes be a typo/mistake, and if not by a mistake the coder, could easily get mistaken by maintainers scanning over it as a comparison operator. Fortunately this isn't a very common problem since it generally only applies to bools. However, all of these examples have more straight-forward, logical implementations by pulling the assignments into their own lines of code. The "shortcuts" being employed by allowing the assignment to be within conditionals are unnecessary and I think we should provide a rule to call this out.
Originally proposed by Andy Reeves: http://stylecop.codeplex.com/workitem/6893
-----------------------------------------------------------
follow up discussion:
xanatos wrote Nov 7, 2011 at 11:38 AM
This is usable:
```
while ((a = b) < 4)
{
// ...
}
```
I often use it for:
```
int count;
while ((count = myStream.Read(buffer, 0, buffer.Length)) > 0)
{
mySecondStream.Write(buffer, 0, count);
}
```
The alternatives are all quite ugly:
This moves away from the while the exit logic:
```
while (true)
{
int count = myStream.Read(buffer, 0, buffer.Length);
if (count == 0)
{
break;
}
mySecondStream.Write(buffer, 0, count);
}
```
or
This is even worse, because it repeats twice the Read line (so the two Reads could become different)
```
int count = myStream.Read(buffer, 0, buffer.Length);
while (count > 0)
{
mySecondStream.Write(buffer, 0, count);
count = myStream.Read(buffer, 0, buffer.Length);
}
``` | main | rule proposal readability maintainablility rule report assignments in condition expressions to carry on the side topic at stylecop does not currently complain about either of the following code patterns where bb and cc are bools if bb cc while bb cc if a b while a b switch bb cc case true default i believe the pattern shown by the first two assignments inside conditionals can sometimes be a typo mistake and if not by a mistake the coder could easily get mistaken by maintainers scanning over it as a comparison operator fortunately this isn t a very common problem since it generally only applies to bools however all of these examples have more straight forward logical implementations by pulling the assignments into their own lines of code the shortcuts being employed by allowing the assignment to be within conditionals are unnecessary and i think we should provide a rule to call this out originally proposed by andy reeves follow up discussion xanatos wrote nov at am this is usable while a b i often use it for int count while count mystream read buffer buffer length mysecondstream write buffer count the alternatives are all quite ugly this moves away from the while the exit logic while true int count mystream read buffer buffer length if count break mysecondstream write buffer count or this is even worse because it repeats twice the read line so the two reads could become different int count mystream read buffer buffer length while count mysecondstream write buffer count count mystream read buffer buffer length | 1 |
20,251 | 3,800,355,968 | IssuesEvent | 2016-03-23 18:48:24 | servo/servo | https://api.github.com/repos/servo/servo | closed | rust.png in reftests is missing | A-testing C-assigned E-easy | It's used in inline_margin_multiple_fragments_a.html and filter_inline_a.html but it's 404. We should replace references to it with one of the other images in http://mxr.mozilla.org/servo/source/tests/wpt/mozilla/tests/css/ . | 1.0 | rust.png in reftests is missing - It's used in inline_margin_multiple_fragments_a.html and filter_inline_a.html but it's 404. We should replace references to it with one of the other images in http://mxr.mozilla.org/servo/source/tests/wpt/mozilla/tests/css/ . | non_main | rust png in reftests is missing it s used in inline margin multiple fragments a html and filter inline a html but it s we should replace references to it with one of the other images in | 0 |
5,434 | 27,243,567,134 | IssuesEvent | 2023-02-21 22:57:10 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | Unable to run docker in ARM Architecture | stage/needs-investigation maintainer/need-followup platform/mac/arm | <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- sam build fails in the Mac Os M1 Chip for lambdas that is using Docker.-->
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
1. Step : 1 Create a YAML for lambda that uses Docker Image
2. Step: 2 Add Docker File in the metadata of yaml which is similar to
`FROM python:3.6
WORKDIR /src
COPY main.py requirements.txt config.json ./
RUN apt-get update && apt-get install make git
RUN apt-get install -y apt-utils
RUN apt-get install -y cmake
RUN apt-get install -y librdkafka-dev
RUN pip install -r requirements.txt
ENTRYPOINT [ "/usr/local/bin/python", "-m", "awslambdaric" ]
CMD ["main.lambda_handler"]`
3. Fails as it tries to build in arm architecture
### Observed result:
<!-- Please provide command output with `--debug` flag set.-->
Fails to build the image
`creating build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src
gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -fPIC -I/usr/local/include/python3.6m -c /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c -o build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.o
In file included from /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c:17:
/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.h:66:2: error: #error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html"
#error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html"`
### Expected result:
<!-- Describe what you expected.-->
SAM CLI should automatically build Docker image for x86 till the support of Lambda is ready for Graviton
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS:MacOs M1
2. If using SAM CLI, `sam --version`: SAM CLI, version 1.24.0
3. AWS region: ap-southeast-2
`Add --debug flag to any SAM CLI commands you are running`
| True | Unable to run docker in ARM Architecture - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- sam build fails in the Mac Os M1 Chip for lambdas that is using Docker.-->
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
1. Step : 1 Create a YAML for lambda that uses Docker Image
2. Step: 2 Add Docker File in the metadata of yaml which is similar to
`FROM python:3.6
WORKDIR /src
COPY main.py requirements.txt config.json ./
RUN apt-get update && apt-get install make git
RUN apt-get install -y apt-utils
RUN apt-get install -y cmake
RUN apt-get install -y librdkafka-dev
RUN pip install -r requirements.txt
ENTRYPOINT [ "/usr/local/bin/python", "-m", "awslambdaric" ]
CMD ["main.lambda_handler"]`
3. Fails as it tries to build in arm architecture
### Observed result:
<!-- Please provide command output with `--debug` flag set.-->
Fails to build the image
`creating build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src
gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -fPIC -I/usr/local/include/python3.6m -c /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c -o build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.o
In file included from /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c:17:
/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.h:66:2: error: #error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html"
#error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html"`
### Expected result:
<!-- Describe what you expected.-->
SAM CLI should automatically build Docker image for x86 till the support of Lambda is ready for Graviton
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS:MacOs M1
2. If using SAM CLI, `sam --version`: SAM CLI, version 1.24.0
3. AWS region: ap-southeast-2
`Add --debug flag to any SAM CLI commands you are running`
| main | unable to run docker in arm architecture make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description steps to reproduce step create a yaml for lambda that uses docker image step add docker file in the metadata of yaml which is similar to from python workdir src copy main py requirements txt config json run apt get update apt get install make git run apt get install y apt utils run apt get install y cmake run apt get install y librdkafka dev run pip install r requirements txt entrypoint cmd fails as it tries to build in arm architecture observed result fails to build the image creating build temp linux tmp pip install confluent kafka src confluent kafka src gcc pthread wno unused result wsign compare dndebug g fwrapv wall fpic i usr local include c tmp pip install confluent kafka src confluent kafka src confluent kafka c o build temp linux tmp pip install confluent kafka src confluent kafka src confluent kafka o in file included from tmp pip install confluent kafka src confluent kafka src confluent kafka c tmp pip install confluent kafka src confluent kafka src confluent kafka h error error confluent kafka python requires librdkafka or later install the latest version of librdkafka from the confluent repositories see error confluent kafka python requires librdkafka or later install the latest version of librdkafka from the confluent repositories see expected result sam cli should automatically build docker image for till the support of lambda is ready for graviton additional environment details ex windows mac amazon linux etc os macos if using sam cli sam version sam cli version aws region ap southeast add debug flag to any sam cli commands you are running | 1 |
4,681 | 24,185,022,837 | IssuesEvent | 2022-09-23 12:35:02 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [Question]: How to dynamically render the expanded row content based on row? | type: question β status: waiting for maintainer response π¬ | ### Question for Carbon
I see in the given examples there are hardcoded values for expanded row like this.
```
{row.isExpanded && (
<TableExpandedRow colSpan={headers.length + 1}>
<h1>Expandable row content</h1>
<p>Description here</p>
</TableExpandedRow>
)}
```
The requirement is I need to pass the data for expanded row as part of rows. like
```
const rows = [
{
id: 'a',
field1: 'Field 1a',
expand: 'Value for Expanded row 1'
},
{
id: 'b',
field1: 'Field 1b',
expand: 'Value for Expanded row 2'
}
];
```
Is this possible?
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) | True | [Question]: How to dynamically render the expanded row content based on row? - ### Question for Carbon
I see in the given examples there are hardcoded values for expanded row like this.
```
{row.isExpanded && (
<TableExpandedRow colSpan={headers.length + 1}>
<h1>Expandable row content</h1>
<p>Description here</p>
</TableExpandedRow>
)}
```
The requirement is I need to pass the data for expanded row as part of rows. like
```
const rows = [
{
id: 'a',
field1: 'Field 1a',
expand: 'Value for Expanded row 1'
},
{
id: 'b',
field1: 'Field 1b',
expand: 'Value for Expanded row 2'
}
];
```
Is this possible?
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) | main | how to dynamically render the expanded row content based on row question for carbon i see in the given examples there are hardcoded values for expanded row like this row isexpanded expandable row content description here the requirement is i need to pass the data for expanded row as part of rows like const rows id a field expand value for expanded row id b field expand value for expanded row is this possible code of conduct i agree to follow this project s | 1 |
17,723 | 4,188,021,266 | IssuesEvent | 2016-06-23 19:19:38 | bignerdranch/expandable-recycler-view | https://api.github.com/repos/bignerdranch/expandable-recycler-view | closed | Documentation error: Need LayoutManager | bug documentation | In the [original blog post](https://www.bignerdranch.com/blog/expand-a-recyclerview-in-four-steps/) you reference adding a layout manager, but [this page with the updated information](http://bignerdranch.github.io/expandable-recycler-view/) doesn't. Took me awhile to figure out why nothing was showing up! | 1.0 | Documentation error: Need LayoutManager - In the [original blog post](https://www.bignerdranch.com/blog/expand-a-recyclerview-in-four-steps/) you reference adding a layout manager, but [this page with the updated information](http://bignerdranch.github.io/expandable-recycler-view/) doesn't. Took me awhile to figure out why nothing was showing up! | non_main | documentation error need layoutmanager in the you reference adding a layout manager but doesn t took me awhile to figure out why nothing was showing up | 0 |
2,530 | 8,657,247,698 | IssuesEvent | 2018-11-27 20:45:31 | arcticicestudio/nord-docs | https://api.github.com/repos/arcticicestudio/nord-docs | opened | Netlify Configuration | context-workflow scope-configurability scope-maintainability type-feature | <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48661237-35d1a000-ea6f-11e8-8e16-f48948969be6.png" width="60%" /></p>
> Related epics: #46
This issue documents a part of the implementation of the [hosting & continuous deployment concept][gh-46] with the [Netlify's configuration file][netlify-docs-toml-ref].
See the βHostingβ and βContinuous Deploymentβ (sub)sections for more details about the architecture.
## Tasks
- Implement Netlify's `netlify.toml` configuration file
- [ ] Define the `command` for the production `[build]` section
- [ ] Define the `publish` path for the production `[build]` section
[gh-46]: https://github.com/arcticicestudio/nord-docs/issues/46
[netlify-docs-toml-ref]: https://www.netlify.com/docs/netlify-toml-reference | True | Netlify Configuration - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48661237-35d1a000-ea6f-11e8-8e16-f48948969be6.png" width="60%" /></p>
> Related epics: #46
This issue documents a part of the implementation of the [hosting & continuous deployment concept][gh-46] with the [Netlify's configuration file][netlify-docs-toml-ref].
See the βHostingβ and βContinuous Deploymentβ (sub)sections for more details about the architecture.
## Tasks
- Implement Netlify's `netlify.toml` configuration file
- [ ] Define the `command` for the production `[build]` section
- [ ] Define the `publish` path for the production `[build]` section
[gh-46]: https://github.com/arcticicestudio/nord-docs/issues/46
[netlify-docs-toml-ref]: https://www.netlify.com/docs/netlify-toml-reference | main | netlify configuration related epics this issue documents a part of the implementation of the with the see the βhostingβ and βcontinuous deploymentβ sub sections for more details about the architecture tasks implement netlify s netlify toml configuration file define the command for the production section define the publish path for the production section | 1 |
1,958 | 6,678,594,345 | IssuesEvent | 2017-10-05 14:42:45 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | closed | What3Words Geocoder: Fallback to StandardBlend (autosuggest) API when no results for three-word address | Low-Hanging Fruit Maintainer Approved Maintainer Submitted Suggestion | When a user misspells a three word address they will currently see no results. The What3Words team suggested we try faling back to their [StandardBlend](https://docs.what3words.com/api/v2/#standardblend) (Autocomplete) API in those cases to show potential matches for the given address.
If the first API call returns no results, we can make another API call and if it has results, display a Places tile view (similar to the [BikeShare Spices](https://duckduckgo.com/?q=nyc+bike+share+locations&ia=bikesharing)) which provides the similar addresses.
This will require creating another `Spice alt_to` endpoint for the StandardBlend API Endpoint.
We should use the `text` template for the tiles. The title should be the 3-Word address, and the subtitle should be the location the place is near to (provided by the API). If possible, it would be nice to use our flag icons as well, if possible but that might be difficult.
---
IA Page: http://duck.co/ia/view/what3words
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @moollaza
| True | What3Words Geocoder: Fallback to StandardBlend (autosuggest) API when no results for three-word address - When a user misspells a three word address they will currently see no results. The What3Words team suggested we try faling back to their [StandardBlend](https://docs.what3words.com/api/v2/#standardblend) (Autocomplete) API in those cases to show potential matches for the given address.
If the first API call returns no results, we can make another API call and if it has results, display a Places tile view (similar to the [BikeShare Spices](https://duckduckgo.com/?q=nyc+bike+share+locations&ia=bikesharing)) which provides the similar addresses.
This will require creating another `Spice alt_to` endpoint for the StandardBlend API Endpoint.
We should use the `text` template for the tiles. The title should be the 3-Word address, and the subtitle should be the location the place is near to (provided by the API). If possible, it would be nice to use our flag icons as well, if possible but that might be difficult.
---
IA Page: http://duck.co/ia/view/what3words
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @moollaza
| main | geocoder fallback to standardblend autosuggest api when no results for three word address when a user misspells a three word address they will currently see no results the team suggested we try faling back to their autocomplete api in those cases to show potential matches for the given address if the first api call returns no results we can make another api call and if it has results display a places tile view similar to the which provides the similar addresses this will require creating another spice alt to endpoint for the standardblend api endpoint we should use the text template for the tiles the title should be the word address and the subtitle should be the location the place is near to provided by the api if possible it would be nice to use our flag icons as well if possible but that might be difficult ia page moollaza | 1 |
208,746 | 16,136,311,721 | IssuesEvent | 2021-04-29 12:19:16 | gatsbyjs/gatsby | https://api.github.com/repos/gatsbyjs/gatsby | closed | [docs] update docs for createRemoteFilenode usage | not stale type: documentation | ## Summary
I was using `createRemoteFilenode` and found some things in the docs that could use an update
- using the examples on [this page](https://www.gatsbyjs.org/docs/preprocessing-external-images/#gatsby-node) I got a warning `warn Deprecation warning - adding inferred resolver for field GhostPost.feature_image_localFile. In Gatsby v3, only fields with an explicit directive/extension will get a resolver.` so I think the example should be changed to (not interily sure about this, since I'm not familiar with `@infer` / `@noInfer`):
```
...
createTypes(`
type MarkdownRemark implements Node @infer(noDefaultResolvers: false) {
frontmatter: Frontmatter
}
type Frontmatter {
title: String!
featuredImgUrl: String
featuredImgAlt: String
}
`)
...
```
- the readme if the [`gatsby-source-filesystem`](https://www.gatsbyjs.org/packages/gatsby-source-filesystem/?=filesys#createremotefilenode) doesn't really have a clear usage example (to me at least) since there is no context given as to where it is recommended to use `createRemoteFilenode` , maybe we can add a link to above mentioned page?
not sure if anyone is working on any of these pages? if not I can make a pr for this. | 1.0 | [docs] update docs for createRemoteFilenode usage - ## Summary
I was using `createRemoteFilenode` and found some things in the docs that could use an update
- using the examples on [this page](https://www.gatsbyjs.org/docs/preprocessing-external-images/#gatsby-node) I got a warning `warn Deprecation warning - adding inferred resolver for field GhostPost.feature_image_localFile. In Gatsby v3, only fields with an explicit directive/extension will get a resolver.` so I think the example should be changed to (not interily sure about this, since I'm not familiar with `@infer` / `@noInfer`):
```
...
createTypes(`
type MarkdownRemark implements Node @infer(noDefaultResolvers: false) {
frontmatter: Frontmatter
}
type Frontmatter {
title: String!
featuredImgUrl: String
featuredImgAlt: String
}
`)
...
```
- the readme if the [`gatsby-source-filesystem`](https://www.gatsbyjs.org/packages/gatsby-source-filesystem/?=filesys#createremotefilenode) doesn't really have a clear usage example (to me at least) since there is no context given as to where it is recommended to use `createRemoteFilenode` , maybe we can add a link to above mentioned page?
not sure if anyone is working on any of these pages? if not I can make a pr for this. | non_main | update docs for createremotefilenode usage summary i was using createremotefilenode and found some things in the docs that could use an update using the examples on i got a warning warn deprecation warning adding inferred resolver for field ghostpost feature image localfile in gatsby only fields with an explicit directive extension will get a resolver so i think the example should be changed to not interily sure about this since i m not familiar with infer noinfer createtypes type markdownremark implements node infer nodefaultresolvers false frontmatter frontmatter type frontmatter title string featuredimgurl string featuredimgalt string the readme if the doesn t really have a clear usage example to me at least since there is no context given as to where it is recommended to use createremotefilenode maybe we can add a link to above mentioned page not sure if anyone is working on any of these pages if not i can make a pr for this | 0 |
72,153 | 8,707,631,539 | IssuesEvent | 2018-12-06 08:32:02 | phetsims/a11y-research | https://api.github.com/repos/phetsims/a11y-research | closed | How well are Arrow keys working in JAWS for 4-way custom move? | design:a11y dev:a11y meeting:a11y type:question | I am having some trouble assessing what is the best way to cue the custom 4-way drag, especially for Friction where pressing the W key is not so useful.
In a recent session with a VoiceOver user familiar with our sims, there seemed to be confusion around cuing only the use of the WASD keys for moving the book.
**Goal of this issue:**
Assess how well the Arrow keys are working in JAWS so we can figure out the best way to provide instructions for the interaction and the best way to provide alternatives for the interaction.
**Question 1 for @jessegreenberg**
Do you know how well the Arrow keys are working with JAWS now? In early prototypes of BASE the Arrow keys were not working at all with JAWS, and then they worked, but not as well as the WASD keys.
**Question 2 for @terracoda & @emily-phet**
If JAWS is no longer having a serious problem with releasing the Arrow keys to the web application, how and where is the best way to provide the alternative WASD key instructions?
- In PDOM help text, first grab alert, **AND** Keyboard Shortcuts dialog?
- Only in PDOM help text **AND** Keyboard Shortcuts? **OR**
- Only in the Keyboard Shortcuts dialog?
Currently, we do option 1, and we only explicitly refer to using the Arrow keys in the Keyboard Shortcuts dialog. Here are examples from BASE and Friction, and soon there will be a grab buttons in GFL and Faraday's Law.
**Cuing text for BASE**
- PDOM help text:
- "Look for grab button to play. Once grabbed, press W, A, S, or D key to move up, left, down, or right. Space to release."
- Initial Grab Alert:
- "Grabbed. At center of Play Area. Has no more negative charges than positive charges. Press W, A, S, or D key to move balloon. Space to release."
- Keyboard Dialog content for grabbed balloon:
- "Move grabbed balloon up, left, down, or right with Arrow keys or with letter keys W, A, S, or D."
**Cuing text for Friction**
- PDOM help text:
- "Look for grab buttons. Once grabbed, use letter keys W, A, S, or D to move book or zoomed-in book up, left, down, or right."
- Initial Grab Alerts before successful interaction:
- Not touching: "Grabbed. Lightly on Physics book. Use W, A, S, or D keys to move book. Space to release."
- Touching: "Grabbed. Rub fast or slow with A or D keys. Space to release."
- and immediately after a grab alert the user also hears, "Atoms jiggle a tiny bit, temperature cool"
Keyboard Dialog content Friction's grabbed book:
- "Move grabbed book up, left, down, or right with Arrow keys, or with letter keys W, A, S, or D."
| 1.0 | How well are Arrow keys working in JAWS for 4-way custom move? - I am having some trouble assessing what is the best way to cue the custom 4-way drag, especially for Friction where pressing the W key is not so useful.
In a recent session with a VoiceOver user familiar with our sims, there seemed to be confusion around cuing only the use of the WASD keys for moving the book.
**Goal of this issue:**
Assess how well the Arrow keys are working in JAWS so we can figure out the best way to provide instructions for the interaction and the best way to provide alternatives for the interaction.
**Question 1 for @jessegreenberg**
Do you know how well the Arrow keys are working with JAWS now? In early prototypes of BASE the Arrow keys were not working at all with JAWS, and then they worked, but not as well as the WASD keys.
**Question 2 for @terracoda & @emily-phet**
If JAWS is no longer having a serious problem with releasing the Arrow keys to the web application, how and where is the best way to provide the alternative WASD key instructions?
- In PDOM help text, first grab alert, **AND** Keyboard Shortcuts dialog?
- Only in PDOM help text **AND** Keyboard Shortcuts? **OR**
- Only in the Keyboard Shortcuts dialog?
Currently, we do option 1, and we only explicitly refer to using the Arrow keys in the Keyboard Shortcuts dialog. Here are examples from BASE and Friction, and soon there will be a grab buttons in GFL and Faraday's Law.
**Cuing text for BASE**
- PDOM help text:
- "Look for grab button to play. Once grabbed, press W, A, S, or D key to move up, left, down, or right. Space to release."
- Initial Grab Alert:
- "Grabbed. At center of Play Area. Has no more negative charges than positive charges. Press W, A, S, or D key to move balloon. Space to release."
- Keyboard Dialog content for grabbed balloon:
- "Move grabbed balloon up, left, down, or right with Arrow keys or with letter keys W, A, S, or D."
**Cuing text for Friction**
- PDOM help text:
- "Look for grab buttons. Once grabbed, use letter keys W, A, S, or D to move book or zoomed-in book up, left, down, or right."
- Initial Grab Alerts before successful interaction:
- Not touching: "Grabbed. Lightly on Physics book. Use W, A, S, or D keys to move book. Space to release."
- Touching: "Grabbed. Rub fast or slow with A or D keys. Space to release."
- and immediately after a grab alert the user also hears, "Atoms jiggle a tiny bit, temperature cool"
Keyboard Dialog content Friction's grabbed book:
- "Move grabbed book up, left, down, or right with Arrow keys, or with letter keys W, A, S, or D."
| non_main | how well are arrow keys working in jaws for way custom move i am having some trouble assessing what is the best way to cue the custom way drag especially for friction where pressing the w key is not so useful in a recent session with a voiceover user familiar with our sims there seemed to be confusion around cuing only the use of the wasd keys for moving the book goal of this issue assess how well the arrow keys are working in jaws so we can figure out the best way to provide instructions for the interaction and the best way to provide alternatives for the interaction question for jessegreenberg do you know how well the arrow keys are working with jaws now in early prototypes of base the arrow keys were not working at all with jaws and then they worked but not as well as the wasd keys question for terracoda emily phet if jaws is no longer having a serious problem with releasing the arrow keys to the web application how and where is the best way to provide the alternative wasd key instructions in pdom help text first grab alert and keyboard shortcuts dialog only in pdom help text and keyboard shortcuts or only in the keyboard shortcuts dialog currently we do option and we only explicitly refer to using the arrow keys in the keyboard shortcuts dialog here are examples from base and friction and soon there will be a grab buttons in gfl and faraday s law cuing text for base pdom help text look for grab button to play once grabbed press w a s or d key to move up left down or right space to release initial grab alert grabbed at center of play area has no more negative charges than positive charges press w a s or d key to move balloon space to release keyboard dialog content for grabbed balloon move grabbed balloon up left down or right with arrow keys or with letter keys w a s or d cuing text for friction pdom help text look for grab buttons once grabbed use letter keys w a s or d to move book or zoomed in book up left down or right initial grab alerts before successful interaction not touching grabbed lightly on physics book use w a s or d keys to move book space to release touching grabbed rub fast or slow with a or d keys space to release and immediately after a grab alert the user also hears atoms jiggle a tiny bit temperature cool keyboard dialog content friction s grabbed book move grabbed book up left down or right with arrow keys or with letter keys w a s or d | 0 |
5,559 | 27,808,349,143 | IssuesEvent | 2023-03-17 22:47:54 | microsoft/DirectXTex | https://api.github.com/repos/microsoft/DirectXTex | closed | Retire legacy Xbox One XDK support | maintainence | The only scenario that still uses VS 2017 is for the legacy Xbox One XDK. This task is drop support for this older Xbox development model and remove the following projects:
```
DirectXTex_XboxOneXDK_2017.sln
DirectXTex_XboxOneXDK_PC_2017.sln
```
> The end-of-life release will be hosted on https://github.com/microsoft/Xbox-ATG-Samples and will likely be the January 2023 release. | True | Retire legacy Xbox One XDK support - The only scenario that still uses VS 2017 is for the legacy Xbox One XDK. This task is drop support for this older Xbox development model and remove the following projects:
```
DirectXTex_XboxOneXDK_2017.sln
DirectXTex_XboxOneXDK_PC_2017.sln
```
> The end-of-life release will be hosted on https://github.com/microsoft/Xbox-ATG-Samples and will likely be the January 2023 release. | main | retire legacy xbox one xdk support the only scenario that still uses vs is for the legacy xbox one xdk this task is drop support for this older xbox development model and remove the following projects directxtex xboxonexdk sln directxtex xboxonexdk pc sln the end of life release will be hosted on and will likely be the january release | 1 |
146,332 | 13,177,933,232 | IssuesEvent | 2020-08-12 08:16:56 | legokor/UPRA-doksik | https://api.github.com/repos/legokor/UPRA-doksik | closed | UPRA for Dummies | documentation | UPRA for Dummies doksik kΓ©szΓtΓ©ser
-~~ΓltalΓ‘nos ΓΆsszefoglalΓ³~~
-~~OBC~~
-~~COM~~
-~~EPS~~
-DAU
-Interface
-BUS
-~~GND~~ | 1.0 | UPRA for Dummies - UPRA for Dummies doksik kΓ©szΓtΓ©ser
-~~ΓltalΓ‘nos ΓΆsszefoglalΓ³~~
-~~OBC~~
-~~COM~~
-~~EPS~~
-DAU
-Interface
-BUS
-~~GND~~ | non_main | upra for dummies upra for dummies doksik kΓ©szΓtΓ©ser Γ‘ltalΓ‘nos ΓΆsszefoglalΓ³ obc com eps dau interface bus gnd | 0 |
4,543 | 23,662,203,313 | IssuesEvent | 2022-08-26 16:42:03 | precice/precice | https://api.github.com/repos/precice/precice | opened | Resetting sent data in coupling scheme to zero after sending affects downstream calculations. Why? | bug maintainability | **Describe your setup**
*will provide link to commit in a moment*
**Describe the problem**
To clean up things and to make sure there are no strange, I changed the source code to reset data after sending and before receiving in the coupling scheme (see link to commit above). However, I observed that some tests are failing after this change.
**Step To Reproduce**
1. Check out branch
2. Run tests
**Expected behaviour**
As far as I know the data should not be used anymore after sending. I also think that this would be the most intuitive behavior. Any idea why this is happening?
**Additional context**
This is a pure software engineering issue. But it might also be a bug. I'm not sure here. | True | Resetting sent data in coupling scheme to zero after sending affects downstream calculations. Why? - **Describe your setup**
*will provide link to commit in a moment*
**Describe the problem**
To clean up things and to make sure there are no strange, I changed the source code to reset data after sending and before receiving in the coupling scheme (see link to commit above). However, I observed that some tests are failing after this change.
**Step To Reproduce**
1. Check out branch
2. Run tests
**Expected behaviour**
As far as I know the data should not be used anymore after sending. I also think that this would be the most intuitive behavior. Any idea why this is happening?
**Additional context**
This is a pure software engineering issue. But it might also be a bug. I'm not sure here. | main | resetting sent data in coupling scheme to zero after sending affects downstream calculations why describe your setup will provide link to commit in a moment describe the problem to clean up things and to make sure there are no strange i changed the source code to reset data after sending and before receiving in the coupling scheme see link to commit above however i observed that some tests are failing after this change step to reproduce check out branch run tests expected behaviour as far as i know the data should not be used anymore after sending i also think that this would be the most intuitive behavior any idea why this is happening additional context this is a pure software engineering issue but it might also be a bug i m not sure here | 1 |
129,772 | 27,559,625,948 | IssuesEvent | 2023-03-07 20:48:48 | openxla/iree | https://api.github.com/repos/openxla/iree | opened | SCF::TileAndFuse produces bad IR for multi result generic op + pack op cases | codegen | I'm working on pack op fusion. I used aggressive fusion to get more data point and found that the SCF tile and fuse does not work well for some cases. If we apply TileAndFuse on `multi result generic ops` and `tensor.pack` ops, it will remain a generic op outside the scf.for loop. E.g.,
Input IR:
```mlir
func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() {
%c256 = arith.constant 256 : index
%c48 = arith.constant 48 : index
%c0 = arith.constant 0 : index
%c786432 = arith.constant 786432 : index
%c1572864 = arith.constant 1572864 : index
%0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
%6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%workgroup_id_x = hal.interface.workgroup.id[0] : index
%workgroup_count_x = hal.interface.workgroup.count[0] : index
%workgroup_id_y = hal.interface.workgroup.id[1] : index
%workgroup_count_y = hal.interface.workgroup.count[1] : index
%8 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_id_y]
%9 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_count_y]
scf.for %arg0 = %8 to %c48 step %9 {
%10 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_id_x]
%11 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_count_x]
scf.for %arg1 = %10 to %c256 step %11 {
%12 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32>
%13 = affine.apply affine_map<(d0) -> (d0 * 8)>(%arg0)
%14 = affine.apply affine_map<(d0) -> (d0 * 2)>(%arg1)
%15 = flow.dispatch.tensor.load %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%16 = flow.dispatch.tensor.load %0, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%17 = flow.dispatch.tensor.load %1, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%18 = flow.dispatch.tensor.load %2, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%19 = flow.dispatch.tensor.load %3, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%20 = flow.dispatch.tensor.load %4, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%21 = tensor.empty() : tensor<128x64xf32>
%22:2 = linalg.generic {indexing_maps = [affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>], iterator_types = ["parallel", "parallel"]} ins(%16, %17, %18, %19, %20 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%21, %15 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} {
^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32):
%23 = arith.addf %in, %in_0 : f32
%24 = arith.addf %23, %in_1 : f32
%25 = arith.mulf %24, %in_2 : f32
%26 = arith.addf %25, %in_3 : f32
linalg.yield %24, %26 : f32, f32
} -> (tensor<128x64xf32>, tensor<128x64xf32>)
%pack = tensor.pack %22#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %12 {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} : tensor<128x64xf32> -> tensor<16x32x8x2xf32>
flow.dispatch.tensor.store %pack, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
flow.dispatch.tensor.store %22#0, %6, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
}
}
return
}
```
After running `iree-opt --linalg-fuse="tile-sizes=1,16" repro.mlir`:
```mlir
#config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>
#map = affine_map<()[s0] -> (s0 * 16)>
#map1 = affine_map<()[s0] -> (s0 * 32)>
#map2 = affine_map<(d0) -> (d0 * 8)>
#map3 = affine_map<(d0) -> (d0 * 2)>
#map4 = affine_map<(d0, d1) -> (d1)>
#map5 = affine_map<(d0, d1) -> (d0, d1)>
module {
func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() {
%c32 = arith.constant 32 : index
%c1 = arith.constant 1 : index
%c16 = arith.constant 16 : index
%c256 = arith.constant 256 : index
%c48 = arith.constant 48 : index
%c0 = arith.constant 0 : index
%c786432 = arith.constant 786432 : index
%c1572864 = arith.constant 1572864 : index
%0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
%6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%workgroup_id_x = hal.interface.workgroup.id[0] : index
%workgroup_count_x = hal.interface.workgroup.count[0] : index
%workgroup_id_y = hal.interface.workgroup.id[1] : index
%workgroup_count_y = hal.interface.workgroup.count[1] : index
%8 = affine.apply #map()[%workgroup_id_y]
%9 = affine.apply #map()[%workgroup_count_y]
%10 = affine.apply #map1()[%workgroup_id_x]
%11 = affine.apply #map1()[%workgroup_count_x]
%12 = tensor.empty() : tensor<128x64xf32>
scf.for %arg0 = %8 to %c48 step %9 {
%13 = affine.apply #map2(%arg0)
scf.for %arg1 = %10 to %c256 step %11 {
%14 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32>
%15 = affine.apply #map3(%arg1)
%16 = flow.dispatch.tensor.load %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%17 = flow.dispatch.tensor.load %0, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%18 = flow.dispatch.tensor.load %1, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%19 = flow.dispatch.tensor.load %2, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%20 = flow.dispatch.tensor.load %3, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%21 = flow.dispatch.tensor.load %4, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%22:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%17, %18, %19, %20, %21 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%12, %16 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #config} {
^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32):
%24 = arith.addf %in, %in_0 : f32
%25 = arith.addf %24, %in_1 : f32
%26 = arith.mulf %25, %in_2 : f32
%27 = arith.addf %26, %in_3 : f32
linalg.yield %25, %27 : f32, f32
} -> (tensor<128x64xf32>, tensor<128x64xf32>)
%23:2 = scf.for %arg2 = %c0 to %c16 step %c1 iter_args(%arg3 = %14, %arg4 = %12) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) {
%24 = affine.apply #map2(%arg2)
%25:2 = scf.for %arg5 = %c0 to %c32 step %c16 iter_args(%arg6 = %arg3, %arg7 = %arg4) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) {
%26 = affine.apply #map3(%arg5)
%extracted_slice = tensor.extract_slice %17[%26] [32] [1] : tensor<64xf32> to tensor<32xf32>
%extracted_slice_0 = tensor.extract_slice %18[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%extracted_slice_1 = tensor.extract_slice %19[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%extracted_slice_2 = tensor.extract_slice %20[%26] [32] [1] : tensor<64xf32> to tensor<32xf32>
%extracted_slice_3 = tensor.extract_slice %21[%26] [32] [1] : tensor<64xf32> to tensor<32xf32>
%extracted_slice_4 = tensor.extract_slice %arg7[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%extracted_slice_5 = tensor.extract_slice %16[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%27:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%extracted_slice, %extracted_slice_0, %extracted_slice_1, %extracted_slice_2, %extracted_slice_3 : tensor<32xf32>, tensor<8x32xf32>, tensor<8x32xf32>, tensor<32xf32>, tensor<32xf32>) outs(%extracted_slice_4, %extracted_slice_5 : tensor<8x32xf32>, tensor<8x32xf32>) attrs = {lowering_config = #config} {
^bb0(%in: f32, %in_8: f32, %in_9: f32, %in_10: f32, %in_11: f32, %out: f32, %out_12: f32):
%28 = arith.addf %in, %in_8 : f32
%29 = arith.addf %28, %in_9 : f32
%30 = arith.mulf %29, %in_10 : f32
%31 = arith.addf %30, %in_11 : f32
linalg.yield %29, %31 : f32, f32
} -> (tensor<8x32xf32>, tensor<8x32xf32>)
%extracted_slice_6 = tensor.extract_slice %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<16x32x8x2xf32> to tensor<1x16x8x2xf32>
%pack = tensor.pack %27#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %extracted_slice_6 {__internal_linalg_transform__ = "1", lowering_config = #config} : tensor<8x32xf32> -> tensor<1x16x8x2xf32>
%inserted_slice = tensor.insert_slice %pack into %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<1x16x8x2xf32> into tensor<16x32x8x2xf32>
%inserted_slice_7 = tensor.insert_slice %27#0 into %arg7[%24, %26] [8, 32] [1, 1] : tensor<8x32xf32> into tensor<128x64xf32>
scf.yield %inserted_slice, %inserted_slice_7 : tensor<16x32x8x2xf32>, tensor<128x64xf32>
}
scf.yield %25#0, %25#1 : tensor<16x32x8x2xf32>, tensor<128x64xf32>
}
flow.dispatch.tensor.store %23#0, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
flow.dispatch.tensor.store %23#1, %6, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
}
}
return
}
}
```
What I'm expecting is that the `%22` is removed and the scf.for loop should return three values. This drops the performance a lot. | 1.0 | SCF::TileAndFuse produces bad IR for multi result generic op + pack op cases - I'm working on pack op fusion. I used aggressive fusion to get more data point and found that the SCF tile and fuse does not work well for some cases. If we apply TileAndFuse on `multi result generic ops` and `tensor.pack` ops, it will remain a generic op outside the scf.for loop. E.g.,
Input IR:
```mlir
func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() {
%c256 = arith.constant 256 : index
%c48 = arith.constant 48 : index
%c0 = arith.constant 0 : index
%c786432 = arith.constant 786432 : index
%c1572864 = arith.constant 1572864 : index
%0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
%6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%workgroup_id_x = hal.interface.workgroup.id[0] : index
%workgroup_count_x = hal.interface.workgroup.count[0] : index
%workgroup_id_y = hal.interface.workgroup.id[1] : index
%workgroup_count_y = hal.interface.workgroup.count[1] : index
%8 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_id_y]
%9 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_count_y]
scf.for %arg0 = %8 to %c48 step %9 {
%10 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_id_x]
%11 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_count_x]
scf.for %arg1 = %10 to %c256 step %11 {
%12 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32>
%13 = affine.apply affine_map<(d0) -> (d0 * 8)>(%arg0)
%14 = affine.apply affine_map<(d0) -> (d0 * 2)>(%arg1)
%15 = flow.dispatch.tensor.load %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%16 = flow.dispatch.tensor.load %0, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%17 = flow.dispatch.tensor.load %1, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%18 = flow.dispatch.tensor.load %2, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%19 = flow.dispatch.tensor.load %3, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%20 = flow.dispatch.tensor.load %4, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%21 = tensor.empty() : tensor<128x64xf32>
%22:2 = linalg.generic {indexing_maps = [affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>], iterator_types = ["parallel", "parallel"]} ins(%16, %17, %18, %19, %20 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%21, %15 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} {
^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32):
%23 = arith.addf %in, %in_0 : f32
%24 = arith.addf %23, %in_1 : f32
%25 = arith.mulf %24, %in_2 : f32
%26 = arith.addf %25, %in_3 : f32
linalg.yield %24, %26 : f32, f32
} -> (tensor<128x64xf32>, tensor<128x64xf32>)
%pack = tensor.pack %22#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %12 {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} : tensor<128x64xf32> -> tensor<16x32x8x2xf32>
flow.dispatch.tensor.store %pack, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
flow.dispatch.tensor.store %22#0, %6, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
}
}
return
}
```
After running `iree-opt --linalg-fuse="tile-sizes=1,16" repro.mlir`:
```mlir
#config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>
#map = affine_map<()[s0] -> (s0 * 16)>
#map1 = affine_map<()[s0] -> (s0 * 32)>
#map2 = affine_map<(d0) -> (d0 * 8)>
#map3 = affine_map<(d0) -> (d0 * 2)>
#map4 = affine_map<(d0, d1) -> (d1)>
#map5 = affine_map<(d0, d1) -> (d0, d1)>
module {
func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() {
%c32 = arith.constant 32 : index
%c1 = arith.constant 1 : index
%c16 = arith.constant 16 : index
%c256 = arith.constant 256 : index
%c48 = arith.constant 48 : index
%c0 = arith.constant 0 : index
%c786432 = arith.constant 786432 : index
%c1572864 = arith.constant 1572864 : index
%0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>>
%3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>>
%5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
%6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
%workgroup_id_x = hal.interface.workgroup.id[0] : index
%workgroup_count_x = hal.interface.workgroup.count[0] : index
%workgroup_id_y = hal.interface.workgroup.id[1] : index
%workgroup_count_y = hal.interface.workgroup.count[1] : index
%8 = affine.apply #map()[%workgroup_id_y]
%9 = affine.apply #map()[%workgroup_count_y]
%10 = affine.apply #map1()[%workgroup_id_x]
%11 = affine.apply #map1()[%workgroup_count_x]
%12 = tensor.empty() : tensor<128x64xf32>
scf.for %arg0 = %8 to %c48 step %9 {
%13 = affine.apply #map2(%arg0)
scf.for %arg1 = %10 to %c256 step %11 {
%14 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32>
%15 = affine.apply #map3(%arg1)
%16 = flow.dispatch.tensor.load %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%17 = flow.dispatch.tensor.load %0, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%18 = flow.dispatch.tensor.load %1, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%19 = flow.dispatch.tensor.load %2, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32>
%20 = flow.dispatch.tensor.load %3, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%21 = flow.dispatch.tensor.load %4, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32>
%22:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%17, %18, %19, %20, %21 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%12, %16 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #config} {
^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32):
%24 = arith.addf %in, %in_0 : f32
%25 = arith.addf %24, %in_1 : f32
%26 = arith.mulf %25, %in_2 : f32
%27 = arith.addf %26, %in_3 : f32
linalg.yield %25, %27 : f32, f32
} -> (tensor<128x64xf32>, tensor<128x64xf32>)
%23:2 = scf.for %arg2 = %c0 to %c16 step %c1 iter_args(%arg3 = %14, %arg4 = %12) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) {
%24 = affine.apply #map2(%arg2)
%25:2 = scf.for %arg5 = %c0 to %c32 step %c16 iter_args(%arg6 = %arg3, %arg7 = %arg4) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) {
%26 = affine.apply #map3(%arg5)
%extracted_slice = tensor.extract_slice %17[%26] [32] [1] : tensor<64xf32> to tensor<32xf32>
%extracted_slice_0 = tensor.extract_slice %18[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%extracted_slice_1 = tensor.extract_slice %19[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%extracted_slice_2 = tensor.extract_slice %20[%26] [32] [1] : tensor<64xf32> to tensor<32xf32>
%extracted_slice_3 = tensor.extract_slice %21[%26] [32] [1] : tensor<64xf32> to tensor<32xf32>
%extracted_slice_4 = tensor.extract_slice %arg7[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%extracted_slice_5 = tensor.extract_slice %16[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32>
%27:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%extracted_slice, %extracted_slice_0, %extracted_slice_1, %extracted_slice_2, %extracted_slice_3 : tensor<32xf32>, tensor<8x32xf32>, tensor<8x32xf32>, tensor<32xf32>, tensor<32xf32>) outs(%extracted_slice_4, %extracted_slice_5 : tensor<8x32xf32>, tensor<8x32xf32>) attrs = {lowering_config = #config} {
^bb0(%in: f32, %in_8: f32, %in_9: f32, %in_10: f32, %in_11: f32, %out: f32, %out_12: f32):
%28 = arith.addf %in, %in_8 : f32
%29 = arith.addf %28, %in_9 : f32
%30 = arith.mulf %29, %in_10 : f32
%31 = arith.addf %30, %in_11 : f32
linalg.yield %29, %31 : f32, f32
} -> (tensor<8x32xf32>, tensor<8x32xf32>)
%extracted_slice_6 = tensor.extract_slice %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<16x32x8x2xf32> to tensor<1x16x8x2xf32>
%pack = tensor.pack %27#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %extracted_slice_6 {__internal_linalg_transform__ = "1", lowering_config = #config} : tensor<8x32xf32> -> tensor<1x16x8x2xf32>
%inserted_slice = tensor.insert_slice %pack into %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<1x16x8x2xf32> into tensor<16x32x8x2xf32>
%inserted_slice_7 = tensor.insert_slice %27#0 into %arg7[%24, %26] [8, 32] [1, 1] : tensor<8x32xf32> into tensor<128x64xf32>
scf.yield %inserted_slice, %inserted_slice_7 : tensor<16x32x8x2xf32>, tensor<128x64xf32>
}
scf.yield %25#0, %25#1 : tensor<16x32x8x2xf32>, tensor<128x64xf32>
}
flow.dispatch.tensor.store %23#0, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>>
flow.dispatch.tensor.store %23#1, %6, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>>
}
}
return
}
}
```
What I'm expecting is that the `%22` is removed and the scf.for loop should return three values. This drops the performance a lot. | non_main | scf tileandfuse produces bad ir for multi result generic op pack op cases i m working on pack op fusion i used aggressive fusion to get more data point and found that the scf tile and fuse does not work well for some cases if we apply tileandfuse on multi result generic ops and tensor pack ops it will remain a generic op outside the scf for loop e g input ir mlir func func main dispatch generic dispatch generic arith constant index arith constant index arith constant index arith constant index arith constant index hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor workgroup id x hal interface workgroup id index workgroup count x hal interface workgroup count index workgroup id y hal interface workgroup id index workgroup count y hal interface workgroup count index affine apply affine map affine apply affine map scf for to step affine apply affine map affine apply affine map scf for to step flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor affine apply affine map affine apply affine map flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor tensor empty tensor linalg generic indexing maps iterator types ins tensor tensor tensor tensor tensor outs tensor tensor attrs lowering config iree codegen lowering config in in in in in out out arith addf in in arith addf in arith mulf in arith addf in linalg yield tensor tensor pack tensor pack inner dims pos inner tiles into lowering config iree codegen lowering config tensor tensor flow dispatch tensor store pack offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor return after running iree opt linalg fuse tile sizes repro mlir mlir config iree codegen lowering config map affine map affine map affine map affine map affine map affine map module func func main dispatch generic dispatch generic arith constant index arith constant index arith constant index arith constant index arith constant index arith constant index arith constant index arith constant index hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor workgroup id x hal interface workgroup id index workgroup count x hal interface workgroup count index workgroup id y hal interface workgroup id index workgroup count y hal interface workgroup count index affine apply map affine apply map affine apply affine apply tensor empty tensor scf for to step affine apply scf for to step flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor affine apply flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor linalg generic indexing maps iterator types ins tensor tensor tensor tensor tensor outs tensor tensor attrs lowering config config in in in in in out out arith addf in in arith addf in arith mulf in arith addf in linalg yield tensor tensor scf for to step iter args tensor tensor affine apply scf for to step iter args tensor tensor affine apply extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor linalg generic indexing maps iterator types ins extracted slice extracted slice extracted slice extracted slice extracted slice tensor tensor tensor tensor tensor outs extracted slice extracted slice tensor tensor attrs lowering config config in in in in in out out arith addf in in arith addf in arith mulf in arith addf in linalg yield tensor tensor extracted slice tensor extract slice tensor to tensor pack tensor pack inner dims pos inner tiles into extracted slice internal linalg transform lowering config config tensor tensor inserted slice tensor insert slice pack into tensor into tensor inserted slice tensor insert slice into tensor into tensor scf yield inserted slice inserted slice tensor tensor scf yield tensor tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor return what i m expecting is that the is removed and the scf for loop should return three values this drops the performance a lot | 0 |
390,964 | 26,876,255,619 | IssuesEvent | 2023-02-05 03:26:24 | RodrigoTeran/syntactically-awesome-react-app | https://api.github.com/repos/RodrigoTeran/syntactically-awesome-react-app | closed | Add strict node and git dependencies to use create-sara-project | documentation priority:low | # β³οΈ Add strict node and git dependencies to use create-sara-project
Dependencies:
- ``` npm@8.9.0 ```
- ``` node@v16.17.1 ```
- ``` git@2.28.0 ```
Add them to the [Readme.md](https://github.com/RodrigoTeran/syntactically-awesome-react-app#readme) | 1.0 | Add strict node and git dependencies to use create-sara-project - # β³οΈ Add strict node and git dependencies to use create-sara-project
Dependencies:
- ``` npm@8.9.0 ```
- ``` node@v16.17.1 ```
- ``` git@2.28.0 ```
Add them to the [Readme.md](https://github.com/RodrigoTeran/syntactically-awesome-react-app#readme) | non_main | add strict node and git dependencies to use create sara project β³οΈ add strict node and git dependencies to use create sara project dependencies npm node git add them to the | 0 |
1,387 | 6,015,202,230 | IssuesEvent | 2017-06-07 00:57:20 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_robocopy module don't handle all posible exit codes | affects_2.2 bug_report waiting_on_maintainer windows | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_robocopy module
##### ANSIBLE VERSION
<!--- Paste verbatim output from βansible --versionβ between quotes below -->
```
2.2.0.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say βN/Aβ for anything that is not platform-specific.
-->
##### SUMMARY
<!--- Explain the problem briefly -->
module handle only 0,1,2,4,8,16 robocopy exit codes
but if custom flags used exit codes may eq combined value:
These can be combined, giving a few extra exit codes:
0Γ03 3 (2+1) Some files were copied. Additional files were present. No failure was encountered.
0Γ05 5 (4+1) Some files were copied. Some files were mismatched. No failure was encountered.
0Γ06 6 (4+2) Additional files and mismatched files exist. No files were copied and no failures were encountered.
This means that the files already exist in the destination directory
0Γ07 7 (4+1+2) Files were copied, a file mismatch was present, and additional files were present.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| True | win_robocopy module don't handle all posible exit codes - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_robocopy module
##### ANSIBLE VERSION
<!--- Paste verbatim output from βansible --versionβ between quotes below -->
```
2.2.0.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say βN/Aβ for anything that is not platform-specific.
-->
##### SUMMARY
<!--- Explain the problem briefly -->
module handle only 0,1,2,4,8,16 robocopy exit codes
but if custom flags used exit codes may eq combined value:
These can be combined, giving a few extra exit codes:
0Γ03 3 (2+1) Some files were copied. Additional files were present. No failure was encountered.
0Γ05 5 (4+1) Some files were copied. Some files were mismatched. No failure was encountered.
0Γ06 6 (4+2) Additional files and mismatched files exist. No files were copied and no failures were encountered.
This means that the files already exist in the destination directory
0Γ07 7 (4+1+2) Files were copied, a file mismatch was present, and additional files were present.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| main | win robocopy module don t handle all posible exit codes issue type bug report component name win robocopy module ansible version configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say βn aβ for anything that is not platform specific summary module handle only robocopy exit codes but if custom flags used exit codes may eq combined value these can be combined giving a few extra exit codes Γ some files were copied additional files were present no failure was encountered Γ some files were copied some files were mismatched no failure was encountered Γ additional files and mismatched files exist no files were copied and no failures were encountered this means that the files already exist in the destination directory Γ files were copied a file mismatch was present and additional files were present steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used expected results actual results | 1 |
77,948 | 22,049,719,371 | IssuesEvent | 2022-05-30 07:31:14 | pulumi/pulumi | https://api.github.com/repos/pulumi/pulumi | opened | Code coverage error should not fail a build | area/build kind/engineering | Failures like this one: https://github.com/pulumi/pulumi/runs/6614209710?check_suite_focus=true#step:40:25

should not fail the whole build. | 1.0 | Code coverage error should not fail a build - Failures like this one: https://github.com/pulumi/pulumi/runs/6614209710?check_suite_focus=true#step:40:25

should not fail the whole build. | non_main | code coverage error should not fail a build failures like this one should not fail the whole build | 0 |
4,491 | 23,387,863,868 | IssuesEvent | 2022-08-11 15:07:52 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Improve BreadcrumbSelector component | type: enhancement work: frontend status: ready restricted: new maintainers | ## Current behavior
- This is the `BreadcrumbSelector` component:

- It is used both for selecting a Schema within the current Database and for selecting a Table or Exploration within the current Schema.
## Desired behavior
- It should look more like this:

Specifically...
- A search input should exist and be focused when the BreadcrumbSelector opens. Search queries should filter the entries across all categories.
- Entries should highlight the substring of their label which matches the search query.
- (Not shown in the mockup) An "Add New" option should appear within each category.
To implement this, I'd recommend starting by changing...
```ts
export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorEntry[]>;
```
to...
```ts
interface BreadcrumbSelectorAddNewViaLink {
type: 'link';
href: string;
}
interface BreadcrumbSelectorAddNewViaButton {
type: 'button';
onSubmit: () => Promise<void>;
}
interface BreadcrumbSelectorSection {
entries: BreadcrumbSelectorEntry[];
addNew?: BreadcrumbSelectorAddNewViaLink | BreadcrumbSelectorAddNewViaButton;
}
export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorSection>;
```
- If the URL for the entry matches _the start_ of the router's current URL, then the entry should visually indicate that it's active. (It's important to match the start because we want to show the active schema when we're on the Table Page, for example.)
- Vertical scrolling should not happen so easily. We'll need to increase `max-height` somewhere. A good value might be something like `calc(100vh - 5em)`.
| True | Improve BreadcrumbSelector component - ## Current behavior
- This is the `BreadcrumbSelector` component:

- It is used both for selecting a Schema within the current Database and for selecting a Table or Exploration within the current Schema.
## Desired behavior
- It should look more like this:

Specifically...
- A search input should exist and be focused when the BreadcrumbSelector opens. Search queries should filter the entries across all categories.
- Entries should highlight the substring of their label which matches the search query.
- (Not shown in the mockup) An "Add New" option should appear within each category.
To implement this, I'd recommend starting by changing...
```ts
export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorEntry[]>;
```
to...
```ts
interface BreadcrumbSelectorAddNewViaLink {
type: 'link';
href: string;
}
interface BreadcrumbSelectorAddNewViaButton {
type: 'button';
onSubmit: () => Promise<void>;
}
interface BreadcrumbSelectorSection {
entries: BreadcrumbSelectorEntry[];
addNew?: BreadcrumbSelectorAddNewViaLink | BreadcrumbSelectorAddNewViaButton;
}
export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorSection>;
```
- If the URL for the entry matches _the start_ of the router's current URL, then the entry should visually indicate that it's active. (It's important to match the start because we want to show the active schema when we're on the Table Page, for example.)
- Vertical scrolling should not happen so easily. We'll need to increase `max-height` somewhere. A good value might be something like `calc(100vh - 5em)`.
| main | improve breadcrumbselector component current behavior this is the breadcrumbselector component it is used both for selecting a schema within the current database and for selecting a table or exploration within the current schema desired behavior it should look more like this specifically a search input should exist and be focused when the breadcrumbselector opens search queries should filter the entries across all categories entries should highlight the substring of their label which matches the search query not shown in the mockup an add new option should appear within each category to implement this i d recommend starting by changing ts export type breadcrumbselectordata map to ts interface breadcrumbselectoraddnewvialink type link href string interface breadcrumbselectoraddnewviabutton type button onsubmit promise interface breadcrumbselectorsection entries breadcrumbselectorentry addnew breadcrumbselectoraddnewvialink breadcrumbselectoraddnewviabutton export type breadcrumbselectordata map if the url for the entry matches the start of the router s current url then the entry should visually indicate that it s active it s important to match the start because we want to show the active schema when we re on the table page for example vertical scrolling should not happen so easily we ll need to increase max height somewhere a good value might be something like calc | 1 |
2,939 | 10,548,952,294 | IssuesEvent | 2019-10-03 07:30:41 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Huawei CloudEngine modules do not work with Ansible 2.8.5 | affects_2.8 bug module needs_maintainer needs_triage source_control support:community traceback |
##### SUMMARY
No ce_* module work with Ansible 2.8.5
I just try your example command :
`
ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=192.168.1.1 port=22 username=huawei password=huawei123" localhost --connection local
`
I have the same issue with all modules 'ce_'
The first time I was this error :
` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/ce.py", line 35, in <module>
from ansible.module_utils.network_common import to_list, ComplexList
ImportError: No module named network_common
`
I replaced in file
`from ansible.module_utils.network_common import to_list, ComplexList`
by
`from ansible.module_utils.network.common.utils import to_list, ComplexList`
Now I have an other error message :
` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__
raise AssertionError('socket_path must be a value')
AssertionError: socket_path must be a value
`
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ce_ modules not working in Ansible 2.8.5
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```
ansible 2.8.5
config file = /etc/ansible/P3/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/dist-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```
ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local
```
##### OS / ENVIRONMENT
Debian 10
##### STEPS TO REPRODUCE
```
ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local
```
##### EXPECTED RESULTS
Vlan List
##### ACTUAL RESULTS
Error Message : **socket_path must be a value**
```
ansible 2.8.5
config file = /etc/ansible/P3/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/dist-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0]
Using /etc/ansible/P3/ansible.cfg as config file
setting up inventory plugins
host_list declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method
script declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method
auto declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method
Parsed /etc/ansible/P3/hosts inventory source with ini plugin
Loading callback plugin minimal of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/minimal.pyc
META: ran handlers
<127.0.0.1> connection transport is cli
<10.10.10.4> using connection plugin network_cli
<10.10.10.4> socket_path: /root/.ansible/pc/cb0005b4c1
<10.10.10.4> exec_command(), socket_path=None
The full traceback is:
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 145, in run
res = self._execute()
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 664, in _execute
result = self._handler.run(task_vars=variables)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/ce.py", line 79, in run
rc, out, err = connection.exec_command('open_shell()')
File "/usr/lib/python2.7/dist-packages/ansible/plugins/connection/persistent.py", line 52, in exec_command
connection = SocketConnection(self.socket_path)
File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__
raise AssertionError('socket_path must be a value')
AssertionError: socket_path must be a value
localhost | FAILED! => {
"msg": "Unexpected failure during module execution.",
"stdout": ""
}
```
| True | Huawei CloudEngine modules do not work with Ansible 2.8.5 -
##### SUMMARY
No ce_* module work with Ansible 2.8.5
I just try your example command :
`
ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=192.168.1.1 port=22 username=huawei password=huawei123" localhost --connection local
`
I have the same issue with all modules 'ce_'
The first time I was this error :
` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/ce.py", line 35, in <module>
from ansible.module_utils.network_common import to_list, ComplexList
ImportError: No module named network_common
`
I replaced in file
`from ansible.module_utils.network_common import to_list, ComplexList`
by
`from ansible.module_utils.network.common.utils import to_list, ComplexList`
Now I have an other error message :
` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__
raise AssertionError('socket_path must be a value')
AssertionError: socket_path must be a value
`
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ce_ modules not working in Ansible 2.8.5
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```
ansible 2.8.5
config file = /etc/ansible/P3/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/dist-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0]
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```
ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local
```
##### OS / ENVIRONMENT
Debian 10
##### STEPS TO REPRODUCE
```
ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local
```
##### EXPECTED RESULTS
Vlan List
##### ACTUAL RESULTS
Error Message : **socket_path must be a value**
```
ansible 2.8.5
config file = /etc/ansible/P3/ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/dist-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0]
Using /etc/ansible/P3/ansible.cfg as config file
setting up inventory plugins
host_list declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method
script declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method
auto declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method
Parsed /etc/ansible/P3/hosts inventory source with ini plugin
Loading callback plugin minimal of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/minimal.pyc
META: ran handlers
<127.0.0.1> connection transport is cli
<10.10.10.4> using connection plugin network_cli
<10.10.10.4> socket_path: /root/.ansible/pc/cb0005b4c1
<10.10.10.4> exec_command(), socket_path=None
The full traceback is:
Traceback (most recent call last):
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 145, in run
res = self._execute()
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 664, in _execute
result = self._handler.run(task_vars=variables)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/ce.py", line 79, in run
rc, out, err = connection.exec_command('open_shell()')
File "/usr/lib/python2.7/dist-packages/ansible/plugins/connection/persistent.py", line 52, in exec_command
connection = SocketConnection(self.socket_path)
File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__
raise AssertionError('socket_path must be a value')
AssertionError: socket_path must be a value
localhost | FAILED! => {
"msg": "Unexpected failure during module execution.",
"stdout": ""
}
```
| main | huawei cloudengine modules do not work with ansible summary no ce module work with ansible i just try your example command ansible m ce command a commands display vlan summary transport cli host port username huawei password localhost connection local i have the same issue with all modules ce the first time i was this error file usr lib dist packages ansible module utils ce py line in from ansible module utils network common import to list complexlist importerror no module named network common i replaced in file from ansible module utils network common import to list complexlist by from ansible module utils network common utils import to list complexlist now i have an other error message file usr lib dist packages ansible module utils connection py line in init raise assertionerror socket path must be a value assertionerror socket path must be a value issue type bug report component name ce modules not working in ansible ansible version ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib dist packages ansible executable location usr bin ansible python version default apr configuration ansible m ce command a commands display vlan summary transport cli host port username admin password admin localhost connection local os environment debian steps to reproduce ansible m ce command a commands display vlan summary transport cli host port username admin password admin localhost connection local expected results vlan list actual results error message socket path must be a value ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib dist packages ansible executable location usr bin ansible python version default apr using etc ansible ansible cfg as config file setting up inventory plugins host list declined parsing etc ansible hosts as it did not pass it s verify file method script declined parsing etc ansible hosts as it did not pass it s verify file method auto declined parsing etc ansible hosts as it did not pass it s verify file method parsed etc ansible hosts inventory source with ini plugin loading callback plugin minimal of type stdout from usr lib dist packages ansible plugins callback minimal pyc meta ran handlers connection transport is cli using connection plugin network cli socket path root ansible pc exec command socket path none the full traceback is traceback most recent call last file usr lib dist packages ansible executor task executor py line in run res self execute file usr lib dist packages ansible executor task executor py line in execute result self handler run task vars variables file usr lib dist packages ansible plugins action ce py line in run rc out err connection exec command open shell file usr lib dist packages ansible plugins connection persistent py line in exec command connection socketconnection self socket path file usr lib dist packages ansible module utils connection py line in init raise assertionerror socket path must be a value assertionerror socket path must be a value localhost failed msg unexpected failure during module execution stdout | 1 |
146,397 | 19,403,573,001 | IssuesEvent | 2021-12-19 16:08:14 | victorlmneves/fed-pug-boilerplate-v2 | https://api.github.com/repos/victorlmneves/fed-pug-boilerplate-v2 | closed | CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz - autoclosed | security vulnerability | ## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: fed-pug-boilerplate-v2/package.json</p>
<p>Path to vulnerable library: fed-pug-boilerplate-v2/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-middleware-0.11.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/victorlmneves/fed-pug-boilerplate-v2/commit/473ea3597a89ac9b7c4f4d251f4b4c119b4643eb">473ea3597a89ac9b7c4f4d251f4b4c119b4643eb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz - autoclosed - ## CVE-2018-20821 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: fed-pug-boilerplate-v2/package.json</p>
<p>Path to vulnerable library: fed-pug-boilerplate-v2/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-middleware-0.11.0.tgz (Root Library)
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/victorlmneves/fed-pug-boilerplate-v2/commit/473ea3597a89ac9b7c4f4d251f4b4c119b4643eb">473ea3597a89ac9b7c4f4d251f4b4c119b4643eb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p>
<p>Release Date: 2019-04-23</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in opennmsopennms source node sass tgz autoclosed cve medium severity vulnerability vulnerable libraries opennmsopennms source node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file fed pug boilerplate package json path to vulnerable library fed pug boilerplate node modules node sass package json dependency hierarchy node sass middleware tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details the parsing component in libsass through allows attackers to cause a denial of service uncontrolled recursion in sass parser parse css variable value in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
9,812 | 3,321,781,139 | IssuesEvent | 2015-11-09 10:53:55 | interactivethings/catalog | https://api.github.com/repos/interactivethings/catalog | opened | Specimen Documentation | documentation | All specimen documentation pages should follow the same structure:
- A brief description
- Markdown API (options & JSON configuration)
- React API (don't worry right now)
- Examples with copy-n-pasteable code snippets | 1.0 | Specimen Documentation - All specimen documentation pages should follow the same structure:
- A brief description
- Markdown API (options & JSON configuration)
- React API (don't worry right now)
- Examples with copy-n-pasteable code snippets | non_main | specimen documentation all specimen documentation pages should follow the same structure a brief description markdown api options json configuration react api don t worry right now examples with copy n pasteable code snippets | 0 |
374,108 | 11,072,081,075 | IssuesEvent | 2019-12-12 09:36:01 | aspnetboilerplate/aspnetboilerplate | https://api.github.com/repos/aspnetboilerplate/aspnetboilerplate | closed | Does UnitTest Project Support EF Core Lazy Loading? | priority:high problem | I want to let Unit Test project support lazy loading feature which introduced in EF Core 2.1.
Ref : https://docs.microsoft.com/en-us/ef/core/querying/related-data#lazy-loading
So, I Performed following steps try to reach the goal:
**Step1.** Install-Package Microsoft.EntityFrameworkCore.Proxies on Unit Test Project.
**Step2.** Add services.AddEntityFrameworkProxies(); statement in function Register(IIocManager iocManager) of ServiceCollectionRegistrar.cs
**Step3.** In same file, add use lazy loading proxy like this : builder.UseLazyLoadingProxies().UseInMemoryDatabase(Guid.NewGuid().ToString()).UseInternalServiceProvider(serviceProvider);
ABP package version: 4.0.1
Base on .Net Core
When I ran the test case, I got this exception message:
Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method.
Any idea or suggestion about this problemοΌ Thanks in advanceοΌ
PS:
````
Result StackTrace:
at Castle.MicroKernel.ComponentActivator.FactoryMethodActivator`1.Instantiate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated)
at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy)
at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21
at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden)
at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired)
at Castle.MicroKernel.Resolvers.DefaultDependencyResolver.Resolve(CreationContext context, ISubDependencyResolver contextHandlerResolver, ComponentModel model, DependencyModel dependency)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateConstructorArguments(ConstructorCandidate constructor, CreationContext context)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.Instantiate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated)
at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy)
at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21
at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden)
at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired)
at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy)
at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.ResolveAll(Type service, IDictionary arguments, IReleasePolicy policy)
at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.ResolveInstanceOrNull(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 86
at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.GetServiceInternal(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 55
at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetService[T](IServiceProvider provider)
at Microsoft.EntityFrameworkCore.Proxies.Internal.ProxiesOptionsExtension.Validate(IDbContextOptions options)
at Microsoft.EntityFrameworkCore.Internal.ServiceProviderCache.GetOrAdd(IDbContextOptions options, Boolean providerRequired)
at Microsoft.EntityFrameworkCore.DbContext..ctor(DbContextOptions options)
at Abp.Zero.EntityFrameworkCore.AbpZeroDbContext`4..ctor(DbContextOptions`1 options) in D:\Github\aspnetboilerplate\src\Abp.ZeroCore.EntityFrameworkCore\Zero\EntityFrameworkCore\AbpZeroDbContext.cs:line 68
at AbpTest.EntityFrameworkCore.AbpTestDbContext..ctor(DbContextOptions`1 options) in C:\Workspace\Lab\4.0.1\aspnet-core\src\AbpTest.EntityFrameworkCore\EntityFrameworkCore\AbpTestDbContext.cs:line 14
at lambda_method(Closure , Object[] )
at Castle.Core.Internal.ReflectionUtil.Instantiate[TBase](Type subtypeofTBase, Object[] ctorArgs)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstanceCore(ConstructorCandidate constructor, Object[] arguments, Type implType)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstance(CreationContext context, ConstructorCandidate constructor, Object[] arguments)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy)
at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden)
at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired)
at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy)
at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.Resolve(Type service, IDictionary arguments, IReleasePolicy policy)
at Castle.Windsor.WindsorContainer.Resolve[T]()
at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Nullable`1 tenantId, Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 84
at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 62
at AbpTest.Tests.AbpTestTestBase..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 33
at AbpTest.Tests.Sessions.SessionAppService_Tests..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\Sessions\SessionAppService_Tests.cs:line 12
Result Message: Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method.
````
| 1.0 | Does UnitTest Project Support EF Core Lazy Loading? - I want to let Unit Test project support lazy loading feature which introduced in EF Core 2.1.
Ref : https://docs.microsoft.com/en-us/ef/core/querying/related-data#lazy-loading
So, I Performed following steps try to reach the goal:
**Step1.** Install-Package Microsoft.EntityFrameworkCore.Proxies on Unit Test Project.
**Step2.** Add services.AddEntityFrameworkProxies(); statement in function Register(IIocManager iocManager) of ServiceCollectionRegistrar.cs
**Step3.** In same file, add use lazy loading proxy like this : builder.UseLazyLoadingProxies().UseInMemoryDatabase(Guid.NewGuid().ToString()).UseInternalServiceProvider(serviceProvider);
ABP package version: 4.0.1
Base on .Net Core
When I ran the test case, I got this exception message:
Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method.
Any idea or suggestion about this problemοΌ Thanks in advanceοΌ
PS:
````
Result StackTrace:
at Castle.MicroKernel.ComponentActivator.FactoryMethodActivator`1.Instantiate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated)
at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy)
at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21
at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden)
at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired)
at Castle.MicroKernel.Resolvers.DefaultDependencyResolver.Resolve(CreationContext context, ISubDependencyResolver contextHandlerResolver, ComponentModel model, DependencyModel dependency)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateConstructorArguments(ConstructorCandidate constructor, CreationContext context)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.Instantiate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated)
at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance)
at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy)
at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21
at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden)
at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired)
at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy)
at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.ResolveAll(Type service, IDictionary arguments, IReleasePolicy policy)
at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.ResolveInstanceOrNull(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 86
at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.GetServiceInternal(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 55
at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetService[T](IServiceProvider provider)
at Microsoft.EntityFrameworkCore.Proxies.Internal.ProxiesOptionsExtension.Validate(IDbContextOptions options)
at Microsoft.EntityFrameworkCore.Internal.ServiceProviderCache.GetOrAdd(IDbContextOptions options, Boolean providerRequired)
at Microsoft.EntityFrameworkCore.DbContext..ctor(DbContextOptions options)
at Abp.Zero.EntityFrameworkCore.AbpZeroDbContext`4..ctor(DbContextOptions`1 options) in D:\Github\aspnetboilerplate\src\Abp.ZeroCore.EntityFrameworkCore\Zero\EntityFrameworkCore\AbpZeroDbContext.cs:line 68
at AbpTest.EntityFrameworkCore.AbpTestDbContext..ctor(DbContextOptions`1 options) in C:\Workspace\Lab\4.0.1\aspnet-core\src\AbpTest.EntityFrameworkCore\EntityFrameworkCore\AbpTestDbContext.cs:line 14
at lambda_method(Closure , Object[] )
at Castle.Core.Internal.ReflectionUtil.Instantiate[TBase](Type subtypeofTBase, Object[] ctorArgs)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstanceCore(ConstructorCandidate constructor, Object[] arguments, Type implType)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstance(CreationContext context, ConstructorCandidate constructor, Object[] arguments)
at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context)
at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally)
at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy)
at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden)
at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired)
at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy)
at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.Resolve(Type service, IDictionary arguments, IReleasePolicy policy)
at Castle.Windsor.WindsorContainer.Resolve[T]()
at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Nullable`1 tenantId, Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 84
at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 62
at AbpTest.Tests.AbpTestTestBase..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 33
at AbpTest.Tests.Sessions.SessionAppService_Tests..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\Sessions\SessionAppService_Tests.cs:line 12
Result Message: Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method.
````
| non_main | does unittest project support ef core lazy loading i want to let unit test project support lazy loading feature which introduced in ef core ref so i performed following steps try to reach the goal install package microsoft entityframeworkcore proxies on unit test project add services addentityframeworkproxies statement in function register iiocmanager iocmanager of servicecollectionregistrar cs in same file add use lazy loading proxy like this builder uselazyloadingproxies useinmemorydatabase guid newguid tostring useinternalserviceprovider serviceprovider abp package version base on net core when i ran the test case i got this exception message castle microkernel componentactivator componentactivatorexception factory method creating instances of component microsoft entityframeworkcore infrastructure idbcontextoptions returned null this is not allowed and most likely a bug in the factory method any idea or suggestion about this problemοΌ thanks in advanceοΌ ps result stacktrace at castle microkernel componentactivator factorymethodactivator instantiate creationcontext context at castle microkernel componentactivator defaultcomponentactivator internalcreate creationcontext context at castle microkernel componentactivator abstractcomponentactivator create creationcontext context burden burden at castle microkernel lifestyle abstractlifestylemanager createinstance creationcontext context boolean trackedexternally at castle microkernel lifestyle scopedlifestylemanager c b action aftercreated at castle microkernel lifestyle scoped defaultlifetimescope getcachedinstance componentmodel model scopedinstanceactivationcallback createinstance at castle microkernel lifestyle scopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy at castle windsor msdependencyinjection msscopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy in d github castle windsor ms adapter src castle windsor msdependencyinjection msscopedlifestylemanager cs line at castle microkernel handlers defaulthandler resolvecore creationcontext context boolean requiresdecommission boolean instancerequired burden burden at castle microkernel handlers defaulthandler resolve creationcontext context boolean instancerequired at castle microkernel resolvers defaultdependencyresolver resolve creationcontext context isubdependencyresolver contexthandlerresolver componentmodel model dependencymodel dependency at castle microkernel componentactivator defaultcomponentactivator createconstructorarguments constructorcandidate constructor creationcontext context at castle microkernel componentactivator defaultcomponentactivator instantiate creationcontext context at castle microkernel componentactivator defaultcomponentactivator internalcreate creationcontext context at castle microkernel componentactivator abstractcomponentactivator create creationcontext context burden burden at castle microkernel lifestyle abstractlifestylemanager createinstance creationcontext context boolean trackedexternally at castle microkernel lifestyle scopedlifestylemanager c b action aftercreated at castle microkernel lifestyle scoped defaultlifetimescope getcachedinstance componentmodel model scopedinstanceactivationcallback createinstance at castle microkernel lifestyle scopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy at castle windsor msdependencyinjection msscopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy in d github castle windsor ms adapter src castle windsor msdependencyinjection msscopedlifestylemanager cs line at castle microkernel handlers defaulthandler resolvecore creationcontext context boolean requiresdecommission boolean instancerequired burden burden at castle microkernel handlers defaulthandler resolve creationcontext context boolean instancerequired at castle microkernel defaultkernel resolvecomponent ihandler handler type service idictionary additionalarguments ireleasepolicy policy at castle microkernel defaultkernel castle microkernel ikernelinternal resolveall type service idictionary arguments ireleasepolicy policy at castle windsor msdependencyinjection scopedwindsorserviceprovider resolveinstanceornull type servicetype boolean isoptional in d github castle windsor ms adapter src castle windsor msdependencyinjection scopedwindsorserviceprovider cs line at castle windsor msdependencyinjection scopedwindsorserviceprovider getserviceinternal type servicetype boolean isoptional in d github castle windsor ms adapter src castle windsor msdependencyinjection scopedwindsorserviceprovider cs line at microsoft extensions dependencyinjection serviceproviderserviceextensions getservice iserviceprovider provider at microsoft entityframeworkcore proxies internal proxiesoptionsextension validate idbcontextoptions options at microsoft entityframeworkcore internal serviceprovidercache getoradd idbcontextoptions options boolean providerrequired at microsoft entityframeworkcore dbcontext ctor dbcontextoptions options at abp zero entityframeworkcore abpzerodbcontext ctor dbcontextoptions options in d github aspnetboilerplate src abp zerocore entityframeworkcore zero entityframeworkcore abpzerodbcontext cs line at abptest entityframeworkcore abptestdbcontext ctor dbcontextoptions options in c workspace lab aspnet core src abptest entityframeworkcore entityframeworkcore abptestdbcontext cs line at lambda method closure object at castle core internal reflectionutil instantiate type subtypeoftbase object ctorargs at castle microkernel componentactivator defaultcomponentactivator createinstancecore constructorcandidate constructor object arguments type impltype at castle microkernel componentactivator defaultcomponentactivator createinstance creationcontext context constructorcandidate constructor object arguments at castle microkernel componentactivator defaultcomponentactivator internalcreate creationcontext context at castle microkernel componentactivator abstractcomponentactivator create creationcontext context burden burden at castle microkernel lifestyle abstractlifestylemanager createinstance creationcontext context boolean trackedexternally at castle microkernel lifestyle abstractlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy at castle microkernel handlers defaulthandler resolvecore creationcontext context boolean requiresdecommission boolean instancerequired burden burden at castle microkernel handlers defaulthandler resolve creationcontext context boolean instancerequired at castle microkernel defaultkernel resolvecomponent ihandler handler type service idictionary additionalarguments ireleasepolicy policy at castle microkernel defaultkernel castle microkernel ikernelinternal resolve type service idictionary arguments ireleasepolicy policy at castle windsor windsorcontainer resolve at abptest tests abptesttestbase usingdbcontext nullable tenantid action action in c workspace lab aspnet core test abptest tests abptesttestbase cs line at abptest tests abptesttestbase usingdbcontext action action in c workspace lab aspnet core test abptest tests abptesttestbase cs line at abptest tests abptesttestbase ctor in c workspace lab aspnet core test abptest tests abptesttestbase cs line at abptest tests sessions sessionappservice tests ctor in c workspace lab aspnet core test abptest tests sessions sessionappservice tests cs line result message castle microkernel componentactivator componentactivatorexception factory method creating instances of component microsoft entityframeworkcore infrastructure idbcontextoptions returned null this is not allowed and most likely a bug in the factory method | 0 |
773,550 | 27,161,536,993 | IssuesEvent | 2023-02-17 12:17:18 | JamieMason/syncpack | https://api.github.com/repos/JamieMason/syncpack | closed | Output semver/version groups in the order they're defined | Priority: Medium Type: Feat good first issue | ## Description
Every command that lists output [has a `.reverse()` call](https://github.com/JamieMason/syncpack/blob/d397ea7d986815e3b5a79bcb017db83670839661/src/bin-list/list.ts#L8) β this is a hangover from a previous implementation before I recently refactored the code β looping in reverse just made it easier because it meant the first item could be known to be the default group.
This makes it harder to debug because the groups don't align with the order they're defined in config.
## Suggested Solution
Remove the `.reverse()` calls from each command and update the expected numbers in the tests.
## Help Needed
This _should_ be an easy one, find and delete the calls, run `yarn jest --watch --no-coverage`, and fix the expected numbers in the log output in the failing tests. | 1.0 | Output semver/version groups in the order they're defined - ## Description
Every command that lists output [has a `.reverse()` call](https://github.com/JamieMason/syncpack/blob/d397ea7d986815e3b5a79bcb017db83670839661/src/bin-list/list.ts#L8) β this is a hangover from a previous implementation before I recently refactored the code β looping in reverse just made it easier because it meant the first item could be known to be the default group.
This makes it harder to debug because the groups don't align with the order they're defined in config.
## Suggested Solution
Remove the `.reverse()` calls from each command and update the expected numbers in the tests.
## Help Needed
This _should_ be an easy one, find and delete the calls, run `yarn jest --watch --no-coverage`, and fix the expected numbers in the log output in the failing tests. | non_main | output semver version groups in the order they re defined description every command that lists output β this is a hangover from a previous implementation before i recently refactored the code β looping in reverse just made it easier because it meant the first item could be known to be the default group this makes it harder to debug because the groups don t align with the order they re defined in config suggested solution remove the reverse calls from each command and update the expected numbers in the tests help needed this should be an easy one find and delete the calls run yarn jest watch no coverage and fix the expected numbers in the log output in the failing tests | 0 |
4,174 | 20,015,665,939 | IssuesEvent | 2022-02-01 11:48:11 | MetaCell/cloud-harness | https://api.github.com/repos/MetaCell/cloud-harness | closed | Add pull request template | maintainance | The pull request template shall include some self sanity checks to help the reviewer and the maintainer contextualize and evaluate the pull request | True | Add pull request template - The pull request template shall include some self sanity checks to help the reviewer and the maintainer contextualize and evaluate the pull request | main | add pull request template the pull request template shall include some self sanity checks to help the reviewer and the maintainer contextualize and evaluate the pull request | 1 |
116,590 | 9,856,618,180 | IssuesEvent | 2019-06-19 22:50:53 | livinglab/webwork-for-wordpress | https://api.github.com/repos/livinglab/webwork-for-wordpress | closed | Additional style adjustments | enhancement testing-needed | @boonebgorges, I've been able to identify a few additional style adjustments. I _think_ they are small and won't affect other elements but should tighten things up and get us a little closer to the original mockups. Feel free to push to "future" if there are cascade issues I'm not aware of.
.item-stats.problem-stats {
font-size: 1rem;
}
h3.ww-header {
font-size: 1.4rem;
}
.ww-ask-question-form label {
font-size: 1.4rem;
}
.ww-question-gloss {
font-size: 1.1rem;
}
.anonymous-toggle label {
line-height: 1.2rem;
font-size: .85rem;
}
.index-intro > p {
line-height: 2rem;
font-size: 1.4rem;
}
.ww-author-name {
font-size: 1.5rem;
} | 1.0 | Additional style adjustments - @boonebgorges, I've been able to identify a few additional style adjustments. I _think_ they are small and won't affect other elements but should tighten things up and get us a little closer to the original mockups. Feel free to push to "future" if there are cascade issues I'm not aware of.
.item-stats.problem-stats {
font-size: 1rem;
}
h3.ww-header {
font-size: 1.4rem;
}
.ww-ask-question-form label {
font-size: 1.4rem;
}
.ww-question-gloss {
font-size: 1.1rem;
}
.anonymous-toggle label {
line-height: 1.2rem;
font-size: .85rem;
}
.index-intro > p {
line-height: 2rem;
font-size: 1.4rem;
}
.ww-author-name {
font-size: 1.5rem;
} | non_main | additional style adjustments boonebgorges i ve been able to identify a few additional style adjustments i think they are small and won t affect other elements but should tighten things up and get us a little closer to the original mockups feel free to push to future if there are cascade issues i m not aware of item stats problem stats font size ww header font size ww ask question form label font size ww question gloss font size anonymous toggle label line height font size index intro p line height font size ww author name font size | 0 |
118,458 | 4,745,601,847 | IssuesEvent | 2016-10-21 08:02:42 | kubernetes/dashboard | https://api.github.com/repos/kubernetes/dashboard | closed | Pod state not correct | area/api kind/bug priority/P3 | #### Issue details
Pod state is not correct in UI
##### Environment
```
Dashboard version: master (12. Aug)
Kubernetes version: v1.4.0-alpha.2
```
##### Steps to reproduce
```
$ kubectl run test --image=debian -- echo hello
$ kubectl get pods
NAME READY STATUS RESTARTS AGE
test-2172490464-elqbi 0/1 CrashLoopBackOff 18 1h
```
##### Observed result
UI claims pod is in "Running" state
##### Expected result
"Crashed"
| 1.0 | Pod state not correct - #### Issue details
Pod state is not correct in UI
##### Environment
```
Dashboard version: master (12. Aug)
Kubernetes version: v1.4.0-alpha.2
```
##### Steps to reproduce
```
$ kubectl run test --image=debian -- echo hello
$ kubectl get pods
NAME READY STATUS RESTARTS AGE
test-2172490464-elqbi 0/1 CrashLoopBackOff 18 1h
```
##### Observed result
UI claims pod is in "Running" state
##### Expected result
"Crashed"
| non_main | pod state not correct issue details pod state is not correct in ui environment dashboard version master aug kubernetes version alpha steps to reproduce kubectl run test image debian echo hello kubectl get pods name ready status restarts age test elqbi crashloopbackoff observed result ui claims pod is in running state expected result crashed | 0 |
54,069 | 3,059,291,851 | IssuesEvent | 2015-08-14 14:16:03 | phetsims/tasks | https://api.github.com/repos/phetsims/tasks | closed | Test Bending Light Performance on iPad2 for a few scenarios | High Priority QA | Please test the performance for Bending Light in the scenarios given in: https://github.com/phetsims/bending-light/issues/153
There is a link there for testing and full details in the issue. Please report your results in https://github.com/phetsims/bending-light/issues/153
@ariel-phet please delegate. | 1.0 | Test Bending Light Performance on iPad2 for a few scenarios - Please test the performance for Bending Light in the scenarios given in: https://github.com/phetsims/bending-light/issues/153
There is a link there for testing and full details in the issue. Please report your results in https://github.com/phetsims/bending-light/issues/153
@ariel-phet please delegate. | non_main | test bending light performance on for a few scenarios please test the performance for bending light in the scenarios given in there is a link there for testing and full details in the issue please report your results in ariel phet please delegate | 0 |
1,174 | 5,096,305,393 | IssuesEvent | 2017-01-03 17:46:22 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | macports module: support for state=latest (upgrade) and for variants | affects_2.0 bug_report feature_idea waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
macports
##### ANSIBLE VERSION
```
$ ansible --version
ansible 2.0.2.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
Orchestrator: Ubuntu trusty
Target: Macos 10.11
##### SUMMARY
from http://docs.ansible.com/ansible/macports_module.html
no state=latest is supported while commands `port upgrade {{ port }}` and `port outdated | grep {{ port }}` allow to upgrade and check if upgrade available.
A `port sync` might be needed as equivalent of update_cache
Macports also allow to have different packages/variant for some compile/install options.
For example
```
port install lftp +ssl
```
a way to support it is needed. can have multiple +variant1 +variant2
##### STEPS TO REPRODUCE
```
- name: Darwin | macports upgrade
macports: name=* present=latest update_cache=yes
when: ansible_os_family == "Darwin"
- name: Darwin | variant install
macports: name=lftp variants=ssl,variant2
when: ansible_os_family == "Darwin"
```
##### EXPECTED RESULTS
Macports is upgraded
Macports has port installed with corresponding variants.
##### ACTUAL RESULTS
Above actions are currently not possible within module and must be executed as command losing idempotency.
| True | macports module: support for state=latest (upgrade) and for variants - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
macports
##### ANSIBLE VERSION
```
$ ansible --version
ansible 2.0.2.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
Orchestrator: Ubuntu trusty
Target: Macos 10.11
##### SUMMARY
from http://docs.ansible.com/ansible/macports_module.html
no state=latest is supported while commands `port upgrade {{ port }}` and `port outdated | grep {{ port }}` allow to upgrade and check if upgrade available.
A `port sync` might be needed as equivalent of update_cache
Macports also allow to have different packages/variant for some compile/install options.
For example
```
port install lftp +ssl
```
a way to support it is needed. can have multiple +variant1 +variant2
##### STEPS TO REPRODUCE
```
- name: Darwin | macports upgrade
macports: name=* present=latest update_cache=yes
when: ansible_os_family == "Darwin"
- name: Darwin | variant install
macports: name=lftp variants=ssl,variant2
when: ansible_os_family == "Darwin"
```
##### EXPECTED RESULTS
Macports is upgraded
Macports has port installed with corresponding variants.
##### ACTUAL RESULTS
Above actions are currently not possible within module and must be executed as command losing idempotency.
| main | macports module support for state latest upgrade and for variants issue type bug report component name macports ansible version ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration n a os environment orchestrator ubuntu trusty target macos summary from no state latest is supported while commands port upgrade port and port outdated grep port allow to upgrade and check if upgrade available a port sync might be needed as equivalent of update cache macports also allow to have different packages variant for some compile install options for example port install lftp ssl a way to support it is needed can have multiple steps to reproduce name darwin macports upgrade macports name present latest update cache yes when ansible os family darwin name darwin variant install macports name lftp variants ssl when ansible os family darwin expected results macports is upgraded macports has port installed with corresponding variants actual results above actions are currently not possible within module and must be executed as command losing idempotency | 1 |
2,786 | 9,985,230,884 | IssuesEvent | 2019-07-10 16:04:27 | DynamoRIO/dynamorio | https://api.github.com/repos/DynamoRIO/dynamorio | opened | Move duplicated CHECK defines in tests to client_tools.h | Maintainability good first issue | We have 15 different tests defining a CHECK macro. We should move all of those to a single define in client_tools.h. | True | Move duplicated CHECK defines in tests to client_tools.h - We have 15 different tests defining a CHECK macro. We should move all of those to a single define in client_tools.h. | main | move duplicated check defines in tests to client tools h we have different tests defining a check macro we should move all of those to a single define in client tools h | 1 |
152,868 | 5,871,404,869 | IssuesEvent | 2017-05-15 08:37:44 | PX4/Firmware | https://api.github.com/repos/PX4/Firmware | closed | Losing GPS does not trigger fail safe | bug priority-critical | On latest master, failsafe when losing GPS does not work. A few test cases:
* When I unclick `use GPS` for the EKF2_AID_MASK (in air), the only thing appearing is `WARN [navigator] global position timeout` after a few seconds but nothing happens. The quad just starts to drift.
* When I stop sending GPS (in air), nothing happens. The quad just starts to drift.
* When the home position was set but has no GPS anymore (on ground), I can still takeoff. Then the quad drifts
I guess the EKF reports local position as still valid? | 1.0 | Losing GPS does not trigger fail safe - On latest master, failsafe when losing GPS does not work. A few test cases:
* When I unclick `use GPS` for the EKF2_AID_MASK (in air), the only thing appearing is `WARN [navigator] global position timeout` after a few seconds but nothing happens. The quad just starts to drift.
* When I stop sending GPS (in air), nothing happens. The quad just starts to drift.
* When the home position was set but has no GPS anymore (on ground), I can still takeoff. Then the quad drifts
I guess the EKF reports local position as still valid? | non_main | losing gps does not trigger fail safe on latest master failsafe when losing gps does not work a few test cases when i unclick use gps for the aid mask in air the only thing appearing is warn global position timeout after a few seconds but nothing happens the quad just starts to drift when i stop sending gps in air nothing happens the quad just starts to drift when the home position was set but has no gps anymore on ground i can still takeoff then the quad drifts i guess the ekf reports local position as still valid | 0 |
4,737 | 24,456,719,413 | IssuesEvent | 2022-10-07 07:29:27 | NaluKit/nalu | https://api.github.com/repos/NaluKit/nalu | closed | Update plugin & dependency versions | maintainance | There are some newer versions of plugins and dependencies.
Especially to avoid the JUnit 5 deprecate warnings | True | Update plugin & dependency versions - There are some newer versions of plugins and dependencies.
Especially to avoid the JUnit 5 deprecate warnings | main | update plugin dependency versions there are some newer versions of plugins and dependencies especially to avoid the junit deprecate warnings | 1 |
435,154 | 30,488,378,132 | IssuesEvent | 2023-07-18 05:27:55 | hwchase17/langchain | https://api.github.com/repos/hwchase17/langchain | closed | DOC: SupabaseVectorStore.from_documents read operation timed out. | area: vector store auto:question auto:documentation | ### Issue with current documentation:
```
# We're using the default `documents` table here. You can modify this by passing in a `table_name` argument to the `from_documents` method.
vector_store = SupabaseVectorStore.from_documents(docs, embeddings, client=supabase)
```
### Idea or request for content:
throw error: httpx.ReadTimeout: The read operation timed out
Is it because the documents are too large? Is there a way to change the timeout? | 1.0 | DOC: SupabaseVectorStore.from_documents read operation timed out. - ### Issue with current documentation:
```
# We're using the default `documents` table here. You can modify this by passing in a `table_name` argument to the `from_documents` method.
vector_store = SupabaseVectorStore.from_documents(docs, embeddings, client=supabase)
```
### Idea or request for content:
throw error: httpx.ReadTimeout: The read operation timed out
Is it because the documents are too large? Is there a way to change the timeout? | non_main | doc supabasevectorstore from documents read operation timed out issue with current documentation we re using the default documents table here you can modify this by passing in a table name argument to the from documents method vector store supabasevectorstore from documents docs embeddings client supabase idea or request for content throw error httpx readtimeout the read operation timed out is it because the documents are too large is there a way to change the timeout | 0 |
637 | 4,155,107,931 | IssuesEvent | 2016-06-16 14:00:10 | duckduckgo/zeroclickinfo-goodies | https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies | closed | Notepad_plus_plus Cheat Sheet: | Maintainer Submitted | In producing this CS, I made the mistake of incorrectly specifying single key combinations with square brackets, e.g. "[F7]" instead of "F7"
------
IA Page: http://duck.co/ia/view/notepad_plus_plus_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @biscuitNinja | True | Notepad_plus_plus Cheat Sheet: - In producing this CS, I made the mistake of incorrectly specifying single key combinations with square brackets, e.g. "[F7]" instead of "F7"
------
IA Page: http://duck.co/ia/view/notepad_plus_plus_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @biscuitNinja | main | notepad plus plus cheat sheet in producing this cs i made the mistake of incorrectly specifying single key combinations with square brackets e g instead of ia page biscuitninja | 1 |
1,982 | 6,694,201,269 | IssuesEvent | 2017-10-10 00:16:50 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | closed | Astrobin Apod: not triggering | Maintainer Input Requested | the example query "astronomy picture of the day" isn't triggering for me.
---
IA Page: http://duck.co/ia/view/apod
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @killerfish
| True | Astrobin Apod: not triggering - the example query "astronomy picture of the day" isn't triggering for me.
---
IA Page: http://duck.co/ia/view/apod
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @killerfish
| main | astrobin apod not triggering the example query astronomy picture of the day isn t triggering for me ia page killerfish | 1 |
64,510 | 15,896,497,500 | IssuesEvent | 2021-04-11 17:38:32 | haskell/text | https://api.github.com/repos/haskell/text | closed | Can't build benchmarks | build failure | When trying to build benchmarks, I get a linker error, as the `cbits.c` from the `text` i have installed clash with the `cbits.c` from the `text` source code i'm currently working on and which i'm trying to benchmark.
```
Linking dist/build/text-benchmarks/text-benchmarks ...
/home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_memcmp':
(.text+0x20): multiple definition of `_hs_text_memcmp'
dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x0): first defined here
/home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8_state':
(.text+0xe0): multiple definition of `_hs_text_decode_utf8_state'
dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x20): first defined here
/home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8':
(.text+0x2a0): multiple definition of `_hs_text_decode_utf8'
dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x150): first defined here
collect2: error: ld returned 1 exit status
```
The problem is, that `criterion` has `text` as a dependency.
Currently, as a workaround, I just rename the functions in `cbits.c`, then `cabal build` just works.
* Is there a better solution?
* If not, I can send a PR adding a compilation flag and do the renaming, so `cabal build` just works.
* Is this related to https://github.com/haskell/cabal/issues/1575 ?
| 1.0 | Can't build benchmarks - When trying to build benchmarks, I get a linker error, as the `cbits.c` from the `text` i have installed clash with the `cbits.c` from the `text` source code i'm currently working on and which i'm trying to benchmark.
```
Linking dist/build/text-benchmarks/text-benchmarks ...
/home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_memcmp':
(.text+0x20): multiple definition of `_hs_text_memcmp'
dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x0): first defined here
/home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8_state':
(.text+0xe0): multiple definition of `_hs_text_decode_utf8_state'
dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x20): first defined here
/home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8':
(.text+0x2a0): multiple definition of `_hs_text_decode_utf8'
dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x150): first defined here
collect2: error: ld returned 1 exit status
```
The problem is, that `criterion` has `text` as a dependency.
Currently, as a workaround, I just rename the functions in `cbits.c`, then `cabal build` just works.
* Is there a better solution?
* If not, I can send a PR adding a compilation flag and do the renaming, so `cabal build` just works.
* Is this related to https://github.com/haskell/cabal/issues/1575 ?
| non_main | can t build benchmarks when trying to build benchmarks i get a linker error as the cbits c from the text i have installed clash with the cbits c from the text source code i m currently working on and which i m trying to benchmark linking dist build text benchmarks text benchmarks home kuko cabal lib linux ghc text libhstext a cbits o in function hs text memcmp text multiple definition of hs text memcmp dist build text benchmarks text benchmarks tmp cbits cbits o cbits c text first defined here home kuko cabal lib linux ghc text libhstext a cbits o in function hs text decode state text multiple definition of hs text decode state dist build text benchmarks text benchmarks tmp cbits cbits o cbits c text first defined here home kuko cabal lib linux ghc text libhstext a cbits o in function hs text decode text multiple definition of hs text decode dist build text benchmarks text benchmarks tmp cbits cbits o cbits c text first defined here error ld returned exit status the problem is that criterion has text as a dependency currently as a workaround i just rename the functions in cbits c then cabal build just works is there a better solution if not i can send a pr adding a compilation flag and do the renaming so cabal build just works is this related to | 0 |
94,781 | 10,854,066,752 | IssuesEvent | 2019-11-13 15:47:51 | dotnet/reactive | https://api.github.com/repos/dotnet/reactive | closed | Make the Documentation available on the Microsoft Docs website | documentation | Rx is an extremely powerful (and in many cases, essential) framework to build complex event-driven / reactive applications or to deal with complex asynchronous workflows. As such an important library, I think it should definitely be available on the [Microsoft Docs](https://docs.microsoft.com) website.
There is a documentation available on [MSDN](https://msdn.microsoft.com/en-us/library/hh242985(v=vs.103).aspx) with API Reference, but I think might be outdated.
It could have its place somewhere alongside LINQ in the [.NET Guide](https://docs.microsoft.com/en-us/dotnet/standard/index) section, or under the [Asynchronous Programming Patterns](https://docs.microsoft.com/en-us/dotnet/standard/asynchronous-programming-patterns/) section, maybe? | 1.0 | Make the Documentation available on the Microsoft Docs website - Rx is an extremely powerful (and in many cases, essential) framework to build complex event-driven / reactive applications or to deal with complex asynchronous workflows. As such an important library, I think it should definitely be available on the [Microsoft Docs](https://docs.microsoft.com) website.
There is a documentation available on [MSDN](https://msdn.microsoft.com/en-us/library/hh242985(v=vs.103).aspx) with API Reference, but I think might be outdated.
It could have its place somewhere alongside LINQ in the [.NET Guide](https://docs.microsoft.com/en-us/dotnet/standard/index) section, or under the [Asynchronous Programming Patterns](https://docs.microsoft.com/en-us/dotnet/standard/asynchronous-programming-patterns/) section, maybe? | non_main | make the documentation available on the microsoft docs website rx is an extremely powerful and in many cases essential framework to build complex event driven reactive applications or to deal with complex asynchronous workflows as such an important library i think it should definitely be available on the website there is a documentation available on with api reference but i think might be outdated it could have its place somewhere alongside linq in the section or under the section maybe | 0 |
3,157 | 12,198,047,592 | IssuesEvent | 2020-04-29 21:59:12 | short-d/short | https://api.github.com/repos/short-d/short | closed | [Refactor] Move envconfig into app framework | maintainability | Current [envconfig](https://github.com/short-d/short/tree/master/backend/envconfig) in located in Short's repo. However, it's reusable across all microservices.
Define `EnvConfig` interface in [fw](https://github.com/short-d/app/tree/master/fw) and create implementation here: https://github.com/short-d/app/tree/master/modern | True | [Refactor] Move envconfig into app framework - Current [envconfig](https://github.com/short-d/short/tree/master/backend/envconfig) in located in Short's repo. However, it's reusable across all microservices.
Define `EnvConfig` interface in [fw](https://github.com/short-d/app/tree/master/fw) and create implementation here: https://github.com/short-d/app/tree/master/modern | main | move envconfig into app framework current in located in short s repo however it s reusable across all microservices define envconfig interface in and create implementation here | 1 |
3,673 | 15,036,008,410 | IssuesEvent | 2021-02-02 14:47:15 | IITIDIDX597/sp_2021_team1 | https://api.github.com/repos/IITIDIDX597/sp_2021_team1 | opened | Sorting information in categories | Epic: 5 Maintaining the system Story Week 3 | **Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care.
**Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform.
**Sub-Hill Statements:**
1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments.
### **Story Details:**
As an: administrator
I want: to be able to add information in their respective categories
So that: I can keep the system updated
| True | Sorting information in categories - **Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care.
**Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform.
**Sub-Hill Statements:**
1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments.
### **Story Details:**
As an: administrator
I want: to be able to add information in their respective categories
So that: I can keep the system updated
| main | sorting information in categories project goal s lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way while at the same time foster deeper learning experiences in order to deliver better abilitylab patient care hill statement individual clinicians can reference relevant continuously evolving information for their patient s therapy needs to self manage their approach patient care plan development in a single platform sub hill statements the learning platform will be routinely updated with s lab s own research advancements as well as outside discoveries and best practices developed for rehabilitation treatments story details as an administrator i want to be able to add information in their respective categories so that i can keep the system updated | 1 |
2,816 | 10,102,930,772 | IssuesEvent | 2019-07-29 12:29:32 | luckyariane/arthas-bot | https://api.github.com/repos/luckyariane/arthas-bot | closed | Fix recent subscriber parsing. | bug maintain | The loss of twitch alert local file creation means I need to find a new source for this data. | True | Fix recent subscriber parsing. - The loss of twitch alert local file creation means I need to find a new source for this data. | main | fix recent subscriber parsing the loss of twitch alert local file creation means i need to find a new source for this data | 1 |
4,607 | 23,855,859,644 | IssuesEvent | 2022-09-06 23:24:56 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [a11y]: Violations found on UI Shell w/ SideNav component in v10 and v11 | severity: 2 type: a11y βΏ component: ui-shell status: waiting for maintainer response π¬ | ### Package
@carbon/react
### Browser
Chrome, Firefox
### Operating System
MacOS
### Package version
v10, v11
### React version
_No response_
### Automated testing tool and ruleset
IBM Accessibility Checker - Latest Deployment
### Assistive technology
_No response_
### Description
I used IBM Accessibility Checker to scan a newly developed page based on React which uses the UI Shell w/ SideNav component. When I scan this component independently (the page on storybook), violations are reported.
### WCAG 2.1 Violation
1.3.1 Info and Relationships, 2.4.1 Bypass Blocks, 4.1.2 Name, Role, Value (v10) and 1.3.1 Info and Relationships, 4.1.2 Name, Role, Value (v11).
### Reproduction/example
Violations are found on Carbon-React storybook for v10 and v11
### Steps to reproduce
To reproduce violations on v10:
1. Go to https://v10-react.carbondesignsystem.com/iframe.html?id=components-ui-shell--header-base-w-side-nav&viewMode=story&args=
2. Scan this page using IBM Accessibility Checkers
3. 8 violations of 4.1.2 Name, Role, Value will be found:

To reproduce violations on v11:
1. Go to https://react.carbondesignsystem.com/iframe.html?viewMode=story&id=components-ui-shell--header-base-w-side-nav&args=
2. Scan this page using IBM Accessibility Checkers
3. 5 violation

s of 2.4.1 Bypass Blocks will be found
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | True | [a11y]: Violations found on UI Shell w/ SideNav component in v10 and v11 - ### Package
@carbon/react
### Browser
Chrome, Firefox
### Operating System
MacOS
### Package version
v10, v11
### React version
_No response_
### Automated testing tool and ruleset
IBM Accessibility Checker - Latest Deployment
### Assistive technology
_No response_
### Description
I used IBM Accessibility Checker to scan a newly developed page based on React which uses the UI Shell w/ SideNav component. When I scan this component independently (the page on storybook), violations are reported.
### WCAG 2.1 Violation
1.3.1 Info and Relationships, 2.4.1 Bypass Blocks, 4.1.2 Name, Role, Value (v10) and 1.3.1 Info and Relationships, 4.1.2 Name, Role, Value (v11).
### Reproduction/example
Violations are found on Carbon-React storybook for v10 and v11
### Steps to reproduce
To reproduce violations on v10:
1. Go to https://v10-react.carbondesignsystem.com/iframe.html?id=components-ui-shell--header-base-w-side-nav&viewMode=story&args=
2. Scan this page using IBM Accessibility Checkers
3. 8 violations of 4.1.2 Name, Role, Value will be found:

To reproduce violations on v11:
1. Go to https://react.carbondesignsystem.com/iframe.html?viewMode=story&id=components-ui-shell--header-base-w-side-nav&args=
2. Scan this page using IBM Accessibility Checkers
3. 5 violation

s of 2.4.1 Bypass Blocks will be found
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | main | violations found on ui shell w sidenav component in and package carbon react browser chrome firefox operating system macos package version react version no response automated testing tool and ruleset ibm accessibility checker latest deployment assistive technology no response description i used ibm accessibility checker to scan a newly developed page based on react which uses the ui shell w sidenav component when i scan this component independently the page on storybook violations are reported wcag violation info and relationships bypass blocks name role value and info and relationships name role value reproduction example violations are found on carbon react storybook for and steps to reproduce to reproduce violations on go to scan this page using ibm accessibility checkers violations of name role value will be found to reproduce violations on go to scan this page using ibm accessibility checkers violation s of bypass blocks will be found code of conduct i agree to follow this project s i checked the for duplicate problems | 1 |
1,124 | 4,995,646,808 | IssuesEvent | 2016-12-09 10:53:16 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | vmware_guest is very slow | affects_2.2 bug_report cloud vmware waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
vmware_guest.py
##### ANSIBLE VERSION
```
2.2.0
```
##### SUMMARY
I have a lot of VMs and folders in my vcenter and _build_folder_map function take approximately 3 minutes to run.
Have you an idea to accelerate it ? | True | vmware_guest is very slow - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
vmware_guest.py
##### ANSIBLE VERSION
```
2.2.0
```
##### SUMMARY
I have a lot of VMs and folders in my vcenter and _build_folder_map function take approximately 3 minutes to run.
Have you an idea to accelerate it ? | main | vmware guest is very slow issue type bug report component name vmware guest py ansible version summary i have a lot of vms and folders in my vcenter and build folder map function take approximately minutes to run have you an idea to accelerate it | 1 |
1,392 | 6,025,276,083 | IssuesEvent | 2017-06-08 08:16:04 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_iis_webbinding fails to assign specified certificate_hash | affects_2.1 bug_report waiting_on_maintainer windows | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
win_iis_webbinding
##### ANSIBLE VERSION
<!--- Paste verbatim output from βansible --versionβ between quotes below -->
```
ansible 2.1.1.0 (detached HEAD 35da6ba9d1) last updated 2016/07/25 09:38:23 (GMT +000)
lib/ansible/modules/core: (detached HEAD 45128c8bab) last updated 2016/07/25 09:38:23 (GMT +000)
lib/ansible/modules/extras: (detached HEAD 511752e53a) last updated 2016/07/25 09:38:23 (GMT +000)
config file = /tmp/***/ansible/ansible.cfg
configured module search path = ['modules']
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
[defaults]
roles_path = roles
library = modules
##### OS / ENVIRONMENT
Server 2012 R2
##### SUMMARY
win_iis_webbinding module is not assigning the certificate specified in the certificate_hash variable.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
- Install certificate on to IIS server (we've tested this with a PS script and installing manually, it's installed properly).
- Paste the certificate hash in to the win_iis_webbinding module
- After the play, the IIS website should have a HTTPS web binding attached with the specified certificate - it doesn't.
<!--- Paste example playbooks or commands between quotes below -->
```
- name: IIS | Add HTTPS binding with cert for Default Website
win_iis_webbinding:
name: "Default Web Site"
protocol: https
port: 443
state: present
certificate_hash: β"***VALID CERT HASH OF INSTALLED CERT*****"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
HTTPS web binding created with specified certificate used.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
A web binding is created successfully for HTTPS, but the certificate is not attached properly.
<!--- Paste verbatim command output between quotes below -->
```
changed: [ec2-**-***-***-***.eu-west-1.compute.amazonaws.com] => {"added": [{"bindingInformation": "*:443:", "certificateHash": "", "certificateStoreName": "", "isDsMapperEnabled": false, "protocol": "https", "sslFlags": 0}], "changed": true, "invocation": {"module_name": "win_iis_webbinding"}, "ip": "0.0.0.0", "matched": [], "parameters": {"Name": "Default Web Site", "Port": 443, "Protocol": "https"}, "port": 443, "removed": []}
```
| True | win_iis_webbinding fails to assign specified certificate_hash - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
win_iis_webbinding
##### ANSIBLE VERSION
<!--- Paste verbatim output from βansible --versionβ between quotes below -->
```
ansible 2.1.1.0 (detached HEAD 35da6ba9d1) last updated 2016/07/25 09:38:23 (GMT +000)
lib/ansible/modules/core: (detached HEAD 45128c8bab) last updated 2016/07/25 09:38:23 (GMT +000)
lib/ansible/modules/extras: (detached HEAD 511752e53a) last updated 2016/07/25 09:38:23 (GMT +000)
config file = /tmp/***/ansible/ansible.cfg
configured module search path = ['modules']
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
[defaults]
roles_path = roles
library = modules
##### OS / ENVIRONMENT
Server 2012 R2
##### SUMMARY
win_iis_webbinding module is not assigning the certificate specified in the certificate_hash variable.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
- Install certificate on to IIS server (we've tested this with a PS script and installing manually, it's installed properly).
- Paste the certificate hash in to the win_iis_webbinding module
- After the play, the IIS website should have a HTTPS web binding attached with the specified certificate - it doesn't.
<!--- Paste example playbooks or commands between quotes below -->
```
- name: IIS | Add HTTPS binding with cert for Default Website
win_iis_webbinding:
name: "Default Web Site"
protocol: https
port: 443
state: present
certificate_hash: β"***VALID CERT HASH OF INSTALLED CERT*****"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
HTTPS web binding created with specified certificate used.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
A web binding is created successfully for HTTPS, but the certificate is not attached properly.
<!--- Paste verbatim command output between quotes below -->
```
changed: [ec2-**-***-***-***.eu-west-1.compute.amazonaws.com] => {"added": [{"bindingInformation": "*:443:", "certificateHash": "", "certificateStoreName": "", "isDsMapperEnabled": false, "protocol": "https", "sslFlags": 0}], "changed": true, "invocation": {"module_name": "win_iis_webbinding"}, "ip": "0.0.0.0", "matched": [], "parameters": {"Name": "Default Web Site", "Port": 443, "Protocol": "https"}, "port": 443, "removed": []}
```
| main | win iis webbinding fails to assign specified certificate hash issue type bug report component name win iis webbinding ansible version ansible detached head last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file tmp ansible ansible cfg configured module search path configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables roles path roles library modules os environment server summary win iis webbinding module is not assigning the certificate specified in the certificate hash variable steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used install certificate on to iis server we ve tested this with a ps script and installing manually it s installed properly paste the certificate hash in to the win iis webbinding module after the play the iis website should have a https web binding attached with the specified certificate it doesn t name iis add https binding with cert for default website win iis webbinding name default web site protocol https port state present certificate hash β valid cert hash of installed cert expected results https web binding created with specified certificate used actual results a web binding is created successfully for https but the certificate is not attached properly changed added changed true invocation module name win iis webbinding ip matched parameters name default web site port protocol https port removed | 1 |
26,409 | 7,835,342,476 | IssuesEvent | 2018-06-17 03:53:10 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | closed | Build-JDK8-win_x86 i686-w64-mingw32-g++: Command not found | comp:build | https://ci.eclipse.org/openj9/job/Build-JDK8-win_x86/3/console
I assume a machine setup issue.
win2012r2-x86-1
The build has previously worked on win2012r2-x86-3
```
07:29:22 make[3]: i686-w64-mingw32-g++: Command not found
07:29:22 make[3]: i686-w64-mingw32-g++: Command not found
07:29:22 make[3]: *** [../makelib/targets.mk:438: BytecodeInterpreter.obj] Error 127
07:29:22 make[3]: *** Waiting for unfinished jobs....
07:29:22 make[3]: *** [../makelib/targets.mk:441: DebugBytecodeInterpreter.obj] Error 127
``` | 1.0 | Build-JDK8-win_x86 i686-w64-mingw32-g++: Command not found - https://ci.eclipse.org/openj9/job/Build-JDK8-win_x86/3/console
I assume a machine setup issue.
win2012r2-x86-1
The build has previously worked on win2012r2-x86-3
```
07:29:22 make[3]: i686-w64-mingw32-g++: Command not found
07:29:22 make[3]: i686-w64-mingw32-g++: Command not found
07:29:22 make[3]: *** [../makelib/targets.mk:438: BytecodeInterpreter.obj] Error 127
07:29:22 make[3]: *** Waiting for unfinished jobs....
07:29:22 make[3]: *** [../makelib/targets.mk:441: DebugBytecodeInterpreter.obj] Error 127
``` | non_main | build win g command not found i assume a machine setup issue the build has previously worked on make g command not found make g command not found make error make waiting for unfinished jobs make error | 0 |
3,074 | 11,642,592,238 | IssuesEvent | 2020-02-29 08:01:06 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | It would be very cool if there is a ftp net_tool. | affects_2.9 deprecated feature infoblox module needs_maintainer needs_triage net_tools networking support:certified support:community support:core | <!--- Verify first that your feature was not already discussed on GitHub -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
[get_url](https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/net_tools/basics/get_url.py]) which can download files from HTTP, HTTPS, or FTP to the remote server.
How about uploading files? It seems that there is no module to upload files to HTTP, HTTPS, or FTP to the remote server.Sometimes ftp is a good net tool for upload/download files.
Can ansible have a module for that ?
Thanks for your attention.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
ansible/lib/ansible/modules/net_tools/basics/ftp.py
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
```
<!--- HINT: You can also paste gist.github.com links for larger files -->
| True | It would be very cool if there is a ftp net_tool. - <!--- Verify first that your feature was not already discussed on GitHub -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
[get_url](https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/net_tools/basics/get_url.py]) which can download files from HTTP, HTTPS, or FTP to the remote server.
How about uploading files? It seems that there is no module to upload files to HTTP, HTTPS, or FTP to the remote server.Sometimes ftp is a good net tool for upload/download files.
Can ansible have a module for that ?
Thanks for your attention.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
ansible/lib/ansible/modules/net_tools/basics/ftp.py
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
```
<!--- HINT: You can also paste gist.github.com links for larger files -->
| main | it would be very cool if there is a ftp net tool summary which can download files from http https or ftp to the remote server how about uploading files it seems that there is no module to upload files to http https or ftp to the remote server sometimes ftp is a good net tool for upload download files can ansible have a module for that thanks for your attention issue type feature idea component name ansible lib ansible modules net tools basics ftp py additional information yaml | 1 |
1,439 | 6,254,208,027 | IssuesEvent | 2017-07-14 01:03:28 | Microsoft/DirectXMath | https://api.github.com/repos/Microsoft/DirectXMath | opened | Remove VS 2013 compiler support | maintainence | At some point it will make sense to retire VS 2013 compiler support. This will enable the ability to remove the ``constexpr`` workaround (assuming I also drop support for VS 2015 RTM):
#if defined(_MSC_VER) && (_MSC_FULL_VER < 190023506)
#define XM_CONST const
#define XM_CONSTEXPR
#else
#define XM_CONST constexpr
#define XM_CONSTEXPR constexpr
#endif
| True | Remove VS 2013 compiler support - At some point it will make sense to retire VS 2013 compiler support. This will enable the ability to remove the ``constexpr`` workaround (assuming I also drop support for VS 2015 RTM):
#if defined(_MSC_VER) && (_MSC_FULL_VER < 190023506)
#define XM_CONST const
#define XM_CONSTEXPR
#else
#define XM_CONST constexpr
#define XM_CONSTEXPR constexpr
#endif
| main | remove vs compiler support at some point it will make sense to retire vs compiler support this will enable the ability to remove the constexpr workaround assuming i also drop support for vs rtm if defined msc ver msc full ver define xm const const define xm constexpr else define xm const constexpr define xm constexpr constexpr endif | 1 |
4,379 | 22,287,203,914 | IssuesEvent | 2022-06-11 20:35:00 | chocolatey-community/chocolatey-package-requests | https://api.github.com/repos/chocolatey-community/chocolatey-package-requests | closed | RFM - privoxy | Status: Available For Maintainer(s) Embeddable | ## I DON'T Want To Become The Maintainer
- [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package;
- [x] There is no existing open maintainer request for this package;
## Checklist
- [x] Issue title starts with 'RFM - '
## Existing Package Details
Package URL: https://chocolatey.org/packages/privoxy
Package source URL: https://github.com/dtgm/chocolatey-packages/tree/master/automatic/privoxy/
Date the maintainer was contacted (in YYYY-MM-DD): 2021-12-13
How the maintainer was contacted: Contacted via Email | True | RFM - privoxy - ## I DON'T Want To Become The Maintainer
- [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package;
- [x] There is no existing open maintainer request for this package;
## Checklist
- [x] Issue title starts with 'RFM - '
## Existing Package Details
Package URL: https://chocolatey.org/packages/privoxy
Package source URL: https://github.com/dtgm/chocolatey-packages/tree/master/automatic/privoxy/
Date the maintainer was contacted (in YYYY-MM-DD): 2021-12-13
How the maintainer was contacted: Contacted via Email | main | rfm privoxy i don t want to become the maintainer i have followed the package triage process and i do not want to become maintainer of the package there is no existing open maintainer request for this package checklist issue title starts with rfm existing package details package url package source url date the maintainer was contacted in yyyy mm dd how the maintainer was contacted contacted via email | 1 |
2,346 | 8,392,686,410 | IssuesEvent | 2018-10-09 18:20:42 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Archive module doesn't work with python 2.7 and LZMA package, asks for backport. | affects_2.7 bug module needs_maintainer python3 support:community traceback | <!---
Verify first that your issue/request is not already reported on GitHub.
THIS FORM WILL BE READ BY A MACHINE, COMPLETE ALL SECTIONS AS DESCRIBED.
Also test if the latest release, and devel branch are affected too.
ALWAYS add information AFTER (OUTSIDE) these html comments.
Otherwise it may end up being automatically closed by our bot. -->
##### SUMMARY
<!--- Explain the problem briefly -->
I've tried using the xz archive extension from ansible and can't seem to make it work.
Whenever i try to execute the script ansible errors out with the message:
` lzma or backports.lzma is required when using xz format.`
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Insert, BELOW THIS COMMENT, the name of the module, plugin, task or feature.
Do not include extra details here, e.g. "vyos_command" not "the network module vyos_command" or the full path-->
archive
##### ANSIBLE VERSION
<!--- Paste, BELOW THIS COMMENT, verbatim output from "ansible --version" between quotes below -->
```
ansible 2.7.0.dev0
config file = /root/.ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python2.7/dist-packages/ansible-2.7.0.dev0-py2.7.egg/ansible
executable location = /usr/local/bin/ansible
python version = 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609]
```
##### CONFIGURATION
<!--- If using Ansible 2.4 or above, paste, BELOW THIS COMMENT, the results of "ansible-config dump --only-changed"
Otherwise, mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).-->
```
DEFAULT_CALLBACK_WHITELIST(/root/.ansible.cfg) = [u'time']
DEFAULT_REMOTE_USER(/root/.ansible.cfg) = root
DEFAULT_STDOUT_CALLBACK(/root/.ansible.cfg) = debug
HOST_KEY_CHECKING(/root/.ansible.cfg) = False
RETRY_FILES_SAVE_PATH(/root/.ansible.cfg) = /root/.ansible/retry
```
##### OS / ENVIRONMENT
<!--- Mention, BELOW THIS COMMENT, the OS you are running Ansible from, and the OS you are
managing, or say "N/A" for anything that is not platform-specific.
Also mention the specific version of what you are trying to control,
e.g. if this is a network bug the version of firmware on the network device.-->
Linux 4.4.0-124-generic #148-Ubuntu SMP Wed May 2 13:00:18 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
##### STEPS TO REPRODUCE
<!--- For bugs, show exactly how to reproduce the problem, using a minimal test-case.
For new features, show how the feature would be used. -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- name: "-- proof of concept"
hosts: localhost
- name: "-- archive something"
archive:
format: xz
path: "./something"
dest: "./something.xz"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Ansible creates package something.xz
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
```
MSG:
lzma or backports.lzma is required when using xz format.
```
---
##### Further testing and other debug information
<!--- Paste verbatim command output between quotes below -->
```
13:56$ > python
Python 2.7.12 (default, Dec 4 2017, 14:50:18)
[GCC 5.4.0 20160609] on linux2
Type "help", "copyright", "credits" or "license" for more information.
>>> from backports import lzma
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: cannot import name lzma
>>> import lzma
>>>
```
---
```
temporary-build-server @ root: ~
13:56$ > python3
Python 3.5.2 (default, Nov 23 2017, 16:37:01)
[GCC 5.4.0 20160609] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> from backports import lzma
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: No module named 'backports'
>>> import lzma
>>>
```
---
```
13:57$ > dpkg -l |grep lzma
ii liblzma5:amd64 5.1.1alpha+20120614-2ubuntu2 amd64 XZ-format compression library
ii lzma 9.22-2ubuntu2 amd64 Compression and decompression in the LZMA format - command line utility
ii python-lzma 0.5.3-3 amd64 Python bindings for liblzma
```
---
##### Source code modifications
__this section seem to be the culprit__
https://github.com/ansible/ansible/blob/9ff20521d1ada2acc64b623875b1d8e51809e0f9/lib/ansible/modules/files/archive.py#L150-L161
Since python is identifying my installation as python 2.7 it defaults to asking for backports, even tho the actual lzma import is there. This might be caused by lzma being installed through `apt` and not `pip`
In my humble opinion the code should check for both lzma and backports.lzma.
---
##### Temporary fix
Installing `backports.lzma` through pip and pip3 seems to fix the issue on my machine.
| True | Archive module doesn't work with python 2.7 and LZMA package, asks for backport. - <!---
Verify first that your issue/request is not already reported on GitHub.
THIS FORM WILL BE READ BY A MACHINE, COMPLETE ALL SECTIONS AS DESCRIBED.
Also test if the latest release, and devel branch are affected too.
ALWAYS add information AFTER (OUTSIDE) these html comments.
Otherwise it may end up being automatically closed by our bot. -->
##### SUMMARY
<!--- Explain the problem briefly -->
I've tried using the xz archive extension from ansible and can't seem to make it work.
Whenever i try to execute the script ansible errors out with the message:
` lzma or backports.lzma is required when using xz format.`
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Insert, BELOW THIS COMMENT, the name of the module, plugin, task or feature.
Do not include extra details here, e.g. "vyos_command" not "the network module vyos_command" or the full path-->
archive
##### ANSIBLE VERSION
<!--- Paste, BELOW THIS COMMENT, verbatim output from "ansible --version" between quotes below -->
```
ansible 2.7.0.dev0
config file = /root/.ansible.cfg
configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python2.7/dist-packages/ansible-2.7.0.dev0-py2.7.egg/ansible
executable location = /usr/local/bin/ansible
python version = 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609]
```
##### CONFIGURATION
<!--- If using Ansible 2.4 or above, paste, BELOW THIS COMMENT, the results of "ansible-config dump --only-changed"
Otherwise, mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).-->
```
DEFAULT_CALLBACK_WHITELIST(/root/.ansible.cfg) = [u'time']
DEFAULT_REMOTE_USER(/root/.ansible.cfg) = root
DEFAULT_STDOUT_CALLBACK(/root/.ansible.cfg) = debug
HOST_KEY_CHECKING(/root/.ansible.cfg) = False
RETRY_FILES_SAVE_PATH(/root/.ansible.cfg) = /root/.ansible/retry
```
##### OS / ENVIRONMENT
<!--- Mention, BELOW THIS COMMENT, the OS you are running Ansible from, and the OS you are
managing, or say "N/A" for anything that is not platform-specific.
Also mention the specific version of what you are trying to control,
e.g. if this is a network bug the version of firmware on the network device.-->
Linux 4.4.0-124-generic #148-Ubuntu SMP Wed May 2 13:00:18 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux
##### STEPS TO REPRODUCE
<!--- For bugs, show exactly how to reproduce the problem, using a minimal test-case.
For new features, show how the feature would be used. -->
<!--- Paste example playbooks or commands between quotes below -->
```yaml
---
- name: "-- proof of concept"
hosts: localhost
- name: "-- archive something"
archive:
format: xz
path: "./something"
dest: "./something.xz"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Ansible creates package something.xz
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
```
MSG:
lzma or backports.lzma is required when using xz format.
```
---
##### Further testing and other debug information
<!--- Paste verbatim command output between quotes below -->
```
13:56$ > python
Python 2.7.12 (default, Dec 4 2017, 14:50:18)
[GCC 5.4.0 20160609] on linux2
Type "help", "copyright", "credits" or "license" for more information.
>>> from backports import lzma
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: cannot import name lzma
>>> import lzma
>>>
```
---
```
temporary-build-server @ root: ~
13:56$ > python3
Python 3.5.2 (default, Nov 23 2017, 16:37:01)
[GCC 5.4.0 20160609] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> from backports import lzma
Traceback (most recent call last):
File "<stdin>", line 1, in <module>
ImportError: No module named 'backports'
>>> import lzma
>>>
```
---
```
13:57$ > dpkg -l |grep lzma
ii liblzma5:amd64 5.1.1alpha+20120614-2ubuntu2 amd64 XZ-format compression library
ii lzma 9.22-2ubuntu2 amd64 Compression and decompression in the LZMA format - command line utility
ii python-lzma 0.5.3-3 amd64 Python bindings for liblzma
```
---
##### Source code modifications
__this section seem to be the culprit__
https://github.com/ansible/ansible/blob/9ff20521d1ada2acc64b623875b1d8e51809e0f9/lib/ansible/modules/files/archive.py#L150-L161
Since python is identifying my installation as python 2.7 it defaults to asking for backports, even tho the actual lzma import is there. This might be caused by lzma being installed through `apt` and not `pip`
In my humble opinion the code should check for both lzma and backports.lzma.
---
##### Temporary fix
Installing `backports.lzma` through pip and pip3 seems to fix the issue on my machine.
| main | archive module doesn t work with python and lzma package asks for backport verify first that your issue request is not already reported on github this form will be read by a machine complete all sections as described also test if the latest release and devel branch are affected too always add information after outside these html comments otherwise it may end up being automatically closed by our bot summary i ve tried using the xz archive extension from ansible and can t seem to make it work whenever i try to execute the script ansible errors out with the message lzma or backports lzma is required when using xz format issue type bug report component name insert below this comment the name of the module plugin task or feature do not include extra details here e g vyos command not the network module vyos command or the full path archive ansible version ansible config file root ansible cfg configured module search path ansible python module location usr local lib dist packages ansible egg ansible executable location usr local bin ansible python version default dec configuration if using ansible or above paste below this comment the results of ansible config dump only changed otherwise mention any settings you have changed added removed in ansible cfg or using the ansible environment variables default callback whitelist root ansible cfg default remote user root ansible cfg root default stdout callback root ansible cfg debug host key checking root ansible cfg false retry files save path root ansible cfg root ansible retry os environment mention below this comment the os you are running ansible from and the os you are managing or say n a for anything that is not platform specific also mention the specific version of what you are trying to control e g if this is a network bug the version of firmware on the network device linux generic ubuntu smp wed may utc gnu linux steps to reproduce for bugs show exactly how to reproduce the problem using a minimal test case for new features show how the feature would be used yaml name proof of concept hosts localhost name archive something archive format xz path something dest something xz expected results ansible creates package something xz actual results msg lzma or backports lzma is required when using xz format further testing and other debug information python python default dec on type help copyright credits or license for more information from backports import lzma traceback most recent call last file line in importerror cannot import name lzma import lzma temporary build server root python default nov on linux type help copyright credits or license for more information from backports import lzma traceback most recent call last file line in importerror no module named backports import lzma dpkg l grep lzma ii xz format compression library ii lzma compression and decompression in the lzma format command line utility ii python lzma python bindings for liblzma source code modifications this section seem to be the culprit since python is identifying my installation as python it defaults to asking for backports even tho the actual lzma import is there this might be caused by lzma being installed through apt and not pip in my humble opinion the code should check for both lzma and backports lzma temporary fix installing backports lzma through pip and seems to fix the issue on my machine | 1 |
49,475 | 12,345,381,157 | IssuesEvent | 2020-05-15 08:51:10 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | opened | There is no synchronization between replicas, neither in readonly state nor in synchronization | build | 2020.05.15 16:50:06.800472 [ 84 ] {} <Error> k19_test.replica_shard (ReplicatedMergeTreeRestartingThread): void DB::ReplicatedMergeTreeRestartingThread::run(): Code: 27, e.displayText() = DB::Exception: Cannot parse input: expected format version: at end of stream., Stack trace (when copying this message, always include the lines below):
0. 0x100ac1bc Poco::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse
1. 0x8e74849 DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse
2. 0x8eaacd5 ? in /usr/bin/clickhouse
3. 0x8ea8caa DB::assertString(char const*, DB::ReadBuffer&) in /usr/bin/clickhouse
4. 0xd78dd1b DB::ReplicatedMergeTreeLogEntryData::readText(DB::ReadBuffer&) in /usr/bin/clickhouse
5. 0xd78f04b DB::ReplicatedMergeTreeLogEntry::parse(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, Coordination::Stat const&) in /usr/bin/clickhouse
6. 0xd7b49c1 DB::ReplicatedMergeTreeQueue::load(std::__1::shared_ptr<zkutil::ZooKeeper>) in /usr/bin/clickhouse
7. 0xd7d7393 DB::ReplicatedMergeTreeRestartingThread::tryStartup() in /usr/bin/clickhouse
8. 0xd7d7cf8 DB::ReplicatedMergeTreeRestartingThread::run() in /usr/bin/clickhouse
9. 0xcd939f1 DB::BackgroundSchedulePoolTaskInfo::execute() in /usr/bin/clickhouse
10. 0xcd93fca DB::BackgroundSchedulePool::threadFunction() in /usr/bin/clickhouse
11. 0xcd94100 ? in /usr/bin/clickhouse
12. 0x8e97347 ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) in /usr/bin/clickhouse
13. 0x8e9580f ? in /usr/bin/clickhouse
14. 0x7e25 start_thread in /usr/lib64/libpthread-2.17.so
15. 0xfebad __clone in /usr/lib64/libc-2.17.so
(version 20.1.6.30 (official build))
| 1.0 | There is no synchronization between replicas, neither in readonly state nor in synchronization - 2020.05.15 16:50:06.800472 [ 84 ] {} <Error> k19_test.replica_shard (ReplicatedMergeTreeRestartingThread): void DB::ReplicatedMergeTreeRestartingThread::run(): Code: 27, e.displayText() = DB::Exception: Cannot parse input: expected format version: at end of stream., Stack trace (when copying this message, always include the lines below):
0. 0x100ac1bc Poco::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse
1. 0x8e74849 DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse
2. 0x8eaacd5 ? in /usr/bin/clickhouse
3. 0x8ea8caa DB::assertString(char const*, DB::ReadBuffer&) in /usr/bin/clickhouse
4. 0xd78dd1b DB::ReplicatedMergeTreeLogEntryData::readText(DB::ReadBuffer&) in /usr/bin/clickhouse
5. 0xd78f04b DB::ReplicatedMergeTreeLogEntry::parse(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, Coordination::Stat const&) in /usr/bin/clickhouse
6. 0xd7b49c1 DB::ReplicatedMergeTreeQueue::load(std::__1::shared_ptr<zkutil::ZooKeeper>) in /usr/bin/clickhouse
7. 0xd7d7393 DB::ReplicatedMergeTreeRestartingThread::tryStartup() in /usr/bin/clickhouse
8. 0xd7d7cf8 DB::ReplicatedMergeTreeRestartingThread::run() in /usr/bin/clickhouse
9. 0xcd939f1 DB::BackgroundSchedulePoolTaskInfo::execute() in /usr/bin/clickhouse
10. 0xcd93fca DB::BackgroundSchedulePool::threadFunction() in /usr/bin/clickhouse
11. 0xcd94100 ? in /usr/bin/clickhouse
12. 0x8e97347 ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) in /usr/bin/clickhouse
13. 0x8e9580f ? in /usr/bin/clickhouse
14. 0x7e25 start_thread in /usr/lib64/libpthread-2.17.so
15. 0xfebad __clone in /usr/lib64/libc-2.17.so
(version 20.1.6.30 (official build))
| non_main | there is no synchronization between replicas neither in readonly state nor in synchronization test replica shard replicatedmergetreerestartingthread void db replicatedmergetreerestartingthread run code e displaytext db exception cannot parse input expected format version at end of stream stack trace when copying this message always include the lines below poco exception exception std basic string std allocator const int in usr bin clickhouse db exception exception std basic string std allocator const int in usr bin clickhouse in usr bin clickhouse db assertstring char const db readbuffer in usr bin clickhouse db replicatedmergetreelogentrydata readtext db readbuffer in usr bin clickhouse db replicatedmergetreelogentry parse std basic string std allocator const coordination stat const in usr bin clickhouse db replicatedmergetreequeue load std shared ptr in usr bin clickhouse db replicatedmergetreerestartingthread trystartup in usr bin clickhouse db replicatedmergetreerestartingthread run in usr bin clickhouse db backgroundschedulepooltaskinfo execute in usr bin clickhouse db backgroundschedulepool threadfunction in usr bin clickhouse in usr bin clickhouse threadpoolimpl worker std list iterator in usr bin clickhouse in usr bin clickhouse start thread in usr libpthread so clone in usr libc so version official build | 0 |
625,902 | 19,770,029,514 | IssuesEvent | 2022-01-17 09:06:16 | scipp/scipp | https://api.github.com/repos/scipp/scipp | closed | Consider wider build matrix options | discussion priority:low | Solving regressions resulting from changes in compiler flags and linking options is time consuming. Would it be better to build with some of these options set to exercise the functionality and find regressions earlier.
A few more recent aspect have made me raise this question.
- With #1670 we had a test segfault, that actually displayed as an assert error in debug mode. Would have been easier found if this was visible in the logs.
- Recent work on dynamic libs for scipp (optional)
- Recent introduction of precombiled headers (optional)
- A thought on packaging -It may actually be desirable to package and distribute via conda Release with Debug information. Would allow users to swap packages on beamlines, re-run and attach debugger
We will have to strike a balance between resource use and time (of CI) and resource use an time of developers.
**Questions**
- What options would we exercise?
- How frequently should we run with these options?
Related issue:
https://github.com/scipp/scipp/issues/1244
| 1.0 | Consider wider build matrix options - Solving regressions resulting from changes in compiler flags and linking options is time consuming. Would it be better to build with some of these options set to exercise the functionality and find regressions earlier.
A few more recent aspect have made me raise this question.
- With #1670 we had a test segfault, that actually displayed as an assert error in debug mode. Would have been easier found if this was visible in the logs.
- Recent work on dynamic libs for scipp (optional)
- Recent introduction of precombiled headers (optional)
- A thought on packaging -It may actually be desirable to package and distribute via conda Release with Debug information. Would allow users to swap packages on beamlines, re-run and attach debugger
We will have to strike a balance between resource use and time (of CI) and resource use an time of developers.
**Questions**
- What options would we exercise?
- How frequently should we run with these options?
Related issue:
https://github.com/scipp/scipp/issues/1244
| non_main | consider wider build matrix options solving regressions resulting from changes in compiler flags and linking options is time consuming would it be better to build with some of these options set to exercise the functionality and find regressions earlier a few more recent aspect have made me raise this question with we had a test segfault that actually displayed as an assert error in debug mode would have been easier found if this was visible in the logs recent work on dynamic libs for scipp optional recent introduction of precombiled headers optional a thought on packaging it may actually be desirable to package and distribute via conda release with debug information would allow users to swap packages on beamlines re run and attach debugger we will have to strike a balance between resource use and time of ci and resource use an time of developers questions what options would we exercise how frequently should we run with these options related issue | 0 |
55,181 | 23,408,153,120 | IssuesEvent | 2022-08-12 14:44:02 | elastic/kibana | https://api.github.com/repos/elastic/kibana | closed | Expressions: Improve test coverage | Feature:ExpressionLanguage loe:week Team:AppServicesSv impact:medium | Improve test coverage in Expressions plugin `src/plugins/expressions`.
- [ ] Add unit tests for everything
- [ ] Add few integration tests for the whole `expressions` plugin
- [ ] Set-up test coverage reporting
- [ ] Look into [moving interpreter functional snapshots to use the unified snapshot interface](https://github.com/elastic/kibana/issues/83955)
Parent issue: https://github.com/elastic/kibana/issues/46909
| 1.0 | Expressions: Improve test coverage - Improve test coverage in Expressions plugin `src/plugins/expressions`.
- [ ] Add unit tests for everything
- [ ] Add few integration tests for the whole `expressions` plugin
- [ ] Set-up test coverage reporting
- [ ] Look into [moving interpreter functional snapshots to use the unified snapshot interface](https://github.com/elastic/kibana/issues/83955)
Parent issue: https://github.com/elastic/kibana/issues/46909
| non_main | expressions improve test coverage improve test coverage in expressions plugin src plugins expressions add unit tests for everything add few integration tests for the whole expressions plugin set up test coverage reporting look into parent issue | 0 |
443,405 | 30,888,764,199 | IssuesEvent | 2023-08-04 01:47:51 | psf/black | https://api.github.com/repos/psf/black | closed | pre-commit hook fails when using `required-version` | T: documentation C: integrations C: configuration | This is attempting to use the pre-commit hook as described under [version control docs](https://black.readthedocs.io/en/stable/integrations/source_version_control.html) along with setting [--required-version](https://black.readthedocs.io/en/stable/usage_and_configuration/the_basics.html#versions) in a configuration file. The pre-commit hook fails with:
```
Oh no! π₯ π π₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`!
```
**To Reproduce**
First, a temporary folder/venv to test things in:
```
mkdir black-pre-commit-test
cd black-pre-commit-test
`which python3.9` -m venv venv
source venv/bin/activate
```
Then this script:
```
git init
pip install pre-commit
pre-commit install
cat > .pre-commit-config.yaml << EOF
repos:
- repo: https://github.com/psf/black
rev: 21.8b0
hooks:
- id: black
language_version: python3
EOF
cat > pyproject.toml << EOF
[tool.black]
required-version = '21.8b0'
EOF
git add .pre-commit-config.yaml
touch test.py
pre-commit run black --files test.py
```
**Expected output**
```
black....................................................................Passed
```
**Actual output**
```
black....................................................................Failed
- hook id: black
- exit code: 1
Oh no! π₯ π π₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`!
```
The same happens if doing a normal "git commit" - it is just easier to run the hook via `pre-commit run` directly. To test with hook:
```
git commit .pre-commit-config.yaml -m "pre-commit"
git add test.py
git commit test.py -m "test"
```
**Environment:**
- Version: 21.8b0
- OS: Linux,
- Python: 3.9.5 (i.e. `which python3.9` used in the beginning of the instructions above)
- However, the version installed by pre-commit in `~/.cache/pre-commit/repo8_9vcuko/py_env-python3/bin/python3` is different - it is whatever is pointed to by `which python3`, which was Python 3.8.10 first time I ran it. By changing `language_version` in `.pre-commit-config.yaml` I can get a different version, e.g. Python 3.9.5, but it made no difference.
**Does this bug also happen on main?**
Haven't tried, because this is specific to pinning to published versions.
**Notes**
I can see the incorrect Black version by doing this:
```
~/.cache/pre-commit/repo8_9vcuko/py_env-python3.9/bin/black --version
```
However, I haven't been able to further debug because I don't know how this line works: https://github.com/psf/black/blob/41e670064063e3e221b1c3c40db5aaae784b5231/src/black/__init__.py#L67 - it appears to be doing some magical import.
| 1.0 | pre-commit hook fails when using `required-version` - This is attempting to use the pre-commit hook as described under [version control docs](https://black.readthedocs.io/en/stable/integrations/source_version_control.html) along with setting [--required-version](https://black.readthedocs.io/en/stable/usage_and_configuration/the_basics.html#versions) in a configuration file. The pre-commit hook fails with:
```
Oh no! π₯ π π₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`!
```
**To Reproduce**
First, a temporary folder/venv to test things in:
```
mkdir black-pre-commit-test
cd black-pre-commit-test
`which python3.9` -m venv venv
source venv/bin/activate
```
Then this script:
```
git init
pip install pre-commit
pre-commit install
cat > .pre-commit-config.yaml << EOF
repos:
- repo: https://github.com/psf/black
rev: 21.8b0
hooks:
- id: black
language_version: python3
EOF
cat > pyproject.toml << EOF
[tool.black]
required-version = '21.8b0'
EOF
git add .pre-commit-config.yaml
touch test.py
pre-commit run black --files test.py
```
**Expected output**
```
black....................................................................Passed
```
**Actual output**
```
black....................................................................Failed
- hook id: black
- exit code: 1
Oh no! π₯ π π₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`!
```
The same happens if doing a normal "git commit" - it is just easier to run the hook via `pre-commit run` directly. To test with hook:
```
git commit .pre-commit-config.yaml -m "pre-commit"
git add test.py
git commit test.py -m "test"
```
**Environment:**
- Version: 21.8b0
- OS: Linux,
- Python: 3.9.5 (i.e. `which python3.9` used in the beginning of the instructions above)
- However, the version installed by pre-commit in `~/.cache/pre-commit/repo8_9vcuko/py_env-python3/bin/python3` is different - it is whatever is pointed to by `which python3`, which was Python 3.8.10 first time I ran it. By changing `language_version` in `.pre-commit-config.yaml` I can get a different version, e.g. Python 3.9.5, but it made no difference.
**Does this bug also happen on main?**
Haven't tried, because this is specific to pinning to published versions.
**Notes**
I can see the incorrect Black version by doing this:
```
~/.cache/pre-commit/repo8_9vcuko/py_env-python3.9/bin/black --version
```
However, I haven't been able to further debug because I don't know how this line works: https://github.com/psf/black/blob/41e670064063e3e221b1c3c40db5aaae784b5231/src/black/__init__.py#L67 - it appears to be doing some magical import.
| non_main | pre commit hook fails when using required version this is attempting to use the pre commit hook as described under along with setting in a configuration file the pre commit hook fails with oh no π₯ π π₯ the required version does not match the running version to reproduce first a temporary folder venv to test things in mkdir black pre commit test cd black pre commit test which m venv venv source venv bin activate then this script git init pip install pre commit pre commit install cat pre commit config yaml eof repos repo rev hooks id black language version eof cat pyproject toml eof required version eof git add pre commit config yaml touch test py pre commit run black files test py expected output black passed actual output black failed hook id black exit code oh no π₯ π π₯ the required version does not match the running version the same happens if doing a normal git commit it is just easier to run the hook via pre commit run directly to test with hook git commit pre commit config yaml m pre commit git add test py git commit test py m test environment version os linux python i e which used in the beginning of the instructions above however the version installed by pre commit in cache pre commit py env bin is different it is whatever is pointed to by which which was python first time i ran it by changing language version in pre commit config yaml i can get a different version e g python but it made no difference does this bug also happen on main haven t tried because this is specific to pinning to published versions notes i can see the incorrect black version by doing this cache pre commit py env bin black version however i haven t been able to further debug because i don t know how this line works it appears to be doing some magical import | 0 |
2,299 | 8,221,956,555 | IssuesEvent | 2018-09-06 05:11:51 | TravisSpark/spark-website | https://api.github.com/repos/TravisSpark/spark-website | closed | Edit Readme | maintainence | ### Checklist
- [X] Searched for, and did not find, duplicate [issue](https://github.com/TravisSpark/spark-website/issues)
- [X] Indicated whether the issue is a bug or a feature
- [X] Focused on one specific bug/feature
- [X] Gave a concise and relevant name
- [X] Created clear and concise description
- [X] Outlined which components are affected
- [X] Assigned issue to project, appropriate contributors, and relevant labels
<!-- Edit as Appropriate -->
### Issue Type:
Feature
### Description
The descriptions in the read me lack impact and clarity. The phrasing should be edited. Use the homepage of travisspark.org as a template.
### Affected Components
.readme
| True | Edit Readme - ### Checklist
- [X] Searched for, and did not find, duplicate [issue](https://github.com/TravisSpark/spark-website/issues)
- [X] Indicated whether the issue is a bug or a feature
- [X] Focused on one specific bug/feature
- [X] Gave a concise and relevant name
- [X] Created clear and concise description
- [X] Outlined which components are affected
- [X] Assigned issue to project, appropriate contributors, and relevant labels
<!-- Edit as Appropriate -->
### Issue Type:
Feature
### Description
The descriptions in the read me lack impact and clarity. The phrasing should be edited. Use the homepage of travisspark.org as a template.
### Affected Components
.readme
| main | edit readme checklist searched for and did not find duplicate indicated whether the issue is a bug or a feature focused on one specific bug feature gave a concise and relevant name created clear and concise description outlined which components are affected assigned issue to project appropriate contributors and relevant labels issue type feature description the descriptions in the read me lack impact and clarity the phrasing should be edited use the homepage of travisspark org as a template affected components readme | 1 |
108,663 | 23,644,045,693 | IssuesEvent | 2022-08-25 20:02:18 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [4.0] No way to change the site name during installation | No Code Attached Yet J4 Issue | ### Steps to reproduce the issue
Start an installation. Enter the site name, go to "setup login data" (and mybe to the database setup)
Did you enter the site name right?
Do you want to update the site name?
### Expected result
You can see the site name you have entered in the first step.
You can change the site name
### Actual result
The sie name is not visible and there is
no way back to change the site name (except restarting the installation)
| 1.0 | [4.0] No way to change the site name during installation - ### Steps to reproduce the issue
Start an installation. Enter the site name, go to "setup login data" (and mybe to the database setup)
Did you enter the site name right?
Do you want to update the site name?
### Expected result
You can see the site name you have entered in the first step.
You can change the site name
### Actual result
The sie name is not visible and there is
no way back to change the site name (except restarting the installation)
| non_main | no way to change the site name during installation steps to reproduce the issue start an installation enter the site name go to setup login data and mybe to the database setup did you enter the site name right do you want to update the site name expected result you can see the site name you have entered in the first step you can change the site name actual result the sie name is not visible and there is no way back to change the site name except restarting the installation | 0 |
16,547 | 2,914,792,045 | IssuesEvent | 2015-06-23 08:23:07 | beefproject/beef | https://api.github.com/repos/beefproject/beef | closed | beef.net.request callbacks aren't being called | Core Defect Priority Low | I think I found an odd issue (Whilst working on Issue #1083) .. it appears that outside of the use of beef.net.request in core/main/client/updater.js (https://github.com/beefproject/beef/blob/master/core/main/client/updater.js#L60) the use of callbacks in beef.net.request() method calls do not actually get used.
I found this in the integration tests this module didn't work properly (https://github.com/beefproject/beef/blob/master/test/integration/tc_debug_modules.rb#L141)
And then subsequently, I couldn't get the coldfusion_dir_traversal_exploit module to work as it relies on this pattern too.
I'm unsure why this isn't working, whether the variables don't get passed into the callback function, or perhaps the module's use is slightly incorrect. Will investigate further. | 1.0 | beef.net.request callbacks aren't being called - I think I found an odd issue (Whilst working on Issue #1083) .. it appears that outside of the use of beef.net.request in core/main/client/updater.js (https://github.com/beefproject/beef/blob/master/core/main/client/updater.js#L60) the use of callbacks in beef.net.request() method calls do not actually get used.
I found this in the integration tests this module didn't work properly (https://github.com/beefproject/beef/blob/master/test/integration/tc_debug_modules.rb#L141)
And then subsequently, I couldn't get the coldfusion_dir_traversal_exploit module to work as it relies on this pattern too.
I'm unsure why this isn't working, whether the variables don't get passed into the callback function, or perhaps the module's use is slightly incorrect. Will investigate further. | non_main | beef net request callbacks aren t being called i think i found an odd issue whilst working on issue it appears that outside of the use of beef net request in core main client updater js the use of callbacks in beef net request method calls do not actually get used i found this in the integration tests this module didn t work properly and then subsequently i couldn t get the coldfusion dir traversal exploit module to work as it relies on this pattern too i m unsure why this isn t working whether the variables don t get passed into the callback function or perhaps the module s use is slightly incorrect will investigate further | 0 |
703 | 4,281,278,518 | IssuesEvent | 2016-07-15 01:45:48 | duckduckgo/zeroclickinfo-goodies | https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies | closed | Counter-Strike Go Cheat Sheet: Add aliases | Maintainer Input Requested Suggestion | This could probably benefit from a few aliases, for example:
"counter-strike go", "counter strike go", "cs:go", "counter-strike global offensive", "counter strike global offensive"
------
IA Page: http://duck.co/ia/view/csgo_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @aiyankovil | True | Counter-Strike Go Cheat Sheet: Add aliases - This could probably benefit from a few aliases, for example:
"counter-strike go", "counter strike go", "cs:go", "counter-strike global offensive", "counter strike global offensive"
------
IA Page: http://duck.co/ia/view/csgo_cheat_sheet
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @aiyankovil | main | counter strike go cheat sheet add aliases this could probably benefit from a few aliases for example counter strike go counter strike go cs go counter strike global offensive counter strike global offensive ia page aiyankovil | 1 |
4,993 | 25,708,406,639 | IssuesEvent | 2022-12-07 03:35:40 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | Accept boolean values in CloudWatchEvent patterns | type/feature maintainer/need-response | I have a Lambda that is triggered by CloudWatchEvents. I'm trying to match events that have "campaign" in their details as in [this example of the documentation](https://docs.aws.amazon.com/eventbridge/latest/userguide/content-filtering-with-event-patterns.html#filtering-exists-matching). This is the code in the lambda's Events property of my SAM template:
```
feedPost:
Type: CloudWatchEvent
Properties:
EventBusName: my_bus
Pattern:
source:
- mySource
detail:
campaign:
- exists: true
```
But when it is deploying, in CloudFormation the following error appears:
> Event pattern is not valid. Reason: exists match pattern must be either true or false. at [Source: (String)"{ "source":["mySource"],"detail":{"campaign":[{"exists":"true"}]}}";
So apparently it is taking the true value as a String and fails to deploy. | True | Accept boolean values in CloudWatchEvent patterns - I have a Lambda that is triggered by CloudWatchEvents. I'm trying to match events that have "campaign" in their details as in [this example of the documentation](https://docs.aws.amazon.com/eventbridge/latest/userguide/content-filtering-with-event-patterns.html#filtering-exists-matching). This is the code in the lambda's Events property of my SAM template:
```
feedPost:
Type: CloudWatchEvent
Properties:
EventBusName: my_bus
Pattern:
source:
- mySource
detail:
campaign:
- exists: true
```
But when it is deploying, in CloudFormation the following error appears:
> Event pattern is not valid. Reason: exists match pattern must be either true or false. at [Source: (String)"{ "source":["mySource"],"detail":{"campaign":[{"exists":"true"}]}}";
So apparently it is taking the true value as a String and fails to deploy. | main | accept boolean values in cloudwatchevent patterns i have a lambda that is triggered by cloudwatchevents i m trying to match events that have campaign in their details as in this is the code in the lambda s events property of my sam template feedpost type cloudwatchevent properties eventbusname my bus pattern source mysource detail campaign exists true but when it is deploying in cloudformation the following error appears event pattern is not valid reason exists match pattern must be either true or false at detail campaign so apparently it is taking the true value as a string and fails to deploy | 1 |
5,680 | 29,833,373,332 | IssuesEvent | 2023-06-18 14:40:37 | Windham-High-School/CubeServer | https://api.github.com/repos/Windham-High-School/CubeServer | closed | Refactor scoring code | maintainability | The current scoring code is ridiculously hard to follow and read, and is written in the same file as a database object mapping model class! | True | Refactor scoring code - The current scoring code is ridiculously hard to follow and read, and is written in the same file as a database object mapping model class! | main | refactor scoring code the current scoring code is ridiculously hard to follow and read and is written in the same file as a database object mapping model class | 1 |
1,089 | 4,939,643,278 | IssuesEvent | 2016-11-29 14:55:51 | numbbo/coco | https://api.github.com/repos/numbbo/coco | opened | data archives folder name convention | Maintainability question Usability | This is how it looks like currently:
<img width="381" alt="screen shot 2016-11-29 at 15 53 46" src="https://cloud.githubusercontent.com/assets/7316439/20714360/177deb12-b64c-11e6-9225-e7a61a78c48c.png">
Shouldn't we rename `bbob-biobj-2016` to `2016-bbob-biobj`? Do we have dependencies which would brake?
Even after this, the naming convention is inconsistent, as we have noiseless and noisy testbeds mixed. How about to have for each year/event also a subfolder for each testbed? | True | data archives folder name convention - This is how it looks like currently:
<img width="381" alt="screen shot 2016-11-29 at 15 53 46" src="https://cloud.githubusercontent.com/assets/7316439/20714360/177deb12-b64c-11e6-9225-e7a61a78c48c.png">
Shouldn't we rename `bbob-biobj-2016` to `2016-bbob-biobj`? Do we have dependencies which would brake?
Even after this, the naming convention is inconsistent, as we have noiseless and noisy testbeds mixed. How about to have for each year/event also a subfolder for each testbed? | main | data archives folder name convention this is how it looks like currently img width alt screen shot at src shouldn t we rename bbob biobj to bbob biobj do we have dependencies which would brake even after this the naming convention is inconsistent as we have noiseless and noisy testbeds mixed how about to have for each year event also a subfolder for each testbed | 1 |
4,659 | 24,097,560,643 | IssuesEvent | 2022-09-19 20:15:13 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | Sort sam init runtimes alphanumerically | type/ux area/init maintainer/need-followup | ### Describe your idea/feature/enhancement

The list of runtimes shown in `sam init` are hard to look through. For example, if I was interested in a node app, I have to hunt through the list to find one up top, and one in the middle (7 in the image).
I think two improvements could help improve time to locate the runtime of interest:
* alphanumeric-sort the runtimes
* right-align the numbers (1-13 in the image)
| True | Sort sam init runtimes alphanumerically - ### Describe your idea/feature/enhancement

The list of runtimes shown in `sam init` are hard to look through. For example, if I was interested in a node app, I have to hunt through the list to find one up top, and one in the middle (7 in the image).
I think two improvements could help improve time to locate the runtime of interest:
* alphanumeric-sort the runtimes
* right-align the numbers (1-13 in the image)
| main | sort sam init runtimes alphanumerically describe your idea feature enhancement the list of runtimes shown in sam init are hard to look through for example if i was interested in a node app i have to hunt through the list to find one up top and one in the middle in the image i think two improvements could help improve time to locate the runtime of interest alphanumeric sort the runtimes right align the numbers in the image | 1 |
6,706 | 9,815,575,684 | IssuesEvent | 2019-06-13 12:58:01 | linnovate/root | https://api.github.com/repos/linnovate/root | closed | in tasks and projects, "no select" option when selecting an assignee changes the status to assigned | 2.0.7 Fixed Process bug Projects Tasks | create a new task/ project
click on select assignee
click on "no select"
the status is changed to assigned

| 1.0 | in tasks and projects, "no select" option when selecting an assignee changes the status to assigned - create a new task/ project
click on select assignee
click on "no select"
the status is changed to assigned

| non_main | in tasks and projects no select option when selecting an assignee changes the status to assigned create a new task project click on select assignee click on no select the status is changed to assigned | 0 |
814,793 | 30,522,160,222 | IssuesEvent | 2023-07-19 08:49:06 | owncloud/web | https://api.github.com/repos/owncloud/web | closed | Search location filter is modified after opening a file in the text editor | Type:Bug Priority:p3-medium | steps:
Precondition: user has files with the same "content" in different places (personal, project, shares jail)
Steps:
- user searches files using fullTextSearch and `all files` filter
- open file from personal space in the text editor
- close text editor
Expected: `all files` filter in the search result.
Actual: `current folder` filter in the search result.
https://github.com/owncloud/web/assets/84779829/0a735e01-be71-42e6-94f7-f317f2247fbc
| 1.0 | Search location filter is modified after opening a file in the text editor - steps:
Precondition: user has files with the same "content" in different places (personal, project, shares jail)
Steps:
- user searches files using fullTextSearch and `all files` filter
- open file from personal space in the text editor
- close text editor
Expected: `all files` filter in the search result.
Actual: `current folder` filter in the search result.
https://github.com/owncloud/web/assets/84779829/0a735e01-be71-42e6-94f7-f317f2247fbc
| non_main | search location filter is modified after opening a file in the text editor steps precondition user has files with the same content in different places personal project shares jail steps user searches files using fulltextsearch and all files filter open file from personal space in the text editor close text editor expected all files filter in the search result actual current folder filter in the search result | 0 |
1,411 | 6,130,961,657 | IssuesEvent | 2017-06-24 11:01:17 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | Remove TimeSeries | API Component-Core Format-DCD maintainability | We discussed recently to remove the TimeSeries objects from MDAnalysis. As far as I know none of the core devs is using it. We also note that it is unlikely we will ever implement TimeSeries for any other readers besides dcd and memory and most of it's function can be replaced by the Analysis Framework.
The only part in the library that uses TimeSeries is encore. So it would be useful get some input from @wouterboomsma and @mtiberti before we go ahead with this.
As a first step I would remove timeseries from DCD during the cython transition.
In a later step we would remove timeseries from memory reader and remove the core TimeSeries class.
| True | Remove TimeSeries - We discussed recently to remove the TimeSeries objects from MDAnalysis. As far as I know none of the core devs is using it. We also note that it is unlikely we will ever implement TimeSeries for any other readers besides dcd and memory and most of it's function can be replaced by the Analysis Framework.
The only part in the library that uses TimeSeries is encore. So it would be useful get some input from @wouterboomsma and @mtiberti before we go ahead with this.
As a first step I would remove timeseries from DCD during the cython transition.
In a later step we would remove timeseries from memory reader and remove the core TimeSeries class.
| main | remove timeseries we discussed recently to remove the timeseries objects from mdanalysis as far as i know none of the core devs is using it we also note that it is unlikely we will ever implement timeseries for any other readers besides dcd and memory and most of it s function can be replaced by the analysis framework the only part in the library that uses timeseries is encore so it would be useful get some input from wouterboomsma and mtiberti before we go ahead with this as a first step i would remove timeseries from dcd during the cython transition in a later step we would remove timeseries from memory reader and remove the core timeseries class | 1 |
4,261 | 21,261,293,875 | IssuesEvent | 2022-04-13 04:45:56 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | ERPNEXT 13 Installation Easy install error URGENT !!!! | blocked/close-if-inactive maintainer/need-followup | I have tried to install ERPnext 13 more than 20 times , using all the tutorials but no good result ,
when I use Easy install [ sudo python3 install.py --verbose --production --user [USER] --mariadb-version 10.5 --frappe-branch version-13 --erpnext-branch version-13 ]
I got error , return none zero status 2
Could you please help me out
the whole result
fatal: [localhost]: FAILED! => {
"changed": true,
"cmd": [
"bench",
"init",
"/home/frappe/frappe-bench",
"--frappe-path",
"https://github.com/frappe/frappe",
"--frappe-branch",
"version-13",
"--python",
"python3"
],
"delta": "0:00:00.378029",
"end": "2022-02-20 10:45:29.953302",
"invocation": {
"module_args": {
"_raw_params": "bench init /home/frappe/frappe-bench --frappe-path https://github.com/frappe/frappe --frappe-branch version-13 --python python3",
"_uses_shell": false,
"argv": null,
"chdir": null,
"creates": "/home/frappe/frappe-bench",
"executable": null,
"removes": null,
"stdin": null,
"stdin_add_newline": true,
"strip_empty_ends": true,
"warn": true
}
},
"msg": "non-zero return code",
"rc": 1,
"start": "2022-02-20 10:45:29.575273",
"stderr": "Traceback (most recent call last):\n File "/usr/bin/bench", line 33, in \n sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())\n File "/usr/bin/bench", line 22, in importlib_load_entry_point\n for entry_point in distribution(dist_name).entry_points\n File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution\n return Distribution.from_name(distribution_name)\n File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name\n raise PackageNotFoundError(name)\nimportlib.metadata.PackageNotFoundError: frappe-bench",
"stderr_lines": [
"Traceback (most recent call last):",
" File "/usr/bin/bench", line 33, in ",
" sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())",
" File "/usr/bin/bench", line 22, in importlib_load_entry_point",
" for entry_point in distribution(dist_name).entry_points",
" File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution",
" return Distribution.from_name(distribution_name)",
" File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name",
" raise PackageNotFoundError(name)",
"importlib.metadata.PackageNotFoundError: frappe-bench"
],
"stdout": "",
"stdout_lines": []
}
PLAY RECAP *****************************************************************************************************************************************************************
localhost : ok=67 changed=17 unreachable=0 failed=1 skipped=62 rescued=0 ignored=0
Traceback (most recent call last):
File "install.py", line 497, in
install_bench(args)
File "install.py", line 278, in install_bench
run_playbook('site.yml', sudo=True, extra_vars=extra_vars)
File "install.py", line 413, in run_playbook
success = subprocess.check_call(args, cwd=playbooks_folder, stdout=log_stream, stderr=sys.stderr)
File "/usr/lib/python3.8/subprocess.py", line 364, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['ansible-playbook', '-c', 'local', 'site.yml', '-vvvv', '-e', '@/tmp/extra_vars.json', '--become', '--become-user=frappe']' returned non-zero exit status 2.
| True | ERPNEXT 13 Installation Easy install error URGENT !!!! - I have tried to install ERPnext 13 more than 20 times , using all the tutorials but no good result ,
when I use Easy install [ sudo python3 install.py --verbose --production --user [USER] --mariadb-version 10.5 --frappe-branch version-13 --erpnext-branch version-13 ]
I got error , return none zero status 2
Could you please help me out
the whole result
fatal: [localhost]: FAILED! => {
"changed": true,
"cmd": [
"bench",
"init",
"/home/frappe/frappe-bench",
"--frappe-path",
"https://github.com/frappe/frappe",
"--frappe-branch",
"version-13",
"--python",
"python3"
],
"delta": "0:00:00.378029",
"end": "2022-02-20 10:45:29.953302",
"invocation": {
"module_args": {
"_raw_params": "bench init /home/frappe/frappe-bench --frappe-path https://github.com/frappe/frappe --frappe-branch version-13 --python python3",
"_uses_shell": false,
"argv": null,
"chdir": null,
"creates": "/home/frappe/frappe-bench",
"executable": null,
"removes": null,
"stdin": null,
"stdin_add_newline": true,
"strip_empty_ends": true,
"warn": true
}
},
"msg": "non-zero return code",
"rc": 1,
"start": "2022-02-20 10:45:29.575273",
"stderr": "Traceback (most recent call last):\n File "/usr/bin/bench", line 33, in \n sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())\n File "/usr/bin/bench", line 22, in importlib_load_entry_point\n for entry_point in distribution(dist_name).entry_points\n File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution\n return Distribution.from_name(distribution_name)\n File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name\n raise PackageNotFoundError(name)\nimportlib.metadata.PackageNotFoundError: frappe-bench",
"stderr_lines": [
"Traceback (most recent call last):",
" File "/usr/bin/bench", line 33, in ",
" sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())",
" File "/usr/bin/bench", line 22, in importlib_load_entry_point",
" for entry_point in distribution(dist_name).entry_points",
" File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution",
" return Distribution.from_name(distribution_name)",
" File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name",
" raise PackageNotFoundError(name)",
"importlib.metadata.PackageNotFoundError: frappe-bench"
],
"stdout": "",
"stdout_lines": []
}
PLAY RECAP *****************************************************************************************************************************************************************
localhost : ok=67 changed=17 unreachable=0 failed=1 skipped=62 rescued=0 ignored=0
Traceback (most recent call last):
File "install.py", line 497, in
install_bench(args)
File "install.py", line 278, in install_bench
run_playbook('site.yml', sudo=True, extra_vars=extra_vars)
File "install.py", line 413, in run_playbook
success = subprocess.check_call(args, cwd=playbooks_folder, stdout=log_stream, stderr=sys.stderr)
File "/usr/lib/python3.8/subprocess.py", line 364, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['ansible-playbook', '-c', 'local', 'site.yml', '-vvvv', '-e', '@/tmp/extra_vars.json', '--become', '--become-user=frappe']' returned non-zero exit status 2.
| main | erpnext installation easy install error urgent i have tried to install erpnext more than times using all the tutorials but no good result when i use easy install mariadb version frappe branch version erpnext branch version i got error return none zero status could you please help me out the whole result fatal failed changed true cmd bench init home frappe frappe bench frappe path frappe branch version python delta end invocation module args raw params bench init home frappe frappe bench frappe path frappe branch version python uses shell false argv null chdir null creates home frappe frappe bench executable null removes null stdin null stdin add newline true strip empty ends true warn true msg non zero return code rc start stderr traceback most recent call last n file usr bin bench line in n sys exit load entry point frappe bench console scripts bench n file usr bin bench line in importlib load entry point n for entry point in distribution dist name entry points n file usr lib importlib metadata py line in distribution n return distribution from name distribution name n file usr lib importlib metadata py line in from name n raise packagenotfounderror name nimportlib metadata packagenotfounderror frappe bench stderr lines traceback most recent call last file usr bin bench line in sys exit load entry point frappe bench console scripts bench file usr bin bench line in importlib load entry point for entry point in distribution dist name entry points file usr lib importlib metadata py line in distribution return distribution from name distribution name file usr lib importlib metadata py line in from name raise packagenotfounderror name importlib metadata packagenotfounderror frappe bench stdout stdout lines play recap localhost ok changed unreachable failed skipped rescued ignored traceback most recent call last file install py line in install bench args file install py line in install bench run playbook site yml sudo true extra vars extra vars file install py line in run playbook success subprocess check call args cwd playbooks folder stdout log stream stderr sys stderr file usr lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status | 1 |
60,407 | 14,543,819,136 | IssuesEvent | 2020-12-15 17:20:58 | bitbar/testdroid-api | https://api.github.com/repos/bitbar/testdroid-api | opened | WS-2019-0379 (Medium) detected in commons-codec-1.10.jar | security vulnerability | ## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: testdroid-api/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- httpclient-4.5.6.jar (Root Library)
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bitbar/testdroid-api/commit/e27a23c3fc46b3858353a24f6550013c097bb0c4">e27a23c3fc46b3858353a24f6550013c097bb0c4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version βcommons-codec-1.13-RC1β is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-12</p>
<p>Fix Resolution: 1.13-RC1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.10","isTransitiveDependency":true,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.6;commons-codec:commons-codec:1.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version βcommons-codec-1.13-RC1β is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | WS-2019-0379 (Medium) detected in commons-codec-1.10.jar - ## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: testdroid-api/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- httpclient-4.5.6.jar (Root Library)
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bitbar/testdroid-api/commit/e27a23c3fc46b3858353a24f6550013c097bb0c4">e27a23c3fc46b3858353a24f6550013c097bb0c4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version βcommons-codec-1.13-RC1β is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-12</p>
<p>Fix Resolution: 1.13-RC1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.10","isTransitiveDependency":true,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.6;commons-codec:commons-codec:1.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version βcommons-codec-1.13-RC1β is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_main | ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file testdroid api pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar dependency hierarchy httpclient jar root library x commons codec jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons codec before version βcommons codec β is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails apache commons codec before version βcommons codec β is vulnerable to information disclosure due to improper input validation vulnerabilityurl | 0 |
2,996 | 10,885,066,939 | IssuesEvent | 2019-11-18 09:39:46 | chocolatey-community/chocolatey-package-requests | https://api.github.com/repos/chocolatey-community/chocolatey-package-requests | closed | RFP - AudioFlyout | Status: Available For Maintainer(s) | <!--
* Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search.
* Please ensure there is no existing open package request.
* Please ensure the issue title starts with 'RFP - ' - for example 'RFP - Adobe Reader'
* Please ensure you have both the Software Project URL and the Software Download URL before continuing.
NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication
PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM.
-->
## Checklist
- [X] The package I am requesting does not already exist on https://chocolatey.org/packages;
- [X] There is no open issue for this package;
- [X] The issue title starts 'RFP - ';
- [X] The download URL is public and not locked behind a paywall / login;
## Package Details
Software project URL : https://github.com/ADeltaX/AudioFlyout
Direct download URL for the software / installer :
https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyout.zip
https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyoutUA.zip
Software summary / short description:
Replace the Volume/SMTC UI with a custom one. fluent design style and more available additional controls.
<!-- ## Package Expectations
Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop
--> | True | RFP - AudioFlyout - <!--
* Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search.
* Please ensure there is no existing open package request.
* Please ensure the issue title starts with 'RFP - ' - for example 'RFP - Adobe Reader'
* Please ensure you have both the Software Project URL and the Software Download URL before continuing.
NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication
PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM.
-->
## Checklist
- [X] The package I am requesting does not already exist on https://chocolatey.org/packages;
- [X] There is no open issue for this package;
- [X] The issue title starts 'RFP - ';
- [X] The download URL is public and not locked behind a paywall / login;
## Package Details
Software project URL : https://github.com/ADeltaX/AudioFlyout
Direct download URL for the software / installer :
https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyout.zip
https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyoutUA.zip
Software summary / short description:
Replace the Volume/SMTC UI with a custom one. fluent design style and more available additional controls.
<!-- ## Package Expectations
Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop
--> | main | rfp audioflyout please ensure the package does not already exist in the chocolatey community repository by using a relevant search please ensure there is no existing open package request please ensure the issue title starts with rfp for example rfp adobe reader please ensure you have both the software project url and the software download url before continuing note keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the chocolatey community please remove all comments once you have read them checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description replace the volume smtc ui with a custom one fluent design style and more available additional controls package expectations here you can make suggestions on what you would expect the package to do outside of installing eg adding icons to the desktop | 1 |
3,669 | 14,997,077,640 | IssuesEvent | 2021-01-29 16:25:07 | exercism/python | https://api.github.com/repos/exercism/python | opened | [CI] Update tooling for v3 | maintainer action required:grey_exclamation: | ## Changed components
- Exercises are no longer in just one directory (exercises/). They are in either exercises/concept/ or exercises/practice/
- Example solutions are no longer in the exercise directory itself (two-fer/example.py). They have been/should be moved to .meta/exemplar.py (two-fer/.meta/exemplar.py)
- config.json's `exercises` field is no longer an array of exercises. It is an object with two properties: `concept` and `practice`. These each have the same format as the old `exercises` property.
- Exercise READMEs are no longer stored as a single, generated file in track repos. They can be removed from exercises if present
## Scripts dependent on changes components:
- [ ] bin/template_status.py
- [ ] bin/generate_tests.py
- [ ] test/check-exercises.py
## Obsolete scripts
The following scripts are no longer needed:
- [ ] bin/check-readmes.sh
- [ ] bin/check-test-version.py (versions no longer used in problem-specifications)
## Ambiguity
@ErikSchierboom, we have these scripts. Neither are needed for CI, but which (if either) should be kept for contributors/maintainers who need to run the new v3 configlet locally
- [ ] bin/fetch-configlet
- [ ] bin/fetch-canonical_data_syncer
| True | [CI] Update tooling for v3 - ## Changed components
- Exercises are no longer in just one directory (exercises/). They are in either exercises/concept/ or exercises/practice/
- Example solutions are no longer in the exercise directory itself (two-fer/example.py). They have been/should be moved to .meta/exemplar.py (two-fer/.meta/exemplar.py)
- config.json's `exercises` field is no longer an array of exercises. It is an object with two properties: `concept` and `practice`. These each have the same format as the old `exercises` property.
- Exercise READMEs are no longer stored as a single, generated file in track repos. They can be removed from exercises if present
## Scripts dependent on changes components:
- [ ] bin/template_status.py
- [ ] bin/generate_tests.py
- [ ] test/check-exercises.py
## Obsolete scripts
The following scripts are no longer needed:
- [ ] bin/check-readmes.sh
- [ ] bin/check-test-version.py (versions no longer used in problem-specifications)
## Ambiguity
@ErikSchierboom, we have these scripts. Neither are needed for CI, but which (if either) should be kept for contributors/maintainers who need to run the new v3 configlet locally
- [ ] bin/fetch-configlet
- [ ] bin/fetch-canonical_data_syncer
| main | update tooling for changed components exercises are no longer in just one directory exercises they are in either exercises concept or exercises practice example solutions are no longer in the exercise directory itself two fer example py they have been should be moved to meta exemplar py two fer meta exemplar py config json s exercises field is no longer an array of exercises it is an object with two properties concept and practice these each have the same format as the old exercises property exercise readmes are no longer stored as a single generated file in track repos they can be removed from exercises if present scripts dependent on changes components bin template status py bin generate tests py test check exercises py obsolete scripts the following scripts are no longer needed bin check readmes sh bin check test version py versions no longer used in problem specifications ambiguity erikschierboom we have these scripts neither are needed for ci but which if either should be kept for contributors maintainers who need to run the new configlet locally bin fetch configlet bin fetch canonical data syncer | 1 |
38,616 | 8,517,113,048 | IssuesEvent | 2018-11-01 06:28:56 | virtual-labs/colloid-and-surface-chemistry-iiith | https://api.github.com/repos/virtual-labs/colloid-and-surface-chemistry-iiith | opened | Missing vendor-prefixed CSS gradients for Webkit (Safari 5+, Chrome), Opera 11.1+. | 2018-Open static-code-analysis | CSS gradients in a cross-browser way requires using many different vendor-prefixed versions. There are currently five different vendor-prefixed versions of CSS gradient:
-ms-linear-gradient and -ms-radial-gradient for Internet Explorer 10+
-moz-linear-gradient and -moz-radial-gradient for Firefox 3.6+
-o-linear-gradient and -o-radial-gradient for Opera 11.10+
-webkit-linear-gradient and -webkit-radial-gradient for Safari 5+ and Chrome
-webkit-gradient for Safari 4+ and Chrome (aka "Old WebKit")
Meaning a simple two-color gradient that works across all browsers must look like this:
```
background: -moz-linear-gradient(...); /* FF3.6+ */
background: -webkit-gradient(...); /* Chrome,Safari4+ */
background: -webkit-linear-gradient(...); /* Chrome10+,Safari5.1+ */
background: -o-linear-gradient(...); /* Opera 11.10+ */
background: -ms-linear-gradient(...); /* IE10+ */
```
It's easy to forget one or more gradient definitions with all of the various vendor prefix gradients available.
**Please refer to the following link to fix similar issues.**
https://app.codacy.com/app/BSravanthi/colloid-and-surface-chemistry-iiith/issues?&filters=W3siaWQiOiJMYW5ndWFnZSIsInZhbHVlcyI6W251bGxdfSx7ImlkIjoiQ2F0ZWdvcnkiLCJ2YWx1ZXMiOlsiQ29tcGF0aWJpbGl0eSJdfSx7ImlkIjoiTGV2ZWwiLCJ2YWx1ZXMiOltudWxsXX0seyJpZCI6IlBhdHRlcm4iLCJ2YWx1ZXMiOltudWxsXX1d | 1.0 | Missing vendor-prefixed CSS gradients for Webkit (Safari 5+, Chrome), Opera 11.1+. - CSS gradients in a cross-browser way requires using many different vendor-prefixed versions. There are currently five different vendor-prefixed versions of CSS gradient:
-ms-linear-gradient and -ms-radial-gradient for Internet Explorer 10+
-moz-linear-gradient and -moz-radial-gradient for Firefox 3.6+
-o-linear-gradient and -o-radial-gradient for Opera 11.10+
-webkit-linear-gradient and -webkit-radial-gradient for Safari 5+ and Chrome
-webkit-gradient for Safari 4+ and Chrome (aka "Old WebKit")
Meaning a simple two-color gradient that works across all browsers must look like this:
```
background: -moz-linear-gradient(...); /* FF3.6+ */
background: -webkit-gradient(...); /* Chrome,Safari4+ */
background: -webkit-linear-gradient(...); /* Chrome10+,Safari5.1+ */
background: -o-linear-gradient(...); /* Opera 11.10+ */
background: -ms-linear-gradient(...); /* IE10+ */
```
It's easy to forget one or more gradient definitions with all of the various vendor prefix gradients available.
**Please refer to the following link to fix similar issues.**
https://app.codacy.com/app/BSravanthi/colloid-and-surface-chemistry-iiith/issues?&filters=W3siaWQiOiJMYW5ndWFnZSIsInZhbHVlcyI6W251bGxdfSx7ImlkIjoiQ2F0ZWdvcnkiLCJ2YWx1ZXMiOlsiQ29tcGF0aWJpbGl0eSJdfSx7ImlkIjoiTGV2ZWwiLCJ2YWx1ZXMiOltudWxsXX0seyJpZCI6IlBhdHRlcm4iLCJ2YWx1ZXMiOltudWxsXX1d | non_main | missing vendor prefixed css gradients for webkit safari chrome opera css gradients in a cross browser way requires using many different vendor prefixed versions there are currently five different vendor prefixed versions of css gradient ms linear gradient and ms radial gradient for internet explorer moz linear gradient and moz radial gradient for firefox o linear gradient and o radial gradient for opera webkit linear gradient and webkit radial gradient for safari and chrome webkit gradient for safari and chrome aka old webkit meaning a simple two color gradient that works across all browsers must look like this background moz linear gradient background webkit gradient chrome background webkit linear gradient background o linear gradient opera background ms linear gradient it s easy to forget one or more gradient definitions with all of the various vendor prefix gradients available please refer to the following link to fix similar issues | 0 |
24,754 | 17,692,928,112 | IssuesEvent | 2021-08-24 12:20:15 | tarantool/tarantool | https://api.github.com/repos/tarantool/tarantool | closed | new release policy: adjust download.tarantool.org infrastructure | infrastructure teamQ | The main discussion about the new release policy is [here](https://github.com/tarantool/tarantool/discussions/6182).
Since we [plan](https://github.com/tarantool/tarantool/issues/6185) to add repositories with a bit different layout (`series-2`, `pre-release`), we should adjust download.tarantool.org redirections and ensure that everything works good.
NB: Don't forget about source tarballs. | 1.0 | new release policy: adjust download.tarantool.org infrastructure - The main discussion about the new release policy is [here](https://github.com/tarantool/tarantool/discussions/6182).
Since we [plan](https://github.com/tarantool/tarantool/issues/6185) to add repositories with a bit different layout (`series-2`, `pre-release`), we should adjust download.tarantool.org redirections and ensure that everything works good.
NB: Don't forget about source tarballs. | non_main | new release policy adjust download tarantool org infrastructure the main discussion about the new release policy is since we to add repositories with a bit different layout series pre release we should adjust download tarantool org redirections and ensure that everything works good nb don t forget about source tarballs | 0 |
635,851 | 20,510,795,378 | IssuesEvent | 2022-03-01 06:08:04 | AY2122S2-CS2103T-T13-1/tp | https://api.github.com/repos/AY2122S2-CS2103T-T13-1/tp | opened | As a pet daycare owner I can get a list of pets which will be staying overnight in the daycare | type.Story priority.Medium | ... so that I can arrange the necessary manpower support required. | 1.0 | As a pet daycare owner I can get a list of pets which will be staying overnight in the daycare - ... so that I can arrange the necessary manpower support required. | non_main | as a pet daycare owner i can get a list of pets which will be staying overnight in the daycare so that i can arrange the necessary manpower support required | 0 |
157,506 | 13,691,125,176 | IssuesEvent | 2020-09-30 15:10:41 | codesankalp/dsalgo | https://api.github.com/repos/codesankalp/dsalgo | opened | PR Protocol | documentation enhancement | required submission protocols for pull request.
type: google doc, word , ppt
must include: format for pull request, test,quality assurance and other neccessary checks | 1.0 | PR Protocol - required submission protocols for pull request.
type: google doc, word , ppt
must include: format for pull request, test,quality assurance and other neccessary checks | non_main | pr protocol required submission protocols for pull request type google doc word ppt must include format for pull request test quality assurance and other neccessary checks | 0 |
299,678 | 22,617,917,411 | IssuesEvent | 2022-06-30 01:22:41 | sonr-io/sonr | https://api.github.com/repos/sonr-io/sonr | closed | Integrate Stripe for Client Side Payment | documentation | We need to do some R&D and handle Fiat to ATOM to SNR conversion. Some questions that need to be answered:
* How do Coins operate in the Cosmos Network?
* Do we need to have an ATOM -> SNR conversion?
* Can we have a direct Fiat -> SNR NFT purchase?
https://dashboard.moonpay.com/getting\_started
βIssue is synchronized with this [Asana task](https://app.asana.com/0/1202528620406598/1202529507214139) by [Unito](https://www.unito.io)
| 1.0 | Integrate Stripe for Client Side Payment - We need to do some R&D and handle Fiat to ATOM to SNR conversion. Some questions that need to be answered:
* How do Coins operate in the Cosmos Network?
* Do we need to have an ATOM -> SNR conversion?
* Can we have a direct Fiat -> SNR NFT purchase?
https://dashboard.moonpay.com/getting\_started
βIssue is synchronized with this [Asana task](https://app.asana.com/0/1202528620406598/1202529507214139) by [Unito](https://www.unito.io)
| non_main | integrate stripe for client side payment we need to do some r d and handle fiat to atom to snr conversion some questions that need to be answered how do coins operate in the cosmos network do we need to have an atom snr conversion can we have a direct fiat snr nft purchase βissue is synchronized with this by | 0 |
528,068 | 15,359,403,350 | IssuesEvent | 2021-03-01 15:49:06 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.pornhubpremium.com - video or audio doesn't play | browser-firefox-mobile browser-firefox-reality engine-gecko priority-normal type-webvr | <!-- @browser: Firefox Mobile 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/67701 -->
**URL**: https://www.pornhubpremium.com/
**Browser / Version**: Firefox Mobile 81.0
**Operating System**: Android 7.1.1
**Tested Another Browser**: Yes Other
**Problem type**: Video or audio doesn't play
**Description**: There is no video
**Steps to Reproduce**:
oculus questfirefox reality when you try to watch something in webr its just a black screen
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with β€οΈ_ | 1.0 | www.pornhubpremium.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/67701 -->
**URL**: https://www.pornhubpremium.com/
**Browser / Version**: Firefox Mobile 81.0
**Operating System**: Android 7.1.1
**Tested Another Browser**: Yes Other
**Problem type**: Video or audio doesn't play
**Description**: There is no video
**Steps to Reproduce**:
oculus questfirefox reality when you try to watch something in webr its just a black screen
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with β€οΈ_ | non_main | video or audio doesn t play url browser version firefox mobile operating system android tested another browser yes other problem type video or audio doesn t play description there is no video steps to reproduce oculus questfirefox reality when you try to watch something in webr its just a black screen browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with β€οΈ | 0 |
390,102 | 11,524,808,568 | IssuesEvent | 2020-02-15 03:06:23 | HW-PlayersPatch/Development | https://api.github.com/repos/HW-PlayersPatch/Development | opened | Salvage Crash Fix | Priority3: Low Status2: Research Needed Type1: Feature Type2: Bug | Build 11: "When salvage corvettes in a formation are given multiple different salvage targets, they often crash the game. To prevent this crash, salvage corvettes will no longer be able to join formations/strikegroups."
https://github.com/HW-PlayersPatch/Development/commit/3b66a70e9b40f5a3b67c48d6b99f339d8561afba
Try to make this an option. | 1.0 | Salvage Crash Fix - Build 11: "When salvage corvettes in a formation are given multiple different salvage targets, they often crash the game. To prevent this crash, salvage corvettes will no longer be able to join formations/strikegroups."
https://github.com/HW-PlayersPatch/Development/commit/3b66a70e9b40f5a3b67c48d6b99f339d8561afba
Try to make this an option. | non_main | salvage crash fix build when salvage corvettes in a formation are given multiple different salvage targets they often crash the game to prevent this crash salvage corvettes will no longer be able to join formations strikegroups try to make this an option | 0 |
4,832 | 24,910,996,393 | IssuesEvent | 2022-10-29 21:24:07 | deislabs/spiderlightning | https://api.github.com/repos/deislabs/spiderlightning | closed | Bin releases of `slight` | π§ maintainer issue | We should start building and publishing releases for `slight`, so that folks don't need to build bins themselves.
We need to add a release action that is triggered by a semantic version tag that would build `slight` bins targeting the major OS/Arch combinations. | True | Bin releases of `slight` - We should start building and publishing releases for `slight`, so that folks don't need to build bins themselves.
We need to add a release action that is triggered by a semantic version tag that would build `slight` bins targeting the major OS/Arch combinations. | main | bin releases of slight we should start building and publishing releases for slight so that folks don t need to build bins themselves we need to add a release action that is triggered by a semantic version tag that would build slight bins targeting the major os arch combinations | 1 |
88,021 | 11,018,306,658 | IssuesEvent | 2019-12-05 10:15:10 | statsmodels/statsmodels | https://api.github.com/repos/statsmodels/statsmodels | closed | use httplib2 for caching downloaded data files ? | design | Skipper recommended https://groups.google.com/group/pystatsmodels/browse_thread/thread/5bfe1e45b3336765?hl=en
http://pypi.python.org/pypi/httplib2/
80,000 downloads looks pretty good
| 1.0 | use httplib2 for caching downloaded data files ? - Skipper recommended https://groups.google.com/group/pystatsmodels/browse_thread/thread/5bfe1e45b3336765?hl=en
http://pypi.python.org/pypi/httplib2/
80,000 downloads looks pretty good
| non_main | use for caching downloaded data files skipper recommended downloads looks pretty good | 0 |
166,141 | 12,891,859,896 | IssuesEvent | 2020-07-13 18:30:47 | ReactiveX/RxJava | https://api.github.com/repos/ReactiveX/RxJava | closed | 3.x: Flaky GroupBy test | 3.x PR welcome Test-Failures good first issue | https://github.com/ReactiveX/RxJava/blob/98acac218cdb04d279b5ac49bb1afc65bc6ec4fe/src/test/java/io/reactivex/rxjava3/internal/operators/flowable/FlowableGroupByTest.java#L2668
```
io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest > issue6974Part2Case1NoEvict FAILED
java.lang.AssertionError: Error(s) present: [io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred. ] (latch = 0, values = 15551, errors = 1, completions = 0)
at io.reactivex.rxjava3.observers.BaseTestConsumer.fail(BaseTestConsumer.java:125)
at io.reactivex.rxjava3.observers.BaseTestConsumer.assertNoErrors(BaseTestConsumer.java:212)
at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974RunPart2NoEvict(FlowableGroupByTest.java:2681)
at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974Part2Case1NoEvict(FlowableGroupByTest.java:2693)
Caused by:
io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred.
Caused by:
io.reactivex.rxjava3.exceptions.MissingBackpressureException: Unable to emit a new group (#71) due to lack of requests. Please make sure the downstream can always accept a new group as well as each group is consumed in order for the whole operator to be able to proceed.
```
Error is an allowed outcome here. | 1.0 | 3.x: Flaky GroupBy test - https://github.com/ReactiveX/RxJava/blob/98acac218cdb04d279b5ac49bb1afc65bc6ec4fe/src/test/java/io/reactivex/rxjava3/internal/operators/flowable/FlowableGroupByTest.java#L2668
```
io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest > issue6974Part2Case1NoEvict FAILED
java.lang.AssertionError: Error(s) present: [io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred. ] (latch = 0, values = 15551, errors = 1, completions = 0)
at io.reactivex.rxjava3.observers.BaseTestConsumer.fail(BaseTestConsumer.java:125)
at io.reactivex.rxjava3.observers.BaseTestConsumer.assertNoErrors(BaseTestConsumer.java:212)
at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974RunPart2NoEvict(FlowableGroupByTest.java:2681)
at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974Part2Case1NoEvict(FlowableGroupByTest.java:2693)
Caused by:
io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred.
Caused by:
io.reactivex.rxjava3.exceptions.MissingBackpressureException: Unable to emit a new group (#71) due to lack of requests. Please make sure the downstream can always accept a new group as well as each group is consumed in order for the whole operator to be able to proceed.
```
Error is an allowed outcome here. | non_main | x flaky groupby test io reactivex internal operators flowable flowablegroupbytest failed java lang assertionerror error s present latch values errors completions at io reactivex observers basetestconsumer fail basetestconsumer java at io reactivex observers basetestconsumer assertnoerrors basetestconsumer java at io reactivex internal operators flowable flowablegroupbytest flowablegroupbytest java at io reactivex internal operators flowable flowablegroupbytest flowablegroupbytest java caused by io reactivex exceptions compositeexception exceptions occurred caused by io reactivex exceptions missingbackpressureexception unable to emit a new group due to lack of requests please make sure the downstream can always accept a new group as well as each group is consumed in order for the whole operator to be able to proceed error is an allowed outcome here | 0 |
2,679 | 9,219,081,694 | IssuesEvent | 2019-03-11 14:44:39 | precice/precice | https://api.github.com/repos/precice/precice | closed | Integrate a testing coverage tool | maintainability | Since we have quite some tests in preCICE, it would be useful to also have a testing coverage tool to show us where tests are missing etc. It would also be a nice code quality metric which we can track over time.
A common tool that several projects use is [code coverage](https://codecov.io/) (e.g. see [Spack](https://codecov.io/gh/spack/spack)). Any other ideas/suggestions? | True | Integrate a testing coverage tool - Since we have quite some tests in preCICE, it would be useful to also have a testing coverage tool to show us where tests are missing etc. It would also be a nice code quality metric which we can track over time.
A common tool that several projects use is [code coverage](https://codecov.io/) (e.g. see [Spack](https://codecov.io/gh/spack/spack)). Any other ideas/suggestions? | main | integrate a testing coverage tool since we have quite some tests in precice it would be useful to also have a testing coverage tool to show us where tests are missing etc it would also be a nice code quality metric which we can track over time a common tool that several projects use is e g see any other ideas suggestions | 1 |
5,526 | 27,632,268,436 | IssuesEvent | 2023-03-10 11:49:27 | jesus2099/konami-command | https://api.github.com/repos/jesus2099/konami-command | opened | Retrieve available localised texts from MBS page | ninja mb_POWER-VOTE minor maintainability | Instead of hardcoding _Yes, No, Abstain, No vote, Approve,_ etc., in several languages, fetch these text from current page. | True | Retrieve available localised texts from MBS page - Instead of hardcoding _Yes, No, Abstain, No vote, Approve,_ etc., in several languages, fetch these text from current page. | main | retrieve available localised texts from mbs page instead of hardcoding yes no abstain no vote approve etc in several languages fetch these text from current page | 1 |
26,387 | 12,404,876,892 | IssuesEvent | 2020-05-21 16:16:37 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | opened | TIA Customer Case Management: Customer Permit Details Page | Need: 1-Must Have Product: TIA Module Service: Apps Status: Done Type: Feature Workgroup: TDSD imported-from-csv | As a TIA Customer I'd like a case details page that allows me to see basic information about my case, especially the current status. | 1.0 | TIA Customer Case Management: Customer Permit Details Page - As a TIA Customer I'd like a case details page that allows me to see basic information about my case, especially the current status. | non_main | tia customer case management customer permit details page as a tia customer i d like a case details page that allows me to see basic information about my case especially the current status | 0 |
493,593 | 14,235,366,822 | IssuesEvent | 2020-11-18 14:44:50 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | opened | Scheduler's navigate event does not trigger on mobile device | Bug C: Scheduler Kendo2 Priority 5 SEV: Medium Touch Devices | ### Bug report
The navigate event does not trigger when changing the view on a mobile device.
The event triggers back in version 2019.1.220.
### Reproduction of the problem
1. Open [this example](https://dojo.telerik.com/OSOCOgEm) on a mobile device.
2. Change the View.
### Current behavior
The navigate event does not trigger when changing the view.
### Expected/desired behavior
The navigate event should trigger.
### Environment
* **Kendo UI version:** 2020.3.1118
* **Browser:** [all]
| 1.0 | Scheduler's navigate event does not trigger on mobile device - ### Bug report
The navigate event does not trigger when changing the view on a mobile device.
The event triggers back in version 2019.1.220.
### Reproduction of the problem
1. Open [this example](https://dojo.telerik.com/OSOCOgEm) on a mobile device.
2. Change the View.
### Current behavior
The navigate event does not trigger when changing the view.
### Expected/desired behavior
The navigate event should trigger.
### Environment
* **Kendo UI version:** 2020.3.1118
* **Browser:** [all]
| non_main | scheduler s navigate event does not trigger on mobile device bug report the navigate event does not trigger when changing the view on a mobile device the event triggers back in version reproduction of the problem open on a mobile device change the view current behavior the navigate event does not trigger when changing the view expected desired behavior the navigate event should trigger environment kendo ui version browser | 0 |
30,432 | 11,825,715,288 | IssuesEvent | 2020-03-21 14:21:50 | stefanfreitag/s3_yum_repository_slides | https://api.github.com/repos/stefanfreitag/s3_yum_repository_slides | opened | CVE-2015-9251 (Medium) detected in jquery-1.7.2.min.js | security vulnerability | ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/s3_yum_repository_slides/node_modules/js-base64/test/index.html</p>
<p>Path to vulnerable library: /s3_yum_repository_slides/node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/s3_yum_repository_slides/commit/a3e7d6e421e87c01267600c253daad2de918d386">a3e7d6e421e87c01267600c253daad2de918d386</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2015-9251 (Medium) detected in jquery-1.7.2.min.js - ## CVE-2015-9251 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/s3_yum_repository_slides/node_modules/js-base64/test/index.html</p>
<p>Path to vulnerable library: /s3_yum_repository_slides/node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.2.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/s3_yum_repository_slides/commit/a3e7d6e421e87c01267600c253daad2de918d386">a3e7d6e421e87c01267600c253daad2de918d386</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm yum repository slides node modules js test index html path to vulnerable library yum repository slides node modules js test index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
182,572 | 14,917,998,820 | IssuesEvent | 2021-01-22 20:51:48 | ZupIT/ritchie-cli | https://api.github.com/repos/ZupIT/ritchie-cli | closed | Add Bitbucket to provider list for rit add repo | :books: documentation :hammer: improvement :heavy_check_mark: refined :sparkles: feature Hacktoberfest | ## What would you like to see
- I would like to be able to import a bitbucket repo on Ritchie through the `rit add repo` command.
## Why is it needed
- Bitbucket is not supported yet, only Github and Gitlab. | 1.0 | Add Bitbucket to provider list for rit add repo - ## What would you like to see
- I would like to be able to import a bitbucket repo on Ritchie through the `rit add repo` command.
## Why is it needed
- Bitbucket is not supported yet, only Github and Gitlab. | non_main | add bitbucket to provider list for rit add repo what would you like to see i would like to be able to import a bitbucket repo on ritchie through the rit add repo command why is it needed bitbucket is not supported yet only github and gitlab | 0 |
99,438 | 8,700,535,103 | IssuesEvent | 2018-12-05 09:02:03 | SME-Issues/issues | https://api.github.com/repos/SME-Issues/issues | closed | General Comprehension None Tests - 04/12/2018 - 5004 | NLP Api pulse_tests | **General Comprehension None Tests**
- Total: 23
- Passed: 17
- **Pass: 17 (77%)**
- Not Understood: 0
- Error (not understood): 1
- Failed but Understood: 5 (23%)
| 1.0 | General Comprehension None Tests - 04/12/2018 - 5004 - **General Comprehension None Tests**
- Total: 23
- Passed: 17
- **Pass: 17 (77%)**
- Not Understood: 0
- Error (not understood): 1
- Failed but Understood: 5 (23%)
| non_main | general comprehension none tests general comprehension none tests total passed pass not understood error not understood failed but understood | 0 |
1,395 | 6,025,335,501 | IssuesEvent | 2017-06-08 08:25:53 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_iis_webapplication: slashes in physical path are not converted to windows style | affects_2.3 bug_report waiting_on_maintainer windows | ##### ISSUE TYPE
Bug Report
##### COMPONENT NAME
win_iis_webapplication module
##### ANSIBLE VERSION
N/A
##### SUMMARY
win_iis_website converts them. Example:
vars somewhere:
```
path: c:/sites/https/develop
```
playbook:
```
- name: create website
win_iis_website:
name: "{{ name }}"
state: started
application_pool: "{{ appdomain }}"
physical_path: "{{ path }}/web"
- name: create rest virtual folder
win_iis_webapplication:
name: rest
state: present
site: "{{ name }}"
physical_path: "{{ path }}/api"
application_pool: "{{ appdomain }}"
```
Website will work ok, webapplication will always fail with 404.
| True | win_iis_webapplication: slashes in physical path are not converted to windows style - ##### ISSUE TYPE
Bug Report
##### COMPONENT NAME
win_iis_webapplication module
##### ANSIBLE VERSION
N/A
##### SUMMARY
win_iis_website converts them. Example:
vars somewhere:
```
path: c:/sites/https/develop
```
playbook:
```
- name: create website
win_iis_website:
name: "{{ name }}"
state: started
application_pool: "{{ appdomain }}"
physical_path: "{{ path }}/web"
- name: create rest virtual folder
win_iis_webapplication:
name: rest
state: present
site: "{{ name }}"
physical_path: "{{ path }}/api"
application_pool: "{{ appdomain }}"
```
Website will work ok, webapplication will always fail with 404.
| main | win iis webapplication slashes in physical path are not converted to windows style issue type bug report component name win iis webapplication module ansible version n a summary win iis website converts them example vars somewhere path c sites https develop playbook name create website win iis website name name state started application pool appdomain physical path path web name create rest virtual folder win iis webapplication name rest state present site name physical path path api application pool appdomain website will work ok webapplication will always fail with | 1 |
247,133 | 18,857,342,416 | IssuesEvent | 2021-11-12 08:29:05 | chongjunwei/pe | https://api.github.com/repos/chongjunwei/pe | opened | Command message ambiguous | severity.Low type.DocumentationBug | 
Instead of "show all modules", perhaps "show students from all the modules" or something similar would be clearer. The current wording is slightly ambiguous and may be confused with the existing "lsmod" command, which shows a list of all the modules.
<!--session: 1636703307298-1218c5c4-1eba-4ef0-95db-9ff00e3700d4-->
<!--Version: Web v3.4.1--> | 1.0 | Command message ambiguous - 
Instead of "show all modules", perhaps "show students from all the modules" or something similar would be clearer. The current wording is slightly ambiguous and may be confused with the existing "lsmod" command, which shows a list of all the modules.
<!--session: 1636703307298-1218c5c4-1eba-4ef0-95db-9ff00e3700d4-->
<!--Version: Web v3.4.1--> | non_main | command message ambiguous instead of show all modules perhaps show students from all the modules or something similar would be clearer the current wording is slightly ambiguous and may be confused with the existing lsmod command which shows a list of all the modules | 0 |
9,454 | 8,639,227,120 | IssuesEvent | 2018-11-23 17:45:18 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Add information about "Diagnose and solve problems" experience | app-service-web/svc assigned-to-author doc-enhancement triaged | We have a great new experience, but not covered in a documentation
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: a14de731-a265-1737-4580-67ca6dede020
* Version Independent ID: a275bb18-dedc-d41c-3541-224b42487b52
* Content: [Slow web app performance in App Service](https://docs.microsoft.com/en-us/azure/app-service/app-service-web-troubleshoot-performance-degradation)
* Content Source: [articles/app-service/app-service-web-troubleshoot-performance-degradation.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-web-troubleshoot-performance-degradation.md)
* Service: **app-service-web**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | 1.0 | Add information about "Diagnose and solve problems" experience - We have a great new experience, but not covered in a documentation
---
#### Document Details
β *Do not edit this section. It is required for docs.microsoft.com β GitHub issue linking.*
* ID: a14de731-a265-1737-4580-67ca6dede020
* Version Independent ID: a275bb18-dedc-d41c-3541-224b42487b52
* Content: [Slow web app performance in App Service](https://docs.microsoft.com/en-us/azure/app-service/app-service-web-troubleshoot-performance-degradation)
* Content Source: [articles/app-service/app-service-web-troubleshoot-performance-degradation.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-web-troubleshoot-performance-degradation.md)
* Service: **app-service-web**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | non_main | add information about diagnose and solve problems experience we have a great new experience but not covered in a documentation document details β do not edit this section it is required for docs microsoft com β github issue linking id version independent id dedc content content source service app service web github login cephalin microsoft alias cephalin | 0 |
4,869 | 25,020,305,365 | IssuesEvent | 2022-11-03 23:27:56 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | "Auth" property do not work with "AWS::Include" DefinitionBody | area/resource/api area/intrinsics maintainer/need-response | **Description:**
We need to use some cloudformation functions in our swagger. So the "AWS::Include" transform is used in the DefnitionBody. I tried to move the authorizer declaration out of the swagger using the new "Auth" property. But I received an error of "Unable to add Auth configuration because 'DefinitionBody' does not contain a valid Swagger". Also, I notice the authorizer caching TTL is not an attribute in the "Auth" property?
** Template snippet:**
```
Type: AWS::Serverless::Api
Properties:
StageName: !Ref "Environment"
EndpointConfiguration: REGIONAL
Auth:
Authorizers:
RequestAuth:
FunctionPayloadType: REQUEST
FunctionArn: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerArn"]
FunctionInvokeRole: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerCredentials"]
Identity:
Headers:
- Authorization
Context:
- httpMethod
- path
ReauthorizeEvery: 0 # OPTIONAL; Service Default: 300
DefinitionBody:
'Fn::Transform':
Name: 'AWS::Include'
Parameters:
Location: s3://accolade-api-swaggers-test-577121982548/users.yml
```
| True | "Auth" property do not work with "AWS::Include" DefinitionBody - **Description:**
We need to use some cloudformation functions in our swagger. So the "AWS::Include" transform is used in the DefnitionBody. I tried to move the authorizer declaration out of the swagger using the new "Auth" property. But I received an error of "Unable to add Auth configuration because 'DefinitionBody' does not contain a valid Swagger". Also, I notice the authorizer caching TTL is not an attribute in the "Auth" property?
** Template snippet:**
```
Type: AWS::Serverless::Api
Properties:
StageName: !Ref "Environment"
EndpointConfiguration: REGIONAL
Auth:
Authorizers:
RequestAuth:
FunctionPayloadType: REQUEST
FunctionArn: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerArn"]
FunctionInvokeRole: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerCredentials"]
Identity:
Headers:
- Authorization
Context:
- httpMethod
- path
ReauthorizeEvery: 0 # OPTIONAL; Service Default: 300
DefinitionBody:
'Fn::Transform':
Name: 'AWS::Include'
Parameters:
Location: s3://accolade-api-swaggers-test-577121982548/users.yml
```
| main | auth property do not work with aws include definitionbody description we need to use some cloudformation functions in our swagger so the aws include transform is used in the defnitionbody i tried to move the authorizer declaration out of the swagger using the new auth property but i received an error of unable to add auth configuration because definitionbody does not contain a valid swagger also i notice the authorizer caching ttl is not an attribute in the auth property template snippet type aws serverless api properties stagename ref environment endpointconfiguration regional auth authorizers requestauth functionpayloadtype request functionarn findinmap functioninvokerole findinmap identity headers authorization context httpmethod path reauthorizeevery optional service default definitionbody fn transform name aws include parameters location accolade api swaggers test users yml | 1 |
255,452 | 21,926,005,845 | IssuesEvent | 2022-05-23 04:18:28 | stores-cedcommerce/Sungin-Internal--Feb21st-2022 | https://api.github.com/repos/stores-cedcommerce/Sungin-Internal--Feb21st-2022 | closed | Password Char length verification error issue. | Account pages Desktop Functional / bug Ready to test | Bug - Password char length verification showing different on create account page and on reset password page.
Exp - Password char length should be visible same in validation on both pages.
Ref Link - https://drive.google.com/file/d/1wQit2JLx4UtunlBWVkgX11DQce7qmysV/view | 1.0 | Password Char length verification error issue. - Bug - Password char length verification showing different on create account page and on reset password page.
Exp - Password char length should be visible same in validation on both pages.
Ref Link - https://drive.google.com/file/d/1wQit2JLx4UtunlBWVkgX11DQce7qmysV/view | non_main | password char length verification error issue bug password char length verification showing different on create account page and on reset password page exp password char length should be visible same in validation on both pages ref link | 0 |
91,529 | 10,722,606,139 | IssuesEvent | 2019-10-27 13:16:50 | riot/riot | https://api.github.com/repos/riot/riot | closed | [Question] How should I migrate mixins to riot 4? | documentation update required good for beginner | 5. How would you tag this issue?
- [x ] Question
I noticed mixins are absent from riot 4 documentation and from what I can tell the code, too. Are there any undocumented features, samples, or ideas for users migrating mixins from riot 3 to riot 4?
| 1.0 | [Question] How should I migrate mixins to riot 4? - 5. How would you tag this issue?
- [x ] Question
I noticed mixins are absent from riot 4 documentation and from what I can tell the code, too. Are there any undocumented features, samples, or ideas for users migrating mixins from riot 3 to riot 4?
| non_main | how should i migrate mixins to riot how would you tag this issue question i noticed mixins are absent from riot documentation and from what i can tell the code too are there any undocumented features samples or ideas for users migrating mixins from riot to riot | 0 |
854 | 4,513,365,927 | IssuesEvent | 2016-09-04 07:56:06 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | ec2_eni is not idempotent | aws bug_report cloud waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
- ec2_eni
##### ANSIBLE VERSION
- devel
##### SUMMARY
For the ec2_eni module to be idempotent on the eni creation, it should create the interface on the first execution and just return the interface with changed=False on the subsequent executions. The current behavior is to return an error on the subsequent executions.
##### STEPS TO REPRODUCE
```
$ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost
localhost | SUCCESS => {
"changed": true,
"interface": {
"description": "",
"groups": {
"sg-6b98a613": "default"
},
"id": "eni-51zzz311",
"mac_address": "0a:ff:2b:b7:0b:91",
"owner_id": "...",
"private_ip_address": "10.137.0.10",
"source_dest_check": true,
"status": "pending",
"subnet_id": "subnet-1b2c696d",
"vpc_id": "vpc-3ee6ff5a"
}
}
$ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost
localhost | FAILED! => {
"changed": false,
"failed": true,
"msg": "The specified address is already in use."
}
```
##### EXPECTED RESULTS
Second execution would return `changed:false` along with the `interface` data structure
| True | ec2_eni is not idempotent - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
- ec2_eni
##### ANSIBLE VERSION
- devel
##### SUMMARY
For the ec2_eni module to be idempotent on the eni creation, it should create the interface on the first execution and just return the interface with changed=False on the subsequent executions. The current behavior is to return an error on the subsequent executions.
##### STEPS TO REPRODUCE
```
$ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost
localhost | SUCCESS => {
"changed": true,
"interface": {
"description": "",
"groups": {
"sg-6b98a613": "default"
},
"id": "eni-51zzz311",
"mac_address": "0a:ff:2b:b7:0b:91",
"owner_id": "...",
"private_ip_address": "10.137.0.10",
"source_dest_check": true,
"status": "pending",
"subnet_id": "subnet-1b2c696d",
"vpc_id": "vpc-3ee6ff5a"
}
}
$ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost
localhost | FAILED! => {
"changed": false,
"failed": true,
"msg": "The specified address is already in use."
}
```
##### EXPECTED RESULTS
Second execution would return `changed:false` along with the `interface` data structure
| main | eni is not idempotent issue type bug report component name eni ansible version devel summary for the eni module to be idempotent on the eni creation it should create the interface on the first execution and just return the interface with changed false on the subsequent executions the current behavior is to return an error on the subsequent executions steps to reproduce ansible i localhost c local m eni a private ip address subnet id subnet zzzzz state present region us east localhost localhost success changed true interface description groups sg default id eni mac address ff owner id private ip address source dest check true status pending subnet id subnet vpc id vpc ansible i localhost c local m eni a private ip address subnet id subnet zzzzz state present region us east localhost localhost failed changed false failed true msg the specified address is already in use expected results second execution would return changed false along with the interface data structure | 1 |
166,569 | 12,962,315,342 | IssuesEvent | 2020-07-20 16:57:44 | longhorn/longhorn | https://api.github.com/repos/longhorn/longhorn | closed | Listing backups may fail during backup deletion | area/engine area/test bug reproduce/rare | ```
clients = {'longhorn-staging-tests-01': <longhorn.Client object at 0x7f0e589e2cd0>, 'longhorn-staging-tests-02': <longhorn.Client object at 0x7f0e58491e10>, 'longhorn-staging-tests-03': <longhorn.Client object at 0x7f0e589e2f50>}
volume_name = 'longhorn-testvol-b4pqea'
@pytest.mark.coretest # NOQA
def test_backup(clients, volume_name): # NOQA
> backup_test(clients, volume_name, SIZE)
test_basic.py:396:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
test_basic.py:430: in backup_test
backupstore_test(client, lht_hostId, volume_name, size)
test_basic.py:538: in backupstore_test
backups = bv.backupList().data
longhorn.py:248: in cb
*args, **kw)
longhorn.py:442: in action
return self._post_and_retry(url, *args, **kw)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <longhorn.Client object at 0x7f0e589e2f50>
url = 'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea?action=backupList'
args = (), kw = {}, retries = 3, last_error = None, i = 0
e = ApiError(ApiError(...), '500 : error listing backups for volume \'longhorn-tes... exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}')
def _post_and_retry(self, url, *args, **kw):
retries = kw.get('retries', 3)
last_error = None
for i in range(retries):
try:
return self._post(url, data=self._to_dict(*args, **kw))
except ApiError as e:
if e.error.code == 409:
last_error = e
time.sleep(.1)
else:
> raise e
E ApiError: (ApiError(...), '500 : error listing backups for volume \'longhorn-testvol-b4pqea\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\n404 \n\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\n AcceptRanges: \\"bytes\\",\\n Body: <nil>,\\n ContentLength: 511,\\n ContentType: \\"application/xml\\",\\n Metadata: {\\n\\n }\\n}" pkg=s3\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n"\n, error exit status 1\n{\'status\': 500, \'code\': 500, \'links\': {\'self\': u\'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea\'}, \'self\': <function cb at 0x7f0e58bd8230>, \'detail\': u\'\', \'actions\': {}, \'message\': u\'error listing backups for volume \\\'longhorn-testvol-b4pqea\\\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n\\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\\\n AcceptRanges: \\\\"bytes\\\\",\\\\n Body: <nil>,\\\\n ContentLength: 511,\\\\n ContentType: \\\\"application/xml\\\\",\\\\n Metadata: {\\\\n\\\\n }\\\\n}" pkg=s3\\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}')
```
When Longhorn tried to list the backup which was being deleted, this error may occur. | 1.0 | Listing backups may fail during backup deletion - ```
clients = {'longhorn-staging-tests-01': <longhorn.Client object at 0x7f0e589e2cd0>, 'longhorn-staging-tests-02': <longhorn.Client object at 0x7f0e58491e10>, 'longhorn-staging-tests-03': <longhorn.Client object at 0x7f0e589e2f50>}
volume_name = 'longhorn-testvol-b4pqea'
@pytest.mark.coretest # NOQA
def test_backup(clients, volume_name): # NOQA
> backup_test(clients, volume_name, SIZE)
test_basic.py:396:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
test_basic.py:430: in backup_test
backupstore_test(client, lht_hostId, volume_name, size)
test_basic.py:538: in backupstore_test
backups = bv.backupList().data
longhorn.py:248: in cb
*args, **kw)
longhorn.py:442: in action
return self._post_and_retry(url, *args, **kw)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <longhorn.Client object at 0x7f0e589e2f50>
url = 'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea?action=backupList'
args = (), kw = {}, retries = 3, last_error = None, i = 0
e = ApiError(ApiError(...), '500 : error listing backups for volume \'longhorn-tes... exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}')
def _post_and_retry(self, url, *args, **kw):
retries = kw.get('retries', 3)
last_error = None
for i in range(retries):
try:
return self._post(url, data=self._to_dict(*args, **kw))
except ApiError as e:
if e.error.code == 409:
last_error = e
time.sleep(.1)
else:
> raise e
E ApiError: (ApiError(...), '500 : error listing backups for volume \'longhorn-testvol-b4pqea\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\n404 \n\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\n AcceptRanges: \\"bytes\\",\\n Body: <nil>,\\n ContentLength: 511,\\n ContentType: \\"application/xml\\",\\n Metadata: {\\n\\n }\\n}" pkg=s3\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n"\n, error exit status 1\n{\'status\': 500, \'code\': 500, \'links\': {\'self\': u\'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea\'}, \'self\': <function cb at 0x7f0e58bd8230>, \'detail\': u\'\', \'actions\': {}, \'message\': u\'error listing backups for volume \\\'longhorn-testvol-b4pqea\\\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n\\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\\\n AcceptRanges: \\\\"bytes\\\\",\\\\n Body: <nil>,\\\\n ContentLength: 511,\\\\n ContentType: \\\\"application/xml\\\\",\\\\n Metadata: {\\\\n\\\\n }\\\\n}" pkg=s3\\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}')
```
When Longhorn tried to list the backup which was being deleted, this error may occur. | non_main | listing backups may fail during backup deletion clients longhorn staging tests longhorn staging tests longhorn staging tests volume name longhorn testvol pytest mark coretest noqa def test backup clients volume name noqa backup test clients volume name size test basic py test basic py in backup test backupstore test client lht hostid volume name size test basic py in backupstore test backups bv backuplist data longhorn py in cb args kw longhorn py in action return self post and retry url args kw self url args kw retries last error none i e apierror apierror error listing backups for volume longhorn tes exist n n error exit status type u error def post and retry self url args kw retries kw get retries last error none for i in range retries try return self post url data self to dict args kw except apierror as e if e error code last error e time sleep else raise e e apierror apierror error listing backups for volume longhorn testvol error listing backups failed to execute var lib rancher longhorn engine binaries longhornio longhorn engine staging longhorn output aws error nosuchkey the specified key does not exist n n stderr time level error msg n acceptranges bytes n body n contentlength n contenttype application xml n metadata n n n pkg ntime level error msg aws error nosuchkey the specified key does not exist n n error exit status n status code links self u self detail u actions message u error listing backups for volume longhorn testvol error listing backups failed to execute var lib rancher longhorn engine binaries longhornio longhorn engine staging longhorn output aws error nosuchkey the specified key does not exist n n stderr time level error msg n acceptranges bytes n body n contentlength n contenttype application xml n metadata n n n pkg ntime level error msg aws error nosuchkey the specified key does not exist n n error exit status type u error when longhorn tried to list the backup which was being deleted this error may occur | 0 |
753 | 4,351,730,564 | IssuesEvent | 2016-08-01 01:10:53 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | azure_rm_virtualmachine module fails creating a virtualmachine when the name of vm contains upper-case. | azure bug_report cloud easyfix waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_virtualmachine
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 221520cbad) last updated 2016/07/13 15:32:29 (GMT +900)
lib/ansible/modules/core: (detached HEAD db8af4c5af) last updated 2016/07/13 15:32:38 (GMT +900)
lib/ansible/modules/extras: (detached HEAD 482b1a640e) last updated 2016/07/13 15:32:38 (GMT +900)
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
Creating a new azure virtualmachine with upper-cased letter fails without setting a specific storage account because `AzureRMVirtualMachine.create_default_storage_account` try to create a storage account with upper-case.
As described in [this document](https://msdn.microsoft.com/en-us/library/azure/hh264518.aspx), the storage account name can only use numbers and lower-case letters.
##### STEPS TO REPRODUCE
Here is a sample task.
```yaml
- azure_rm_virtualmachine:
name: nameWithUpper
resource_group: Testing
vm_size: Standard_D1
public_ip_allocation_method: Dynamic
admin_username: AdminUserName
admin_password: AdminP@ssw0rd
open_ports:
- 3389
- 5986
os_type: Windows
image:
publisher: MicrosoftWindowsServer
offer: WindowsServer
sku: Windows-Server-Technical-Preview
version: latest
```
##### EXPECTED RESULTS
The module should convert the vm name to lowercase before trying to create a default storage account.
##### ACTUAL RESULTS
Creating storage account always fails as below.
```
TASK [azure_rm_virtualmachine] *************************************************
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to create a unique storage account name for nameWithUpper. Try using a different VM name."}
```
| True | azure_rm_virtualmachine module fails creating a virtualmachine when the name of vm contains upper-case. - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_virtualmachine
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 221520cbad) last updated 2016/07/13 15:32:29 (GMT +900)
lib/ansible/modules/core: (detached HEAD db8af4c5af) last updated 2016/07/13 15:32:38 (GMT +900)
lib/ansible/modules/extras: (detached HEAD 482b1a640e) last updated 2016/07/13 15:32:38 (GMT +900)
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
Creating a new azure virtualmachine with upper-cased letter fails without setting a specific storage account because `AzureRMVirtualMachine.create_default_storage_account` try to create a storage account with upper-case.
As described in [this document](https://msdn.microsoft.com/en-us/library/azure/hh264518.aspx), the storage account name can only use numbers and lower-case letters.
##### STEPS TO REPRODUCE
Here is a sample task.
```yaml
- azure_rm_virtualmachine:
name: nameWithUpper
resource_group: Testing
vm_size: Standard_D1
public_ip_allocation_method: Dynamic
admin_username: AdminUserName
admin_password: AdminP@ssw0rd
open_ports:
- 3389
- 5986
os_type: Windows
image:
publisher: MicrosoftWindowsServer
offer: WindowsServer
sku: Windows-Server-Technical-Preview
version: latest
```
##### EXPECTED RESULTS
The module should convert the vm name to lowercase before trying to create a default storage account.
##### ACTUAL RESULTS
Creating storage account always fails as below.
```
TASK [azure_rm_virtualmachine] *************************************************
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to create a unique storage account name for nameWithUpper. Try using a different VM name."}
```
| main | azure rm virtualmachine module fails creating a virtualmachine when the name of vm contains upper case issue type bug report component name azure rm virtualmachine ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file configured module search path default w o overrides configuration n a os environment n a summary creating a new azure virtualmachine with upper cased letter fails without setting a specific storage account because azurermvirtualmachine create default storage account try to create a storage account with upper case as described in the storage account name can only use numbers and lower case letters steps to reproduce here is a sample task yaml azure rm virtualmachine name namewithupper resource group testing vm size standard public ip allocation method dynamic admin username adminusername admin password adminp open ports os type windows image publisher microsoftwindowsserver offer windowsserver sku windows server technical preview version latest expected results the module should convert the vm name to lowercase before trying to create a default storage account actual results creating storage account always fails as below task fatal failed changed false failed true msg failed to create a unique storage account name for namewithupper try using a different vm name | 1 |
2,884 | 10,319,589,571 | IssuesEvent | 2019-08-30 17:57:26 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | I would like to join the Backdrop Contrib community! | Maintainer application | I have a port of Multiple Selects (D7) ready, and I'm a (co)maintainer of a couple of Drupal projects which I can port, and maintain in the (near) future.
-tnx!-
Harold
| True | I would like to join the Backdrop Contrib community! - I have a port of Multiple Selects (D7) ready, and I'm a (co)maintainer of a couple of Drupal projects which I can port, and maintain in the (near) future.
-tnx!-
Harold
| main | i would like to join the backdrop contrib community i have a port of multiple selects ready and i m a co maintainer of a couple of drupal projects which i can port and maintain in the near future tnx harold | 1 |
2,154 | 7,481,259,120 | IssuesEvent | 2018-04-04 20:04:12 | lansuite/lansuite | https://api.github.com/repos/lansuite/lansuite | reopened | TLD .bayern domain not accepted | bug pending-maintainer-response | <!--- Provide a general summary of the issue in the Title above -->
<!-- Formatting tips:
GitHub supports Markdown: https://guides.github.com/features/mastering-markdown/
Multi-line code blocks either with three back ticks, or four space indent.
```php
<?php
$foo = "bar";
...
```
-->
## Expected Behavior
Register with domain.bayern without an error
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
## Current Behavior
Top level domain .bayern not accepted
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include configuration, logs, etc. to reproduce, if relevant -->
1. Register on a fresh installation on lansuite with a .bayern domain
## Your Environment
It is an default behaviour and not only in my environment. | True | TLD .bayern domain not accepted - <!--- Provide a general summary of the issue in the Title above -->
<!-- Formatting tips:
GitHub supports Markdown: https://guides.github.com/features/mastering-markdown/
Multi-line code blocks either with three back ticks, or four space indent.
```php
<?php
$foo = "bar";
...
```
-->
## Expected Behavior
Register with domain.bayern without an error
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
## Current Behavior
Top level domain .bayern not accepted
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include configuration, logs, etc. to reproduce, if relevant -->
1. Register on a fresh installation on lansuite with a .bayern domain
## Your Environment
It is an default behaviour and not only in my environment. | main | tld bayern domain not accepted formatting tips github supports markdown multi line code blocks either with three back ticks or four space indent php php foo bar expected behavior register with domain bayern without an error current behavior top level domain bayern not accepted steps to reproduce for bugs register on a fresh installation on lansuite with a bayern domain your environment it is an default behaviour and not only in my environment | 1 |
922 | 4,622,717,663 | IssuesEvent | 2016-09-27 08:36:17 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | nxos_config isn't idempotent (in some cases) | affects_2.2 bug_report networking P2 waiting_on_maintainer |
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
nxos_config
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 547cea556f) last updated 2016/09/20 12:12:18 (GMT +100)
lib/ansible/modules/core: (devel 12a7027c49) last updated 2016/09/20 15:11:43 (GMT +100)
lib/ansible/modules/extras: (devel db7a3f48e1) last updated 2016/09/20 11:53:00 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
##### STEPS TO REPRODUCE
```
- name: setup
nxos_config:
commands:
- no description
- no shutdown
parents:
- interface Ethernet2/5
match: none
provider: "{{ cli }}"
- name: configure device with config
nxos_config:
src: basic/config.j2
provider: "{{ cli }}"
match: none
register: result
- assert:
that:
- "result.changed == true"
# https://github.com/ansible/ansible-modules-core/issues/4807
- "result.updates is not defined"
- name: check device with config
nxos_config:
src: basic/config.j2
provider: "{{ cli }}"
match: none
register: result
- assert:
that:
# Idempotent test
# https://github.com/ansible/ansible-modules-core/issues/4807
- "result.changed == false"
- "result.updates is not defined"
```
```
cat templates/basic/config.j2
interface Ethernet2/5
description this is a test
shutdown
```
##### EXPECTED RESULTS
##### ACTUAL RESULTS
```
TASK [test_nxos_config : configure device with config] *************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:14
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py
<nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" && echo ansible-tmp-1474483552.07-278464852480214="` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" ) && sleep 0'
<nxos01> PUT /tmp/tmpmSyMj7 TO /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py
<nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/ /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py && sleep 0'
<nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/" > /dev/null 2>&1 && sleep 0'
changed: [nxos01] => {
"changed": true,
"invocation": {
"module_args": {
"after": null,
"auth_pass": null,
"authorize": false,
"backup": false,
"before": null,
"config": null,
"defaults": false,
"force": false,
"host": "nxos01",
"lines": null,
"match": "none",
"parents": null,
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"port": null,
"provider": {
"host": "nxos01",
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"transport": "cli",
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
},
"replace": "line",
"save": false,
"src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n",
"ssh_keyfile": null,
"timeout": 10,
"transport": "cli",
"use_ssl": false,
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"validate_certs": true
}
},
"warnings": []
}
TASK [test_nxos_config : assert] ***********************************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:21
ok: [nxos01] => {
"changed": false,
"invocation": {
"module_args": {
"that": [
"result.changed == true",
"result.updates is not defined"
]
},
"module_name": "assert"
},
"msg": "all assertions passed"
}
TASK [test_nxos_config : check device with config] *****************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:27
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py
<nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" && echo ansible-tmp-1474483562.39-244603509576702="` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" ) && sleep 0'
<nxos01> PUT /tmp/tmpTQ_P7m TO /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py
<nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/ /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py && sleep 0'
<nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/" > /dev/null 2>&1 && sleep 0'
changed: [nxos01] => {
"changed": true,
"invocation": {
"module_args": {
"after": null,
"auth_pass": null,
"authorize": false,
"backup": false,
"before": null,
"config": null,
"defaults": false,
"force": false,
"host": "nxos01",
"lines": null,
"match": "none",
"parents": null,
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"port": null,
"provider": {
"host": "nxos01",
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"transport": "cli",
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
},
"replace": "line",
"save": false,
"src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n",
"ssh_keyfile": null,
"timeout": 10,
"transport": "cli",
"use_ssl": false,
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"validate_certs": true
}
},
"warnings": []
}
TASK [test_nxos_config : assert] ***********************************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:34
fatal: [nxos01]: FAILED! => {
"assertion": "result.changed == false",
"changed": false,
"evaluated_to": false,
"failed": true,
"invocation": {
"module_args": {
"that": [
"result.changed == false",
"result.updates is not defined"
]
},
"module_name": "assert"
}
}
to retry, use: --limit @/home/johnb/git/ansible-inc/test-network-modules/nxos.retry
PLAY RECAP *********************************************************************
nxos01 : ok=46 changed=12 unreachable=0 failed=1
```
| True | nxos_config isn't idempotent (in some cases) -
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
nxos_config
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 547cea556f) last updated 2016/09/20 12:12:18 (GMT +100)
lib/ansible/modules/core: (devel 12a7027c49) last updated 2016/09/20 15:11:43 (GMT +100)
lib/ansible/modules/extras: (devel db7a3f48e1) last updated 2016/09/20 11:53:00 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
##### STEPS TO REPRODUCE
```
- name: setup
nxos_config:
commands:
- no description
- no shutdown
parents:
- interface Ethernet2/5
match: none
provider: "{{ cli }}"
- name: configure device with config
nxos_config:
src: basic/config.j2
provider: "{{ cli }}"
match: none
register: result
- assert:
that:
- "result.changed == true"
# https://github.com/ansible/ansible-modules-core/issues/4807
- "result.updates is not defined"
- name: check device with config
nxos_config:
src: basic/config.j2
provider: "{{ cli }}"
match: none
register: result
- assert:
that:
# Idempotent test
# https://github.com/ansible/ansible-modules-core/issues/4807
- "result.changed == false"
- "result.updates is not defined"
```
```
cat templates/basic/config.j2
interface Ethernet2/5
description this is a test
shutdown
```
##### EXPECTED RESULTS
##### ACTUAL RESULTS
```
TASK [test_nxos_config : configure device with config] *************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:14
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py
<nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" && echo ansible-tmp-1474483552.07-278464852480214="` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" ) && sleep 0'
<nxos01> PUT /tmp/tmpmSyMj7 TO /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py
<nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/ /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py && sleep 0'
<nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/" > /dev/null 2>&1 && sleep 0'
changed: [nxos01] => {
"changed": true,
"invocation": {
"module_args": {
"after": null,
"auth_pass": null,
"authorize": false,
"backup": false,
"before": null,
"config": null,
"defaults": false,
"force": false,
"host": "nxos01",
"lines": null,
"match": "none",
"parents": null,
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"port": null,
"provider": {
"host": "nxos01",
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"transport": "cli",
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
},
"replace": "line",
"save": false,
"src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n",
"ssh_keyfile": null,
"timeout": 10,
"transport": "cli",
"use_ssl": false,
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"validate_certs": true
}
},
"warnings": []
}
TASK [test_nxos_config : assert] ***********************************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:21
ok: [nxos01] => {
"changed": false,
"invocation": {
"module_args": {
"that": [
"result.changed == true",
"result.updates is not defined"
]
},
"module_name": "assert"
},
"msg": "all assertions passed"
}
TASK [test_nxos_config : check device with config] *****************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:27
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py
<nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" && echo ansible-tmp-1474483562.39-244603509576702="` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" ) && sleep 0'
<nxos01> PUT /tmp/tmpTQ_P7m TO /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py
<nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/ /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py && sleep 0'
<nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/" > /dev/null 2>&1 && sleep 0'
changed: [nxos01] => {
"changed": true,
"invocation": {
"module_args": {
"after": null,
"auth_pass": null,
"authorize": false,
"backup": false,
"before": null,
"config": null,
"defaults": false,
"force": false,
"host": "nxos01",
"lines": null,
"match": "none",
"parents": null,
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"port": null,
"provider": {
"host": "nxos01",
"password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"transport": "cli",
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER"
},
"replace": "line",
"save": false,
"src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n",
"ssh_keyfile": null,
"timeout": 10,
"transport": "cli",
"use_ssl": false,
"username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER",
"validate_certs": true
}
},
"warnings": []
}
TASK [test_nxos_config : assert] ***********************************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:34
fatal: [nxos01]: FAILED! => {
"assertion": "result.changed == false",
"changed": false,
"evaluated_to": false,
"failed": true,
"invocation": {
"module_args": {
"that": [
"result.changed == false",
"result.updates is not defined"
]
},
"module_name": "assert"
}
}
to retry, use: --limit @/home/johnb/git/ansible-inc/test-network-modules/nxos.retry
PLAY RECAP *********************************************************************
nxos01 : ok=46 changed=12 unreachable=0 failed=1
```
| main | nxos config isn t idempotent in some cases issue type bug report component name nxos config ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary steps to reproduce name setup nxos config commands no description no shutdown parents interface match none provider cli name configure device with config nxos config src basic config provider cli match none register result assert that result changed true result updates is not defined name check device with config nxos config src basic config provider cli match none register result assert that idempotent test result changed false result updates is not defined cat templates basic config interface description this is a test shutdown expected results actual results task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml using module file home johnb git ansible inc ansible lib ansible modules core network nxos nxos config py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home johnb ansible tmp ansible tmp nxos config py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp nxos config py sleep exec bin sh c python home johnb ansible tmp ansible tmp nxos config py rm rf home johnb ansible tmp ansible tmp dev null sleep changed changed true invocation module args after null auth pass null authorize false backup false before null config null defaults false force false host lines null match none parents null password value specified in no log parameter port null provider host password value specified in no log parameter transport cli username value specified in no log parameter replace line save false src interface n description this is a test n shutdown n n ssh keyfile null timeout transport cli use ssl false username value specified in no log parameter validate certs true warnings task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml ok changed false invocation module args that result changed true result updates is not defined module name assert msg all assertions passed task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml using module file home johnb git ansible inc ansible lib ansible modules core network nxos nxos config py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmptq to home johnb ansible tmp ansible tmp nxos config py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp nxos config py sleep exec bin sh c python home johnb ansible tmp ansible tmp nxos config py rm rf home johnb ansible tmp ansible tmp dev null sleep changed changed true invocation module args after null auth pass null authorize false backup false before null config null defaults false force false host lines null match none parents null password value specified in no log parameter port null provider host password value specified in no log parameter transport cli username value specified in no log parameter replace line save false src interface n description this is a test n shutdown n n ssh keyfile null timeout transport cli use ssl false username value specified in no log parameter validate certs true warnings task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml fatal failed assertion result changed false changed false evaluated to false failed true invocation module args that result changed false result updates is not defined module name assert to retry use limit home johnb git ansible inc test network modules nxos retry play recap ok changed unreachable failed | 1 |
4,167 | 6,963,655,800 | IssuesEvent | 2017-12-08 18:17:41 | mlibrary/cozy-sun-bear | https://api.github.com/repos/mlibrary/cozy-sun-bear | opened | Build in support for IE11 and IE10 | browser compatibility EPUB | CSB must support older versions of IE to IE10. Full functionality is not a requirement.
- [ ] Reading and paging must work
- [ ] Table of contents must work
- [ ] Preferences should work, but not mandatory | True | Build in support for IE11 and IE10 - CSB must support older versions of IE to IE10. Full functionality is not a requirement.
- [ ] Reading and paging must work
- [ ] Table of contents must work
- [ ] Preferences should work, but not mandatory | non_main | build in support for and csb must support older versions of ie to full functionality is not a requirement reading and paging must work table of contents must work preferences should work but not mandatory | 0 |
294,706 | 22,161,017,938 | IssuesEvent | 2022-06-04 14:07:15 | nigelmann/res-http | https://api.github.com/repos/nigelmann/res-http | opened | Why is a reverse proxy useful to improve the security of the infrastructure ? | documentation | Hello m8 !
There is just one part of the documentation we need to discuss.
https://github.com/HEIGVD-Course-API/API-2021-HTTP-Infra#step-4-ajax-requests-with-jquery
Maybe tomorrow you have a little time for that ? I would be gr8ful
In the mean time if you have some spare time you can check the documentation. I have added the missing parts.
Cheers m8 ! | 1.0 | Why is a reverse proxy useful to improve the security of the infrastructure ? - Hello m8 !
There is just one part of the documentation we need to discuss.
https://github.com/HEIGVD-Course-API/API-2021-HTTP-Infra#step-4-ajax-requests-with-jquery
Maybe tomorrow you have a little time for that ? I would be gr8ful
In the mean time if you have some spare time you can check the documentation. I have added the missing parts.
Cheers m8 ! | non_main | why is a reverse proxy useful to improve the security of the infrastructure hello there is just one part of the documentation we need to discuss maybe tomorrow you have a little time for that i would be in the mean time if you have some spare time you can check the documentation i have added the missing parts cheers | 0 |
235,298 | 19,322,232,906 | IssuesEvent | 2021-12-14 07:29:07 | pingcap/tidb | https://api.github.com/repos/pingcap/tidb | closed | TiDB CI hang for more then 10 min | type/bug component/test component/tikv severity/major | ## Bug Report
```
[2021-11-23T14:16:43.094Z] FAIL github.com/pingcap/tidb/session 600.096s
```
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
in ci https://ci.pingcap.net/blue/organizations/jenkins/tidb_ghpr_check_2/detail/tidb_ghpr_check_2/47625/pipeline/64
<!-- a step by step guide for reproducing the bug. -->
### 2. What did you expect to see? (Required)
### 3. What did you see instead (Required)
### 4. What is your TiDB version? (Required)
master
<!-- Paste the output of SELECT tidb_version() -->
| 1.0 | TiDB CI hang for more then 10 min - ## Bug Report
```
[2021-11-23T14:16:43.094Z] FAIL github.com/pingcap/tidb/session 600.096s
```
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
in ci https://ci.pingcap.net/blue/organizations/jenkins/tidb_ghpr_check_2/detail/tidb_ghpr_check_2/47625/pipeline/64
<!-- a step by step guide for reproducing the bug. -->
### 2. What did you expect to see? (Required)
### 3. What did you see instead (Required)
### 4. What is your TiDB version? (Required)
master
<!-- Paste the output of SELECT tidb_version() -->
| non_main | tidb ci hang for more then min bug report fail github com pingcap tidb session please answer these questions before submitting your issue thanks minimal reproduce step required in ci what did you expect to see required what did you see instead required what is your tidb version required master | 0 |
5,642 | 28,369,700,374 | IssuesEvent | 2023-04-12 16:03:48 | deislabs/spiderlightning | https://api.github.com/repos/deislabs/spiderlightning | opened | installation fails on v0.4.1 in wsl (ubuntu 20.04) with "`GLIBC_2.32' not found (required by slight)" | π bug π§ maintainer issue | **Description of the bug**
βββ ~
β°βββββΆ slight
slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.32' not found (required by slight)
slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.34' not found (required by slight)
slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.33' not found (required by slight)
**To Reproduce**
the "unix" installation instructions. "unix?"
βββ ~
β°βββββΆ /bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/deislabs/spiderlightning/main/install.sh)"
>>> LATEST RELEASE: v0.4.1...
>>> DONLOADING FROM: https://github.com/deislabs/spiderlightning/releases/download/v0.4.1/slight-linux-x86_64.tar.gz...
>>> DOWNLOADED BINARY TAR.
>>> EXTRACTED BINARY TAR.
>>> INSTALLED BINARY.
>>> CLEANED UP.
**Additional context**
βββ ~
β°βββββΆ uname -a
Linux 5.15.90.1-microsoft-standard-WSL2 #1 SMP Fri Jan 27 02:56:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Li | True | installation fails on v0.4.1 in wsl (ubuntu 20.04) with "`GLIBC_2.32' not found (required by slight)" - **Description of the bug**
βββ ~
β°βββββΆ slight
slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.32' not found (required by slight)
slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.34' not found (required by slight)
slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.33' not found (required by slight)
**To Reproduce**
the "unix" installation instructions. "unix?"
βββ ~
β°βββββΆ /bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/deislabs/spiderlightning/main/install.sh)"
>>> LATEST RELEASE: v0.4.1...
>>> DONLOADING FROM: https://github.com/deislabs/spiderlightning/releases/download/v0.4.1/slight-linux-x86_64.tar.gz...
>>> DOWNLOADED BINARY TAR.
>>> EXTRACTED BINARY TAR.
>>> INSTALLED BINARY.
>>> CLEANED UP.
**Additional context**
βββ ~
β°βββββΆ uname -a
Linux 5.15.90.1-microsoft-standard-WSL2 #1 SMP Fri Jan 27 02:56:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Li | main | installation fails on in wsl ubuntu with glibc not found required by slight description of the bug βββ β°βββββΆ slight slight lib linux gnu libc so version glibc not found required by slight slight lib linux gnu libc so version glibc not found required by slight slight lib linux gnu libc so version glibc not found required by slight to reproduce the unix installation instructions unix βββ β°βββββΆ bin bash c curl fssl latest release donloading from downloaded binary tar extracted binary tar installed binary cleaned up additional context βββ β°βββββΆ uname a linux microsoft standard smp fri jan utc gnu li | 1 |
35,745 | 7,988,291,372 | IssuesEvent | 2018-07-19 10:31:40 | Microsoft/devkit-sdk | https://api.github.com/repos/Microsoft/devkit-sdk | closed | [NOTE] Deprecate existing installation package | Code Ready Feature P1 | As we are moving to the new [IoT Workbench](https://marketplace.visualstudio.com/items?itemName=vsciot-vscode.vscode-iot-workbench), the DevKit will be full supported by this new VS Code extension, which will be more convenient for developing, less tooltrain acquisition paints and more straightforward on creating a IoT Solution.
We will deprecate current installation package, which includes:
- Installation scripts
- Task scripts for all examples
- All binaries
We will keep delivering new SDK library for IoT DevKit, add new features and fix bugs.
This is not a 'deprecation' of the IoT DevKit, but a 'evolution' which will lead to a new IoT E2E dev experience.
> 'Deprecation' not mean 'Disappearance', all historic packages are still keeping for downloading. | 1.0 | [NOTE] Deprecate existing installation package - As we are moving to the new [IoT Workbench](https://marketplace.visualstudio.com/items?itemName=vsciot-vscode.vscode-iot-workbench), the DevKit will be full supported by this new VS Code extension, which will be more convenient for developing, less tooltrain acquisition paints and more straightforward on creating a IoT Solution.
We will deprecate current installation package, which includes:
- Installation scripts
- Task scripts for all examples
- All binaries
We will keep delivering new SDK library for IoT DevKit, add new features and fix bugs.
This is not a 'deprecation' of the IoT DevKit, but a 'evolution' which will lead to a new IoT E2E dev experience.
> 'Deprecation' not mean 'Disappearance', all historic packages are still keeping for downloading. | non_main | deprecate existing installation package as we are moving to the new the devkit will be full supported by this new vs code extension which will be more convenient for developing less tooltrain acquisition paints and more straightforward on creating a iot solution we will deprecate current installation package which includes installation scripts task scripts for all examples all binaries we will keep delivering new sdk library for iot devkit add new features and fix bugs this is not a deprecation of the iot devkit but a evolution which will lead to a new iot dev experience deprecation not mean disappearance all historic packages are still keeping for downloading | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.