Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
2,441
8,638,204,691
IssuesEvent
2018-11-23 14:00:05
invertase/react-native-firebase
https://api.github.com/repos/invertase/react-native-firebase
closed
Can`t receive credential-already-in-use from linkWithCredential on Android
await-maintainer-feedback await-react-native-pr ios πŸ‘‰ await-user-feedback πŸ€– android
<!--- Hello there you awesome person; Please note that the issue list of this repo is exclusively for bug reports; 1) For feature requests please visit our [Feature Request Board](https://boards.invertase.io/react-native-firebase). 2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase 3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed. --> <!-- NOTE: You can change any of the `[ ]` to `[x]` to mark an option(s) as selected --> <!-- PLEASE DO NOT REMOVE ANY SECTIONS FROM THIS ISSUE TEMPLATE --> <!-- Leave them as they are even if they're irrelevant to your issue --> ## Issue I have code which should authtentificate user if it already exists or link to anonymous ``` export const authorizeSkippedUser = function* authorizeSkippedUser( user, credentials ) { try { let result = yield call( [user, user.linkAndRetrieveDataWithCredential], credentials ); let userProfile = yield call(updateAndRetrieveUserProfile, user.uid); yield put(AuthActions.authSuccess(userProfile, true)); } catch (ex) { if ( ex.code && ex.code === 'auth/credential-already-in-use' && isSocialProvider(credentials.providerId) ) { yield put(AuthActions.authSignOut()); yield call( [firebaseAuth, firebaseAuth.signInAndRetrieveDataWithCredential], credentials ); } else if (ex.code && ex.code === 'auth/email-already-in-use') { ex.message = yield* getProvideMessage(ex.email); throw ex; } else { throw ex; } } }; ``` it work fine before but now i can`t receive error auth/credential-already-in-use and I don`t know why, I think maybe after somewhere update <!-- Please describe your issue here --^ and provide as much detail as you can. --> <!-- Include code snippets that show your usages of the library in the context of your project. --> <!-- Snippets that also show how and where the library is imported in JS are useful to debug issues relating to importing or methods not found issues --> --- ## Project Files <!-- Provide the contents of key project files which will help to debug --> <!-- For Example: --> <!-- - iOS: `Podfile` contents. --> <!-- - Android: `android/build.gradle` contents. --> <!-- - Android: `android/app/build.gradle` contents. --> <!-- - Android: `AndroidManifest.xml` contents. --> <!-- ADD THE CONTENTS OF THE FILES IN THE PROVIDED CODE BLOCKS BELOW --> ### Android #### `android/build.gradle`: ```groovy // Top-level build file where you can add configuration options common to all sub-projects/modules. buildscript { ext { buildToolsVersion = "27.0.3" minSdkVersion = 16 compileSdkVersion = 27 targetSdkVersion = 26 supportLibVersion = "27.1.1" } repositories { jcenter() maven { url 'https://maven.fabric.io/public' } maven { url 'https://maven.google.com/' name 'Google' } google() } dependencies { classpath 'com.android.tools.build:gradle:3.1.4' classpath 'com.google.gms:google-services:3.2.1' classpath 'io.fabric.tools:gradle:1.25.4' classpath 'com.google.firebase:firebase-plugins:1.1.5' // NOTE: Do not place your application dependencies here; they belong // in the individual module build.gradle files } } allprojects { repositories { google() mavenLocal() jcenter() maven { // All of React Native (JS, Obj-C sources, Android binaries) is installed from npm url "$rootDir/../node_modules/react-native/android" } maven { url 'https://maven.google.com/' name 'Google' } configurations.all { resolutionStrategy { force 'com.facebook.android:facebook-android-sdk:4.22.1' eachDependency { DependencyResolveDetails details -> if (details.requested.group == 'com.facebook.react' && details.requested.name == 'react-native') { def file = new File("$rootDir/../node_modules/react-native/package.json") def version = new groovy.json.JsonSlurper().parseText(file.text).version details.useVersion version } } } } } } task wrapper(type: Wrapper) { gradleVersion = '4.4' distributionUrl = distributionUrl.replace("bin", "all") } ``` #### `android/app/build.gradle`: ```groovy dependencies { compile project(':react-native-config') implementation project(':react-native-fabric') implementation(project(':react-native-firebase')) { transitive = false } implementation project(':react-native-gesture-handler') implementation project(':appcenter-crashes') implementation project(':appcenter-analytics') implementation project(':appcenter') implementation project(':react-native-android-location-services-dialog-box') implementation project(':react-native-background-timer') implementation project(':react-native-code-push') implementation(project(':react-native-device-info')) { exclude group: "com.google.android.gms" // very important } implementation project(':react-native-fbsdk') implementation project(':react-native-fetch-blob') implementation(project(':react-native-google-signin')) { exclude group: "com.google.android.gms" // very important } implementation project(':react-native-keep-awake') implementation project(':react-native-linear-gradient') implementation project(':react-native-shake-event') implementation project(':react-native-splash-screen') implementation project(':react-native-svg') implementation project(':react-native-vector-icons') implementation project(':react-native-view-shot') implementation fileTree(dir: "libs", include: ["*.jar"]) //noinspection GradleCompatible implementation "com.android.support:appcompat-v7:${rootProject.ext.supportLibVersion}" implementation 'com.android.support:multidex:1.0.1' //noinspection GradleDynamicVersion implementation 'com.facebook.react:react-native:+' // From node_modules implementation 'com.twitter.sdk.android:twitter:3.0.0' implementation 'com.google.code.gson:gson:2.8.0' implementation('com.google.firebase:firebase-core:16.0.1') { force = true } implementation('com.google.firebase:firebase-messaging:17.1.0') { force = true } implementation "com.google.firebase:firebase-ads:15.0.1" implementation "com.google.firebase:firebase-storage:16.0.1" implementation "com.google.firebase:firebase-perf:16.1.0" implementation "com.google.firebase:firebase-database:16.0.1" implementation "com.google.firebase:firebase-auth:16.0.3" implementation 'me.leolin:ShortcutBadger:1.1.21@aar' implementation("com.google.android.gms:play-services-base:15.0.1") { force = true } implementation("com.google.android.gms:play-services-base:15.0.1") { force = true } implementation('com.google.android.gms:play-services-ads:15.0.1') { force = true } implementation('com.google.android.gms:play-services-auth:15.0.1') { force = true } implementation('com.google.android.gms:play-services-gcm:15.0.1') { force = true } implementation('com.crashlytics.sdk.android:crashlytics:2.9.3@aar') { transitive = true } } ``` --- ## Environment <!-- change `[ ]` to `[x]` to select an option(s) --> - **Platform that you're experiencing the issue on**: - [ ] iOS - [x] Android - [ ] **iOS** but have not tested behavior on Android - [ ] **Android** but have not tested behavior on iOS - [ ] Both - **Operating System:** - [x] MacOS, version: `10.12.6` - [ ] Windows, version: `N/A` - [ ] Other, please specify: `N/A` - **Build Tools:** - `Android Studio 3.2` - **`React Native` version:** - `0.57.1` - **`React Native Firebase` library version:** - `4.3.8` - **`Firebase` module(s) you're using that has the issue:** - [] **N/A** - [x] Authentication - [ ] Analytics - [ ] Cloud **Firestore** - [ ] Cloud **Messaging** (FCM) - [ ] Crashlytics - [ ] Dynamic **Links** - [ ] **Functions** Callable - [ ] Invites - [ ] Instance ID - [ ] Notifications - [ ] Performance Monitoring - [ ] Realtime **Database** - [ ] Remote **Config** - [ ] Storage - **Are you using `TypeScript`?** - [x] No - [ ] Yes, version: `N/A` - **Are you using Expo, e.g. `ExpoKit`?** - [x] No - [ ] Yes, I've _not_ ejected - [ ] Yes, but I **have** ejected to `ExpoKit` - [ ] Yes, but I **have** ejected to vanilla React Native - Expo version: `N/A` <!-- Thanks for reading this far down ❀️ --> <!-- High quality, detailed issues are much easier and quicker to triage for maintainers --> <!-- For bonus points, if you put a πŸ”₯ (:fire:) emojii at the start of the issue title we'll know --> <!-- that you took the time to fill this out correctly, or, at least read this far --> --- Think `react-native-firebase` is great? Please consider supporting the project with any of the below: - πŸ‘‰ Donate via [Open Collective](https://opencollective.com/react-native-firebase/donate) - πŸ‘‰ Follow [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter - πŸ‘‰ Star this repo on GitHub ⭐️ - πŸ‘‰ Contribute; see our [contributing guide](./../../CONTRIBUTING.md)
True
Can`t receive credential-already-in-use from linkWithCredential on Android - <!--- Hello there you awesome person; Please note that the issue list of this repo is exclusively for bug reports; 1) For feature requests please visit our [Feature Request Board](https://boards.invertase.io/react-native-firebase). 2) For questions and support please use our Discord chat: https://discord.gg/C9aK28N or Stack Overflow: https://stackoverflow.com/questions/tagged/react-native-firebase 3) If this is a setup issue then please make sure you've correctly followed the setup guides, most setup issues such as 'duplicate dex files', 'default app has not been initialized' etc are all down to an incorrect setup as the guides haven't been correctly followed. --> <!-- NOTE: You can change any of the `[ ]` to `[x]` to mark an option(s) as selected --> <!-- PLEASE DO NOT REMOVE ANY SECTIONS FROM THIS ISSUE TEMPLATE --> <!-- Leave them as they are even if they're irrelevant to your issue --> ## Issue I have code which should authtentificate user if it already exists or link to anonymous ``` export const authorizeSkippedUser = function* authorizeSkippedUser( user, credentials ) { try { let result = yield call( [user, user.linkAndRetrieveDataWithCredential], credentials ); let userProfile = yield call(updateAndRetrieveUserProfile, user.uid); yield put(AuthActions.authSuccess(userProfile, true)); } catch (ex) { if ( ex.code && ex.code === 'auth/credential-already-in-use' && isSocialProvider(credentials.providerId) ) { yield put(AuthActions.authSignOut()); yield call( [firebaseAuth, firebaseAuth.signInAndRetrieveDataWithCredential], credentials ); } else if (ex.code && ex.code === 'auth/email-already-in-use') { ex.message = yield* getProvideMessage(ex.email); throw ex; } else { throw ex; } } }; ``` it work fine before but now i can`t receive error auth/credential-already-in-use and I don`t know why, I think maybe after somewhere update <!-- Please describe your issue here --^ and provide as much detail as you can. --> <!-- Include code snippets that show your usages of the library in the context of your project. --> <!-- Snippets that also show how and where the library is imported in JS are useful to debug issues relating to importing or methods not found issues --> --- ## Project Files <!-- Provide the contents of key project files which will help to debug --> <!-- For Example: --> <!-- - iOS: `Podfile` contents. --> <!-- - Android: `android/build.gradle` contents. --> <!-- - Android: `android/app/build.gradle` contents. --> <!-- - Android: `AndroidManifest.xml` contents. --> <!-- ADD THE CONTENTS OF THE FILES IN THE PROVIDED CODE BLOCKS BELOW --> ### Android #### `android/build.gradle`: ```groovy // Top-level build file where you can add configuration options common to all sub-projects/modules. buildscript { ext { buildToolsVersion = "27.0.3" minSdkVersion = 16 compileSdkVersion = 27 targetSdkVersion = 26 supportLibVersion = "27.1.1" } repositories { jcenter() maven { url 'https://maven.fabric.io/public' } maven { url 'https://maven.google.com/' name 'Google' } google() } dependencies { classpath 'com.android.tools.build:gradle:3.1.4' classpath 'com.google.gms:google-services:3.2.1' classpath 'io.fabric.tools:gradle:1.25.4' classpath 'com.google.firebase:firebase-plugins:1.1.5' // NOTE: Do not place your application dependencies here; they belong // in the individual module build.gradle files } } allprojects { repositories { google() mavenLocal() jcenter() maven { // All of React Native (JS, Obj-C sources, Android binaries) is installed from npm url "$rootDir/../node_modules/react-native/android" } maven { url 'https://maven.google.com/' name 'Google' } configurations.all { resolutionStrategy { force 'com.facebook.android:facebook-android-sdk:4.22.1' eachDependency { DependencyResolveDetails details -> if (details.requested.group == 'com.facebook.react' && details.requested.name == 'react-native') { def file = new File("$rootDir/../node_modules/react-native/package.json") def version = new groovy.json.JsonSlurper().parseText(file.text).version details.useVersion version } } } } } } task wrapper(type: Wrapper) { gradleVersion = '4.4' distributionUrl = distributionUrl.replace("bin", "all") } ``` #### `android/app/build.gradle`: ```groovy dependencies { compile project(':react-native-config') implementation project(':react-native-fabric') implementation(project(':react-native-firebase')) { transitive = false } implementation project(':react-native-gesture-handler') implementation project(':appcenter-crashes') implementation project(':appcenter-analytics') implementation project(':appcenter') implementation project(':react-native-android-location-services-dialog-box') implementation project(':react-native-background-timer') implementation project(':react-native-code-push') implementation(project(':react-native-device-info')) { exclude group: "com.google.android.gms" // very important } implementation project(':react-native-fbsdk') implementation project(':react-native-fetch-blob') implementation(project(':react-native-google-signin')) { exclude group: "com.google.android.gms" // very important } implementation project(':react-native-keep-awake') implementation project(':react-native-linear-gradient') implementation project(':react-native-shake-event') implementation project(':react-native-splash-screen') implementation project(':react-native-svg') implementation project(':react-native-vector-icons') implementation project(':react-native-view-shot') implementation fileTree(dir: "libs", include: ["*.jar"]) //noinspection GradleCompatible implementation "com.android.support:appcompat-v7:${rootProject.ext.supportLibVersion}" implementation 'com.android.support:multidex:1.0.1' //noinspection GradleDynamicVersion implementation 'com.facebook.react:react-native:+' // From node_modules implementation 'com.twitter.sdk.android:twitter:3.0.0' implementation 'com.google.code.gson:gson:2.8.0' implementation('com.google.firebase:firebase-core:16.0.1') { force = true } implementation('com.google.firebase:firebase-messaging:17.1.0') { force = true } implementation "com.google.firebase:firebase-ads:15.0.1" implementation "com.google.firebase:firebase-storage:16.0.1" implementation "com.google.firebase:firebase-perf:16.1.0" implementation "com.google.firebase:firebase-database:16.0.1" implementation "com.google.firebase:firebase-auth:16.0.3" implementation 'me.leolin:ShortcutBadger:1.1.21@aar' implementation("com.google.android.gms:play-services-base:15.0.1") { force = true } implementation("com.google.android.gms:play-services-base:15.0.1") { force = true } implementation('com.google.android.gms:play-services-ads:15.0.1') { force = true } implementation('com.google.android.gms:play-services-auth:15.0.1') { force = true } implementation('com.google.android.gms:play-services-gcm:15.0.1') { force = true } implementation('com.crashlytics.sdk.android:crashlytics:2.9.3@aar') { transitive = true } } ``` --- ## Environment <!-- change `[ ]` to `[x]` to select an option(s) --> - **Platform that you're experiencing the issue on**: - [ ] iOS - [x] Android - [ ] **iOS** but have not tested behavior on Android - [ ] **Android** but have not tested behavior on iOS - [ ] Both - **Operating System:** - [x] MacOS, version: `10.12.6` - [ ] Windows, version: `N/A` - [ ] Other, please specify: `N/A` - **Build Tools:** - `Android Studio 3.2` - **`React Native` version:** - `0.57.1` - **`React Native Firebase` library version:** - `4.3.8` - **`Firebase` module(s) you're using that has the issue:** - [] **N/A** - [x] Authentication - [ ] Analytics - [ ] Cloud **Firestore** - [ ] Cloud **Messaging** (FCM) - [ ] Crashlytics - [ ] Dynamic **Links** - [ ] **Functions** Callable - [ ] Invites - [ ] Instance ID - [ ] Notifications - [ ] Performance Monitoring - [ ] Realtime **Database** - [ ] Remote **Config** - [ ] Storage - **Are you using `TypeScript`?** - [x] No - [ ] Yes, version: `N/A` - **Are you using Expo, e.g. `ExpoKit`?** - [x] No - [ ] Yes, I've _not_ ejected - [ ] Yes, but I **have** ejected to `ExpoKit` - [ ] Yes, but I **have** ejected to vanilla React Native - Expo version: `N/A` <!-- Thanks for reading this far down ❀️ --> <!-- High quality, detailed issues are much easier and quicker to triage for maintainers --> <!-- For bonus points, if you put a πŸ”₯ (:fire:) emojii at the start of the issue title we'll know --> <!-- that you took the time to fill this out correctly, or, at least read this far --> --- Think `react-native-firebase` is great? Please consider supporting the project with any of the below: - πŸ‘‰ Donate via [Open Collective](https://opencollective.com/react-native-firebase/donate) - πŸ‘‰ Follow [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter - πŸ‘‰ Star this repo on GitHub ⭐️ - πŸ‘‰ Contribute; see our [contributing guide](./../../CONTRIBUTING.md)
main
can t receive credential already in use from linkwithcredential on android hello there you awesome person please note that the issue list of this repo is exclusively for bug reports for feature requests please visit our for questions and support please use our discord chat or stack overflow if this is a setup issue then please make sure you ve correctly followed the setup guides most setup issues such as duplicate dex files default app has not been initialized etc are all down to an incorrect setup as the guides haven t been correctly followed issue i have code which should authtentificate user if it already exists or link to anonymous export const authorizeskippeduser function authorizeskippeduser user credentials try let result yield call credentials let userprofile yield call updateandretrieveuserprofile user uid yield put authactions authsuccess userprofile true catch ex if ex code ex code auth credential already in use issocialprovider credentials providerid yield put authactions authsignout yield call credentials else if ex code ex code auth email already in use ex message yield getprovidemessage ex email throw ex else throw ex it work fine before but now i can t receive error auth credential already in use and i don t know why i think maybe after somewhere update project files android android build gradle groovy top level build file where you can add configuration options common to all sub projects modules buildscript ext buildtoolsversion minsdkversion compilesdkversion targetsdkversion supportlibversion repositories jcenter maven url maven url name google google dependencies classpath com android tools build gradle classpath com google gms google services classpath io fabric tools gradle classpath com google firebase firebase plugins note do not place your application dependencies here they belong in the individual module build gradle files allprojects repositories google mavenlocal jcenter maven all of react native js obj c sources android binaries is installed from npm url rootdir node modules react native android maven url name google configurations all resolutionstrategy force com facebook android facebook android sdk eachdependency dependencyresolvedetails details if details requested group com facebook react details requested name react native def file new file rootdir node modules react native package json def version new groovy json jsonslurper parsetext file text version details useversion version task wrapper type wrapper gradleversion distributionurl distributionurl replace bin all android app build gradle groovy dependencies compile project react native config implementation project react native fabric implementation project react native firebase transitive false implementation project react native gesture handler implementation project appcenter crashes implementation project appcenter analytics implementation project appcenter implementation project react native android location services dialog box implementation project react native background timer implementation project react native code push implementation project react native device info exclude group com google android gms very important implementation project react native fbsdk implementation project react native fetch blob implementation project react native google signin exclude group com google android gms very important implementation project react native keep awake implementation project react native linear gradient implementation project react native shake event implementation project react native splash screen implementation project react native svg implementation project react native vector icons implementation project react native view shot implementation filetree dir libs include noinspection gradlecompatible implementation com android support appcompat rootproject ext supportlibversion implementation com android support multidex noinspection gradledynamicversion implementation com facebook react react native from node modules implementation com twitter sdk android twitter implementation com google code gson gson implementation com google firebase firebase core force true implementation com google firebase firebase messaging force true implementation com google firebase firebase ads implementation com google firebase firebase storage implementation com google firebase firebase perf implementation com google firebase firebase database implementation com google firebase firebase auth implementation me leolin shortcutbadger aar implementation com google android gms play services base force true implementation com google android gms play services base force true implementation com google android gms play services ads force true implementation com google android gms play services auth force true implementation com google android gms play services gcm force true implementation com crashlytics sdk android crashlytics aar transitive true environment platform that you re experiencing the issue on ios android ios but have not tested behavior on android android but have not tested behavior on ios both operating system macos version windows version n a other please specify n a build tools android studio react native version react native firebase library version firebase module s you re using that has the issue n a authentication analytics cloud firestore cloud messaging fcm crashlytics dynamic links functions callable invites instance id notifications performance monitoring realtime database remote config storage are you using typescript no yes version n a are you using expo e g expokit no yes i ve not ejected yes but i have ejected to expokit yes but i have ejected to vanilla react native expo version n a think react native firebase is great please consider supporting the project with any of the below πŸ‘‰ donate via πŸ‘‰ follow and on twitter πŸ‘‰ star this repo on github ⭐️ πŸ‘‰ contribute see our contributing md
1
192
2,814,796,774
IssuesEvent
2015-05-18 22:05:39
DotNetAnalyzers/StyleCopAnalyzers
https://api.github.com/repos/DotNetAnalyzers/StyleCopAnalyzers
closed
Rule proposal - Readability (Maintainablility) rule - Report assignments in condition expressions
maintainability needs discussion new rule proposal
To carry on the side topic at http://stylecop.codeplex.com/discussions/252502... StyleCop does not currently complain about either of the following code patterns. Where bb and cc are bools: ``` if (bb = cc) { // ... } while (bb = cc) { // ... } if ((a = b) > 3) { // ... } while ((a = b) < 4) { // ... } switch (bb = cc) { case true: default: } ``` I believe the pattern shown by the first two assignments-inside-conditionals can sometimes be a typo/mistake, and if not by a mistake the coder, could easily get mistaken by maintainers scanning over it as a comparison operator. Fortunately this isn't a very common problem since it generally only applies to bools. However, all of these examples have more straight-forward, logical implementations by pulling the assignments into their own lines of code. The "shortcuts" being employed by allowing the assignment to be within conditionals are unnecessary and I think we should provide a rule to call this out. Originally proposed by Andy Reeves: http://stylecop.codeplex.com/workitem/6893 ----------------------------------------------------------- follow up discussion: xanatos wrote Nov 7, 2011 at 11:38 AM This is usable: ``` while ((a = b) < 4) { // ... } ``` I often use it for: ``` int count; while ((count = myStream.Read(buffer, 0, buffer.Length)) > 0) { mySecondStream.Write(buffer, 0, count); } ``` The alternatives are all quite ugly: This moves away from the while the exit logic: ``` while (true) { int count = myStream.Read(buffer, 0, buffer.Length); if (count == 0) { break; } mySecondStream.Write(buffer, 0, count); } ``` or This is even worse, because it repeats twice the Read line (so the two Reads could become different) ``` int count = myStream.Read(buffer, 0, buffer.Length); while (count > 0) { mySecondStream.Write(buffer, 0, count); count = myStream.Read(buffer, 0, buffer.Length); } ```
True
Rule proposal - Readability (Maintainablility) rule - Report assignments in condition expressions - To carry on the side topic at http://stylecop.codeplex.com/discussions/252502... StyleCop does not currently complain about either of the following code patterns. Where bb and cc are bools: ``` if (bb = cc) { // ... } while (bb = cc) { // ... } if ((a = b) > 3) { // ... } while ((a = b) < 4) { // ... } switch (bb = cc) { case true: default: } ``` I believe the pattern shown by the first two assignments-inside-conditionals can sometimes be a typo/mistake, and if not by a mistake the coder, could easily get mistaken by maintainers scanning over it as a comparison operator. Fortunately this isn't a very common problem since it generally only applies to bools. However, all of these examples have more straight-forward, logical implementations by pulling the assignments into their own lines of code. The "shortcuts" being employed by allowing the assignment to be within conditionals are unnecessary and I think we should provide a rule to call this out. Originally proposed by Andy Reeves: http://stylecop.codeplex.com/workitem/6893 ----------------------------------------------------------- follow up discussion: xanatos wrote Nov 7, 2011 at 11:38 AM This is usable: ``` while ((a = b) < 4) { // ... } ``` I often use it for: ``` int count; while ((count = myStream.Read(buffer, 0, buffer.Length)) > 0) { mySecondStream.Write(buffer, 0, count); } ``` The alternatives are all quite ugly: This moves away from the while the exit logic: ``` while (true) { int count = myStream.Read(buffer, 0, buffer.Length); if (count == 0) { break; } mySecondStream.Write(buffer, 0, count); } ``` or This is even worse, because it repeats twice the Read line (so the two Reads could become different) ``` int count = myStream.Read(buffer, 0, buffer.Length); while (count > 0) { mySecondStream.Write(buffer, 0, count); count = myStream.Read(buffer, 0, buffer.Length); } ```
main
rule proposal readability maintainablility rule report assignments in condition expressions to carry on the side topic at stylecop does not currently complain about either of the following code patterns where bb and cc are bools if bb cc while bb cc if a b while a b switch bb cc case true default i believe the pattern shown by the first two assignments inside conditionals can sometimes be a typo mistake and if not by a mistake the coder could easily get mistaken by maintainers scanning over it as a comparison operator fortunately this isn t a very common problem since it generally only applies to bools however all of these examples have more straight forward logical implementations by pulling the assignments into their own lines of code the shortcuts being employed by allowing the assignment to be within conditionals are unnecessary and i think we should provide a rule to call this out originally proposed by andy reeves follow up discussion xanatos wrote nov at am this is usable while a b i often use it for int count while count mystream read buffer buffer length mysecondstream write buffer count the alternatives are all quite ugly this moves away from the while the exit logic while true int count mystream read buffer buffer length if count break mysecondstream write buffer count or this is even worse because it repeats twice the read line so the two reads could become different int count mystream read buffer buffer length while count mysecondstream write buffer count count mystream read buffer buffer length
1
20,251
3,800,355,968
IssuesEvent
2016-03-23 18:48:24
servo/servo
https://api.github.com/repos/servo/servo
closed
rust.png in reftests is missing
A-testing C-assigned E-easy
It's used in inline_margin_multiple_fragments_a.html and filter_inline_a.html but it's 404. We should replace references to it with one of the other images in http://mxr.mozilla.org/servo/source/tests/wpt/mozilla/tests/css/ .
1.0
rust.png in reftests is missing - It's used in inline_margin_multiple_fragments_a.html and filter_inline_a.html but it's 404. We should replace references to it with one of the other images in http://mxr.mozilla.org/servo/source/tests/wpt/mozilla/tests/css/ .
non_main
rust png in reftests is missing it s used in inline margin multiple fragments a html and filter inline a html but it s we should replace references to it with one of the other images in
0
5,434
27,243,567,134
IssuesEvent
2023-02-21 22:57:10
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Unable to run docker in ARM Architecture
stage/needs-investigation maintainer/need-followup platform/mac/arm
<!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. --> ### Description: <!-- sam build fails in the Mac Os M1 Chip for lambdas that is using Docker.--> ### Steps to reproduce: <!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) --> 1. Step : 1 Create a YAML for lambda that uses Docker Image 2. Step: 2 Add Docker File in the metadata of yaml which is similar to `FROM python:3.6 WORKDIR /src COPY main.py requirements.txt config.json ./ RUN apt-get update && apt-get install make git RUN apt-get install -y apt-utils RUN apt-get install -y cmake RUN apt-get install -y librdkafka-dev RUN pip install -r requirements.txt ENTRYPOINT [ "/usr/local/bin/python", "-m", "awslambdaric" ] CMD ["main.lambda_handler"]` 3. Fails as it tries to build in arm architecture ### Observed result: <!-- Please provide command output with `--debug` flag set.--> Fails to build the image `creating build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -fPIC -I/usr/local/include/python3.6m -c /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c -o build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.o In file included from /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c:17: /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.h:66:2: error: #error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html" #error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html"` ### Expected result: <!-- Describe what you expected.--> SAM CLI should automatically build Docker image for x86 till the support of Lambda is ready for Graviton ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS:MacOs M1 2. If using SAM CLI, `sam --version`: SAM CLI, version 1.24.0 3. AWS region: ap-southeast-2 `Add --debug flag to any SAM CLI commands you are running`
True
Unable to run docker in ARM Architecture - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. --> ### Description: <!-- sam build fails in the Mac Os M1 Chip for lambdas that is using Docker.--> ### Steps to reproduce: <!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) --> 1. Step : 1 Create a YAML for lambda that uses Docker Image 2. Step: 2 Add Docker File in the metadata of yaml which is similar to `FROM python:3.6 WORKDIR /src COPY main.py requirements.txt config.json ./ RUN apt-get update && apt-get install make git RUN apt-get install -y apt-utils RUN apt-get install -y cmake RUN apt-get install -y librdkafka-dev RUN pip install -r requirements.txt ENTRYPOINT [ "/usr/local/bin/python", "-m", "awslambdaric" ] CMD ["main.lambda_handler"]` 3. Fails as it tries to build in arm architecture ### Observed result: <!-- Please provide command output with `--debug` flag set.--> Fails to build the image `creating build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src gcc -pthread -Wno-unused-result -Wsign-compare -DNDEBUG -g -fwrapv -O3 -Wall -fPIC -I/usr/local/include/python3.6m -c /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c -o build/temp.linux-aarch64-3.6/tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.o In file included from /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.c:17: /tmp/pip-install-bur4_y1q/confluent-kafka_3a16b52446ce4d7a82d5dbf75653e9f4/src/confluent_kafka/src/confluent_kafka.h:66:2: error: #error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html" #error "confluent-kafka-python requires librdkafka v1.6.0 or later. Install the latest version of librdkafka from the Confluent repositories, see http://docs.confluent.io/current/installation.html"` ### Expected result: <!-- Describe what you expected.--> SAM CLI should automatically build Docker image for x86 till the support of Lambda is ready for Graviton ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS:MacOs M1 2. If using SAM CLI, `sam --version`: SAM CLI, version 1.24.0 3. AWS region: ap-southeast-2 `Add --debug flag to any SAM CLI commands you are running`
main
unable to run docker in arm architecture make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description steps to reproduce step create a yaml for lambda that uses docker image step add docker file in the metadata of yaml which is similar to from python workdir src copy main py requirements txt config json run apt get update apt get install make git run apt get install y apt utils run apt get install y cmake run apt get install y librdkafka dev run pip install r requirements txt entrypoint cmd fails as it tries to build in arm architecture observed result fails to build the image creating build temp linux tmp pip install confluent kafka src confluent kafka src gcc pthread wno unused result wsign compare dndebug g fwrapv wall fpic i usr local include c tmp pip install confluent kafka src confluent kafka src confluent kafka c o build temp linux tmp pip install confluent kafka src confluent kafka src confluent kafka o in file included from tmp pip install confluent kafka src confluent kafka src confluent kafka c tmp pip install confluent kafka src confluent kafka src confluent kafka h error error confluent kafka python requires librdkafka or later install the latest version of librdkafka from the confluent repositories see error confluent kafka python requires librdkafka or later install the latest version of librdkafka from the confluent repositories see expected result sam cli should automatically build docker image for till the support of lambda is ready for graviton additional environment details ex windows mac amazon linux etc os macos if using sam cli sam version sam cli version aws region ap southeast add debug flag to any sam cli commands you are running
1
4,681
24,185,022,837
IssuesEvent
2022-09-23 12:35:02
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Question]: How to dynamically render the expanded row content based on row?
type: question ❓ status: waiting for maintainer response πŸ’¬
### Question for Carbon I see in the given examples there are hardcoded values for expanded row like this. ``` {row.isExpanded && ( <TableExpandedRow colSpan={headers.length + 1}> <h1>Expandable row content</h1> <p>Description here</p> </TableExpandedRow> )} ``` The requirement is I need to pass the data for expanded row as part of rows. like ``` const rows = [ { id: 'a', field1: 'Field 1a', expand: 'Value for Expanded row 1' }, { id: 'b', field1: 'Field 1b', expand: 'Value for Expanded row 2' } ]; ``` Is this possible? ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
True
[Question]: How to dynamically render the expanded row content based on row? - ### Question for Carbon I see in the given examples there are hardcoded values for expanded row like this. ``` {row.isExpanded && ( <TableExpandedRow colSpan={headers.length + 1}> <h1>Expandable row content</h1> <p>Description here</p> </TableExpandedRow> )} ``` The requirement is I need to pass the data for expanded row as part of rows. like ``` const rows = [ { id: 'a', field1: 'Field 1a', expand: 'Value for Expanded row 1' }, { id: 'b', field1: 'Field 1b', expand: 'Value for Expanded row 2' } ]; ``` Is this possible? ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
main
how to dynamically render the expanded row content based on row question for carbon i see in the given examples there are hardcoded values for expanded row like this row isexpanded expandable row content description here the requirement is i need to pass the data for expanded row as part of rows like const rows id a field expand value for expanded row id b field expand value for expanded row is this possible code of conduct i agree to follow this project s
1
17,723
4,188,021,266
IssuesEvent
2016-06-23 19:19:38
bignerdranch/expandable-recycler-view
https://api.github.com/repos/bignerdranch/expandable-recycler-view
closed
Documentation error: Need LayoutManager
bug documentation
In the [original blog post](https://www.bignerdranch.com/blog/expand-a-recyclerview-in-four-steps/) you reference adding a layout manager, but [this page with the updated information](http://bignerdranch.github.io/expandable-recycler-view/) doesn't. Took me awhile to figure out why nothing was showing up!
1.0
Documentation error: Need LayoutManager - In the [original blog post](https://www.bignerdranch.com/blog/expand-a-recyclerview-in-four-steps/) you reference adding a layout manager, but [this page with the updated information](http://bignerdranch.github.io/expandable-recycler-view/) doesn't. Took me awhile to figure out why nothing was showing up!
non_main
documentation error need layoutmanager in the you reference adding a layout manager but doesn t took me awhile to figure out why nothing was showing up
0
2,530
8,657,247,698
IssuesEvent
2018-11-27 20:45:31
arcticicestudio/nord-docs
https://api.github.com/repos/arcticicestudio/nord-docs
opened
Netlify Configuration
context-workflow scope-configurability scope-maintainability type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/48661237-35d1a000-ea6f-11e8-8e16-f48948969be6.png" width="60%" /></p> > Related epics: #46 This issue documents a part of the implementation of the [hosting & continuous deployment concept][gh-46] with the [Netlify's configuration file][netlify-docs-toml-ref]. See the β€œHosting” and β€œContinuous Deployment” (sub)sections for more details about the architecture. ## Tasks - Implement Netlify's `netlify.toml` configuration file - [ ] Define the `command` for the production `[build]` section - [ ] Define the `publish` path for the production `[build]` section [gh-46]: https://github.com/arcticicestudio/nord-docs/issues/46 [netlify-docs-toml-ref]: https://www.netlify.com/docs/netlify-toml-reference
True
Netlify Configuration - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48661237-35d1a000-ea6f-11e8-8e16-f48948969be6.png" width="60%" /></p> > Related epics: #46 This issue documents a part of the implementation of the [hosting & continuous deployment concept][gh-46] with the [Netlify's configuration file][netlify-docs-toml-ref]. See the β€œHosting” and β€œContinuous Deployment” (sub)sections for more details about the architecture. ## Tasks - Implement Netlify's `netlify.toml` configuration file - [ ] Define the `command` for the production `[build]` section - [ ] Define the `publish` path for the production `[build]` section [gh-46]: https://github.com/arcticicestudio/nord-docs/issues/46 [netlify-docs-toml-ref]: https://www.netlify.com/docs/netlify-toml-reference
main
netlify configuration related epics this issue documents a part of the implementation of the with the see the β€œhosting” and β€œcontinuous deployment” sub sections for more details about the architecture tasks implement netlify s netlify toml configuration file define the command for the production section define the publish path for the production section
1
1,958
6,678,594,345
IssuesEvent
2017-10-05 14:42:45
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
What3Words Geocoder: Fallback to StandardBlend (autosuggest) API when no results for three-word address
Low-Hanging Fruit Maintainer Approved Maintainer Submitted Suggestion
When a user misspells a three word address they will currently see no results. The What3Words team suggested we try faling back to their [StandardBlend](https://docs.what3words.com/api/v2/#standardblend) (Autocomplete) API in those cases to show potential matches for the given address. If the first API call returns no results, we can make another API call and if it has results, display a Places tile view (similar to the [BikeShare Spices](https://duckduckgo.com/?q=nyc+bike+share+locations&ia=bikesharing)) which provides the similar addresses. This will require creating another `Spice alt_to` endpoint for the StandardBlend API Endpoint. We should use the `text` template for the tiles. The title should be the 3-Word address, and the subtitle should be the location the place is near to (provided by the API). If possible, it would be nice to use our flag icons as well, if possible but that might be difficult. --- IA Page: http://duck.co/ia/view/what3words [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @moollaza
True
What3Words Geocoder: Fallback to StandardBlend (autosuggest) API when no results for three-word address - When a user misspells a three word address they will currently see no results. The What3Words team suggested we try faling back to their [StandardBlend](https://docs.what3words.com/api/v2/#standardblend) (Autocomplete) API in those cases to show potential matches for the given address. If the first API call returns no results, we can make another API call and if it has results, display a Places tile view (similar to the [BikeShare Spices](https://duckduckgo.com/?q=nyc+bike+share+locations&ia=bikesharing)) which provides the similar addresses. This will require creating another `Spice alt_to` endpoint for the StandardBlend API Endpoint. We should use the `text` template for the tiles. The title should be the 3-Word address, and the subtitle should be the location the place is near to (provided by the API). If possible, it would be nice to use our flag icons as well, if possible but that might be difficult. --- IA Page: http://duck.co/ia/view/what3words [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @moollaza
main
geocoder fallback to standardblend autosuggest api when no results for three word address when a user misspells a three word address they will currently see no results the team suggested we try faling back to their autocomplete api in those cases to show potential matches for the given address if the first api call returns no results we can make another api call and if it has results display a places tile view similar to the which provides the similar addresses this will require creating another spice alt to endpoint for the standardblend api endpoint we should use the text template for the tiles the title should be the word address and the subtitle should be the location the place is near to provided by the api if possible it would be nice to use our flag icons as well if possible but that might be difficult ia page moollaza
1
208,746
16,136,311,721
IssuesEvent
2021-04-29 12:19:16
gatsbyjs/gatsby
https://api.github.com/repos/gatsbyjs/gatsby
closed
[docs] update docs for createRemoteFilenode usage
not stale type: documentation
## Summary I was using `createRemoteFilenode` and found some things in the docs that could use an update - using the examples on [this page](https://www.gatsbyjs.org/docs/preprocessing-external-images/#gatsby-node) I got a warning `warn Deprecation warning - adding inferred resolver for field GhostPost.feature_image_localFile. In Gatsby v3, only fields with an explicit directive/extension will get a resolver.` so I think the example should be changed to (not interily sure about this, since I'm not familiar with `@infer` / `@noInfer`): ``` ... createTypes(` type MarkdownRemark implements Node @infer(noDefaultResolvers: false) { frontmatter: Frontmatter } type Frontmatter { title: String! featuredImgUrl: String featuredImgAlt: String } `) ... ``` - the readme if the [`gatsby-source-filesystem`](https://www.gatsbyjs.org/packages/gatsby-source-filesystem/?=filesys#createremotefilenode) doesn't really have a clear usage example (to me at least) since there is no context given as to where it is recommended to use `createRemoteFilenode` , maybe we can add a link to above mentioned page? not sure if anyone is working on any of these pages? if not I can make a pr for this.
1.0
[docs] update docs for createRemoteFilenode usage - ## Summary I was using `createRemoteFilenode` and found some things in the docs that could use an update - using the examples on [this page](https://www.gatsbyjs.org/docs/preprocessing-external-images/#gatsby-node) I got a warning `warn Deprecation warning - adding inferred resolver for field GhostPost.feature_image_localFile. In Gatsby v3, only fields with an explicit directive/extension will get a resolver.` so I think the example should be changed to (not interily sure about this, since I'm not familiar with `@infer` / `@noInfer`): ``` ... createTypes(` type MarkdownRemark implements Node @infer(noDefaultResolvers: false) { frontmatter: Frontmatter } type Frontmatter { title: String! featuredImgUrl: String featuredImgAlt: String } `) ... ``` - the readme if the [`gatsby-source-filesystem`](https://www.gatsbyjs.org/packages/gatsby-source-filesystem/?=filesys#createremotefilenode) doesn't really have a clear usage example (to me at least) since there is no context given as to where it is recommended to use `createRemoteFilenode` , maybe we can add a link to above mentioned page? not sure if anyone is working on any of these pages? if not I can make a pr for this.
non_main
update docs for createremotefilenode usage summary i was using createremotefilenode and found some things in the docs that could use an update using the examples on i got a warning warn deprecation warning adding inferred resolver for field ghostpost feature image localfile in gatsby only fields with an explicit directive extension will get a resolver so i think the example should be changed to not interily sure about this since i m not familiar with infer noinfer createtypes type markdownremark implements node infer nodefaultresolvers false frontmatter frontmatter type frontmatter title string featuredimgurl string featuredimgalt string the readme if the doesn t really have a clear usage example to me at least since there is no context given as to where it is recommended to use createremotefilenode maybe we can add a link to above mentioned page not sure if anyone is working on any of these pages if not i can make a pr for this
0
72,153
8,707,631,539
IssuesEvent
2018-12-06 08:32:02
phetsims/a11y-research
https://api.github.com/repos/phetsims/a11y-research
closed
How well are Arrow keys working in JAWS for 4-way custom move?
design:a11y dev:a11y meeting:a11y type:question
I am having some trouble assessing what is the best way to cue the custom 4-way drag, especially for Friction where pressing the W key is not so useful. In a recent session with a VoiceOver user familiar with our sims, there seemed to be confusion around cuing only the use of the WASD keys for moving the book. **Goal of this issue:** Assess how well the Arrow keys are working in JAWS so we can figure out the best way to provide instructions for the interaction and the best way to provide alternatives for the interaction. **Question 1 for @jessegreenberg** Do you know how well the Arrow keys are working with JAWS now? In early prototypes of BASE the Arrow keys were not working at all with JAWS, and then they worked, but not as well as the WASD keys. **Question 2 for @terracoda & @emily-phet** If JAWS is no longer having a serious problem with releasing the Arrow keys to the web application, how and where is the best way to provide the alternative WASD key instructions? - In PDOM help text, first grab alert, **AND** Keyboard Shortcuts dialog? - Only in PDOM help text **AND** Keyboard Shortcuts? **OR** - Only in the Keyboard Shortcuts dialog? Currently, we do option 1, and we only explicitly refer to using the Arrow keys in the Keyboard Shortcuts dialog. Here are examples from BASE and Friction, and soon there will be a grab buttons in GFL and Faraday's Law. **Cuing text for BASE** - PDOM help text: - "Look for grab button to play. Once grabbed, press W, A, S, or D key to move up, left, down, or right. Space to release." - Initial Grab Alert: - "Grabbed. At center of Play Area. Has no more negative charges than positive charges. Press W, A, S, or D key to move balloon. Space to release." - Keyboard Dialog content for grabbed balloon: - "Move grabbed balloon up, left, down, or right with Arrow keys or with letter keys W, A, S, or D." **Cuing text for Friction** - PDOM help text: - "Look for grab buttons. Once grabbed, use letter keys W, A, S, or D to move book or zoomed-in book up, left, down, or right." - Initial Grab Alerts before successful interaction: - Not touching: "Grabbed. Lightly on Physics book. Use W, A, S, or D keys to move book. Space to release." - Touching: "Grabbed. Rub fast or slow with A or D keys. Space to release." - and immediately after a grab alert the user also hears, "Atoms jiggle a tiny bit, temperature cool" Keyboard Dialog content Friction's grabbed book: - "Move grabbed book up, left, down, or right with Arrow keys, or with letter keys W, A, S, or D."
1.0
How well are Arrow keys working in JAWS for 4-way custom move? - I am having some trouble assessing what is the best way to cue the custom 4-way drag, especially for Friction where pressing the W key is not so useful. In a recent session with a VoiceOver user familiar with our sims, there seemed to be confusion around cuing only the use of the WASD keys for moving the book. **Goal of this issue:** Assess how well the Arrow keys are working in JAWS so we can figure out the best way to provide instructions for the interaction and the best way to provide alternatives for the interaction. **Question 1 for @jessegreenberg** Do you know how well the Arrow keys are working with JAWS now? In early prototypes of BASE the Arrow keys were not working at all with JAWS, and then they worked, but not as well as the WASD keys. **Question 2 for @terracoda & @emily-phet** If JAWS is no longer having a serious problem with releasing the Arrow keys to the web application, how and where is the best way to provide the alternative WASD key instructions? - In PDOM help text, first grab alert, **AND** Keyboard Shortcuts dialog? - Only in PDOM help text **AND** Keyboard Shortcuts? **OR** - Only in the Keyboard Shortcuts dialog? Currently, we do option 1, and we only explicitly refer to using the Arrow keys in the Keyboard Shortcuts dialog. Here are examples from BASE and Friction, and soon there will be a grab buttons in GFL and Faraday's Law. **Cuing text for BASE** - PDOM help text: - "Look for grab button to play. Once grabbed, press W, A, S, or D key to move up, left, down, or right. Space to release." - Initial Grab Alert: - "Grabbed. At center of Play Area. Has no more negative charges than positive charges. Press W, A, S, or D key to move balloon. Space to release." - Keyboard Dialog content for grabbed balloon: - "Move grabbed balloon up, left, down, or right with Arrow keys or with letter keys W, A, S, or D." **Cuing text for Friction** - PDOM help text: - "Look for grab buttons. Once grabbed, use letter keys W, A, S, or D to move book or zoomed-in book up, left, down, or right." - Initial Grab Alerts before successful interaction: - Not touching: "Grabbed. Lightly on Physics book. Use W, A, S, or D keys to move book. Space to release." - Touching: "Grabbed. Rub fast or slow with A or D keys. Space to release." - and immediately after a grab alert the user also hears, "Atoms jiggle a tiny bit, temperature cool" Keyboard Dialog content Friction's grabbed book: - "Move grabbed book up, left, down, or right with Arrow keys, or with letter keys W, A, S, or D."
non_main
how well are arrow keys working in jaws for way custom move i am having some trouble assessing what is the best way to cue the custom way drag especially for friction where pressing the w key is not so useful in a recent session with a voiceover user familiar with our sims there seemed to be confusion around cuing only the use of the wasd keys for moving the book goal of this issue assess how well the arrow keys are working in jaws so we can figure out the best way to provide instructions for the interaction and the best way to provide alternatives for the interaction question for jessegreenberg do you know how well the arrow keys are working with jaws now in early prototypes of base the arrow keys were not working at all with jaws and then they worked but not as well as the wasd keys question for terracoda emily phet if jaws is no longer having a serious problem with releasing the arrow keys to the web application how and where is the best way to provide the alternative wasd key instructions in pdom help text first grab alert and keyboard shortcuts dialog only in pdom help text and keyboard shortcuts or only in the keyboard shortcuts dialog currently we do option and we only explicitly refer to using the arrow keys in the keyboard shortcuts dialog here are examples from base and friction and soon there will be a grab buttons in gfl and faraday s law cuing text for base pdom help text look for grab button to play once grabbed press w a s or d key to move up left down or right space to release initial grab alert grabbed at center of play area has no more negative charges than positive charges press w a s or d key to move balloon space to release keyboard dialog content for grabbed balloon move grabbed balloon up left down or right with arrow keys or with letter keys w a s or d cuing text for friction pdom help text look for grab buttons once grabbed use letter keys w a s or d to move book or zoomed in book up left down or right initial grab alerts before successful interaction not touching grabbed lightly on physics book use w a s or d keys to move book space to release touching grabbed rub fast or slow with a or d keys space to release and immediately after a grab alert the user also hears atoms jiggle a tiny bit temperature cool keyboard dialog content friction s grabbed book move grabbed book up left down or right with arrow keys or with letter keys w a s or d
0
5,559
27,808,349,143
IssuesEvent
2023-03-17 22:47:54
microsoft/DirectXTex
https://api.github.com/repos/microsoft/DirectXTex
closed
Retire legacy Xbox One XDK support
maintainence
The only scenario that still uses VS 2017 is for the legacy Xbox One XDK. This task is drop support for this older Xbox development model and remove the following projects: ``` DirectXTex_XboxOneXDK_2017.sln DirectXTex_XboxOneXDK_PC_2017.sln ``` > The end-of-life release will be hosted on https://github.com/microsoft/Xbox-ATG-Samples and will likely be the January 2023 release.
True
Retire legacy Xbox One XDK support - The only scenario that still uses VS 2017 is for the legacy Xbox One XDK. This task is drop support for this older Xbox development model and remove the following projects: ``` DirectXTex_XboxOneXDK_2017.sln DirectXTex_XboxOneXDK_PC_2017.sln ``` > The end-of-life release will be hosted on https://github.com/microsoft/Xbox-ATG-Samples and will likely be the January 2023 release.
main
retire legacy xbox one xdk support the only scenario that still uses vs is for the legacy xbox one xdk this task is drop support for this older xbox development model and remove the following projects directxtex xboxonexdk sln directxtex xboxonexdk pc sln the end of life release will be hosted on and will likely be the january release
1
146,332
13,177,933,232
IssuesEvent
2020-08-12 08:16:56
legokor/UPRA-doksik
https://api.github.com/repos/legokor/UPRA-doksik
closed
UPRA for Dummies
documentation
UPRA for Dummies doksik készítéser -~~ÁltalÑnos âsszefoglaló~~ -~~OBC~~ -~~COM~~ -~~EPS~~ -DAU -Interface -BUS -~~GND~~
1.0
UPRA for Dummies - UPRA for Dummies doksik készítéser -~~ÁltalÑnos âsszefoglaló~~ -~~OBC~~ -~~COM~~ -~~EPS~~ -DAU -Interface -BUS -~~GND~~
non_main
upra for dummies upra for dummies doksik kΓ©szΓ­tΓ©ser Γ‘ltalΓ‘nos ΓΆsszefoglalΓ³ obc com eps dau interface bus gnd
0
4,543
23,662,203,313
IssuesEvent
2022-08-26 16:42:03
precice/precice
https://api.github.com/repos/precice/precice
opened
Resetting sent data in coupling scheme to zero after sending affects downstream calculations. Why?
bug maintainability
**Describe your setup** *will provide link to commit in a moment* **Describe the problem** To clean up things and to make sure there are no strange, I changed the source code to reset data after sending and before receiving in the coupling scheme (see link to commit above). However, I observed that some tests are failing after this change. **Step To Reproduce** 1. Check out branch 2. Run tests **Expected behaviour** As far as I know the data should not be used anymore after sending. I also think that this would be the most intuitive behavior. Any idea why this is happening? **Additional context** This is a pure software engineering issue. But it might also be a bug. I'm not sure here.
True
Resetting sent data in coupling scheme to zero after sending affects downstream calculations. Why? - **Describe your setup** *will provide link to commit in a moment* **Describe the problem** To clean up things and to make sure there are no strange, I changed the source code to reset data after sending and before receiving in the coupling scheme (see link to commit above). However, I observed that some tests are failing after this change. **Step To Reproduce** 1. Check out branch 2. Run tests **Expected behaviour** As far as I know the data should not be used anymore after sending. I also think that this would be the most intuitive behavior. Any idea why this is happening? **Additional context** This is a pure software engineering issue. But it might also be a bug. I'm not sure here.
main
resetting sent data in coupling scheme to zero after sending affects downstream calculations why describe your setup will provide link to commit in a moment describe the problem to clean up things and to make sure there are no strange i changed the source code to reset data after sending and before receiving in the coupling scheme see link to commit above however i observed that some tests are failing after this change step to reproduce check out branch run tests expected behaviour as far as i know the data should not be used anymore after sending i also think that this would be the most intuitive behavior any idea why this is happening additional context this is a pure software engineering issue but it might also be a bug i m not sure here
1
129,772
27,559,625,948
IssuesEvent
2023-03-07 20:48:48
openxla/iree
https://api.github.com/repos/openxla/iree
opened
SCF::TileAndFuse produces bad IR for multi result generic op + pack op cases
codegen
I'm working on pack op fusion. I used aggressive fusion to get more data point and found that the SCF tile and fuse does not work well for some cases. If we apply TileAndFuse on `multi result generic ops` and `tensor.pack` ops, it will remain a generic op outside the scf.for loop. E.g., Input IR: ```mlir func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() { %c256 = arith.constant 256 : index %c48 = arith.constant 48 : index %c0 = arith.constant 0 : index %c786432 = arith.constant 786432 : index %c1572864 = arith.constant 1572864 : index %0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> %6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %workgroup_id_x = hal.interface.workgroup.id[0] : index %workgroup_count_x = hal.interface.workgroup.count[0] : index %workgroup_id_y = hal.interface.workgroup.id[1] : index %workgroup_count_y = hal.interface.workgroup.count[1] : index %8 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_id_y] %9 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_count_y] scf.for %arg0 = %8 to %c48 step %9 { %10 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_id_x] %11 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_count_x] scf.for %arg1 = %10 to %c256 step %11 { %12 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32> %13 = affine.apply affine_map<(d0) -> (d0 * 8)>(%arg0) %14 = affine.apply affine_map<(d0) -> (d0 * 2)>(%arg1) %15 = flow.dispatch.tensor.load %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32> %16 = flow.dispatch.tensor.load %0, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %17 = flow.dispatch.tensor.load %1, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %18 = flow.dispatch.tensor.load %2, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %19 = flow.dispatch.tensor.load %3, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %20 = flow.dispatch.tensor.load %4, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %21 = tensor.empty() : tensor<128x64xf32> %22:2 = linalg.generic {indexing_maps = [affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>], iterator_types = ["parallel", "parallel"]} ins(%16, %17, %18, %19, %20 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%21, %15 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} { ^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32): %23 = arith.addf %in, %in_0 : f32 %24 = arith.addf %23, %in_1 : f32 %25 = arith.mulf %24, %in_2 : f32 %26 = arith.addf %25, %in_3 : f32 linalg.yield %24, %26 : f32, f32 } -> (tensor<128x64xf32>, tensor<128x64xf32>) %pack = tensor.pack %22#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %12 {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} : tensor<128x64xf32> -> tensor<16x32x8x2xf32> flow.dispatch.tensor.store %pack, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> flow.dispatch.tensor.store %22#0, %6, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> } } return } ``` After running `iree-opt --linalg-fuse="tile-sizes=1,16" repro.mlir`: ```mlir #config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]> #map = affine_map<()[s0] -> (s0 * 16)> #map1 = affine_map<()[s0] -> (s0 * 32)> #map2 = affine_map<(d0) -> (d0 * 8)> #map3 = affine_map<(d0) -> (d0 * 2)> #map4 = affine_map<(d0, d1) -> (d1)> #map5 = affine_map<(d0, d1) -> (d0, d1)> module { func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() { %c32 = arith.constant 32 : index %c1 = arith.constant 1 : index %c16 = arith.constant 16 : index %c256 = arith.constant 256 : index %c48 = arith.constant 48 : index %c0 = arith.constant 0 : index %c786432 = arith.constant 786432 : index %c1572864 = arith.constant 1572864 : index %0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> %6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %workgroup_id_x = hal.interface.workgroup.id[0] : index %workgroup_count_x = hal.interface.workgroup.count[0] : index %workgroup_id_y = hal.interface.workgroup.id[1] : index %workgroup_count_y = hal.interface.workgroup.count[1] : index %8 = affine.apply #map()[%workgroup_id_y] %9 = affine.apply #map()[%workgroup_count_y] %10 = affine.apply #map1()[%workgroup_id_x] %11 = affine.apply #map1()[%workgroup_count_x] %12 = tensor.empty() : tensor<128x64xf32> scf.for %arg0 = %8 to %c48 step %9 { %13 = affine.apply #map2(%arg0) scf.for %arg1 = %10 to %c256 step %11 { %14 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32> %15 = affine.apply #map3(%arg1) %16 = flow.dispatch.tensor.load %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32> %17 = flow.dispatch.tensor.load %0, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %18 = flow.dispatch.tensor.load %1, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %19 = flow.dispatch.tensor.load %2, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %20 = flow.dispatch.tensor.load %3, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %21 = flow.dispatch.tensor.load %4, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %22:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%17, %18, %19, %20, %21 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%12, %16 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #config} { ^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32): %24 = arith.addf %in, %in_0 : f32 %25 = arith.addf %24, %in_1 : f32 %26 = arith.mulf %25, %in_2 : f32 %27 = arith.addf %26, %in_3 : f32 linalg.yield %25, %27 : f32, f32 } -> (tensor<128x64xf32>, tensor<128x64xf32>) %23:2 = scf.for %arg2 = %c0 to %c16 step %c1 iter_args(%arg3 = %14, %arg4 = %12) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) { %24 = affine.apply #map2(%arg2) %25:2 = scf.for %arg5 = %c0 to %c32 step %c16 iter_args(%arg6 = %arg3, %arg7 = %arg4) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) { %26 = affine.apply #map3(%arg5) %extracted_slice = tensor.extract_slice %17[%26] [32] [1] : tensor<64xf32> to tensor<32xf32> %extracted_slice_0 = tensor.extract_slice %18[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %extracted_slice_1 = tensor.extract_slice %19[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %extracted_slice_2 = tensor.extract_slice %20[%26] [32] [1] : tensor<64xf32> to tensor<32xf32> %extracted_slice_3 = tensor.extract_slice %21[%26] [32] [1] : tensor<64xf32> to tensor<32xf32> %extracted_slice_4 = tensor.extract_slice %arg7[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %extracted_slice_5 = tensor.extract_slice %16[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %27:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%extracted_slice, %extracted_slice_0, %extracted_slice_1, %extracted_slice_2, %extracted_slice_3 : tensor<32xf32>, tensor<8x32xf32>, tensor<8x32xf32>, tensor<32xf32>, tensor<32xf32>) outs(%extracted_slice_4, %extracted_slice_5 : tensor<8x32xf32>, tensor<8x32xf32>) attrs = {lowering_config = #config} { ^bb0(%in: f32, %in_8: f32, %in_9: f32, %in_10: f32, %in_11: f32, %out: f32, %out_12: f32): %28 = arith.addf %in, %in_8 : f32 %29 = arith.addf %28, %in_9 : f32 %30 = arith.mulf %29, %in_10 : f32 %31 = arith.addf %30, %in_11 : f32 linalg.yield %29, %31 : f32, f32 } -> (tensor<8x32xf32>, tensor<8x32xf32>) %extracted_slice_6 = tensor.extract_slice %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<16x32x8x2xf32> to tensor<1x16x8x2xf32> %pack = tensor.pack %27#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %extracted_slice_6 {__internal_linalg_transform__ = "1", lowering_config = #config} : tensor<8x32xf32> -> tensor<1x16x8x2xf32> %inserted_slice = tensor.insert_slice %pack into %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<1x16x8x2xf32> into tensor<16x32x8x2xf32> %inserted_slice_7 = tensor.insert_slice %27#0 into %arg7[%24, %26] [8, 32] [1, 1] : tensor<8x32xf32> into tensor<128x64xf32> scf.yield %inserted_slice, %inserted_slice_7 : tensor<16x32x8x2xf32>, tensor<128x64xf32> } scf.yield %25#0, %25#1 : tensor<16x32x8x2xf32>, tensor<128x64xf32> } flow.dispatch.tensor.store %23#0, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> flow.dispatch.tensor.store %23#1, %6, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> } } return } } ``` What I'm expecting is that the `%22` is removed and the scf.for loop should return three values. This drops the performance a lot.
1.0
SCF::TileAndFuse produces bad IR for multi result generic op + pack op cases - I'm working on pack op fusion. I used aggressive fusion to get more data point and found that the SCF tile and fuse does not work well for some cases. If we apply TileAndFuse on `multi result generic ops` and `tensor.pack` ops, it will remain a generic op outside the scf.for loop. E.g., Input IR: ```mlir func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() { %c256 = arith.constant 256 : index %c48 = arith.constant 48 : index %c0 = arith.constant 0 : index %c786432 = arith.constant 786432 : index %c1572864 = arith.constant 1572864 : index %0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> %6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %workgroup_id_x = hal.interface.workgroup.id[0] : index %workgroup_count_x = hal.interface.workgroup.count[0] : index %workgroup_id_y = hal.interface.workgroup.id[1] : index %workgroup_count_y = hal.interface.workgroup.count[1] : index %8 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_id_y] %9 = affine.apply affine_map<()[s0] -> (s0 * 16)>()[%workgroup_count_y] scf.for %arg0 = %8 to %c48 step %9 { %10 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_id_x] %11 = affine.apply affine_map<()[s0] -> (s0 * 32)>()[%workgroup_count_x] scf.for %arg1 = %10 to %c256 step %11 { %12 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32> %13 = affine.apply affine_map<(d0) -> (d0 * 8)>(%arg0) %14 = affine.apply affine_map<(d0) -> (d0 * 2)>(%arg1) %15 = flow.dispatch.tensor.load %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32> %16 = flow.dispatch.tensor.load %0, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %17 = flow.dispatch.tensor.load %1, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %18 = flow.dispatch.tensor.load %2, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %19 = flow.dispatch.tensor.load %3, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %20 = flow.dispatch.tensor.load %4, offsets = [%14], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %21 = tensor.empty() : tensor<128x64xf32> %22:2 = linalg.generic {indexing_maps = [affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d1)>, affine_map<(d0, d1) -> (d0, d1)>, affine_map<(d0, d1) -> (d0, d1)>], iterator_types = ["parallel", "parallel"]} ins(%16, %17, %18, %19, %20 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%21, %15 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} { ^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32): %23 = arith.addf %in, %in_0 : f32 %24 = arith.addf %23, %in_1 : f32 %25 = arith.mulf %24, %in_2 : f32 %26 = arith.addf %25, %in_3 : f32 linalg.yield %24, %26 : f32, f32 } -> (tensor<128x64xf32>, tensor<128x64xf32>) %pack = tensor.pack %22#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %12 {lowering_config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]>} : tensor<128x64xf32> -> tensor<16x32x8x2xf32> flow.dispatch.tensor.store %pack, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> flow.dispatch.tensor.store %22#0, %6, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %14], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> } } return } ``` After running `iree-opt --linalg-fuse="tile-sizes=1,16" repro.mlir`: ```mlir #config = #iree_codegen.lowering_config<tile_sizes = [[16, 32], [1, 16], [0, 0]]> #map = affine_map<()[s0] -> (s0 * 16)> #map1 = affine_map<()[s0] -> (s0 * 32)> #map2 = affine_map<(d0) -> (d0 * 8)> #map3 = affine_map<(d0) -> (d0 * 2)> #map4 = affine_map<(d0, d1) -> (d1)> #map5 = affine_map<(d0, d1) -> (d0, d1)> module { func.func @main_dispatch_114_generic_384x512_dispatch_0_generic_384x512() { %c32 = arith.constant 32 : index %c1 = arith.constant 1 : index %c16 = arith.constant 16 : index %c256 = arith.constant 256 : index %c48 = arith.constant 48 : index %c0 = arith.constant 0 : index %c786432 = arith.constant 786432 : index %c1572864 = arith.constant 1572864 : index %0 = hal.interface.binding.subspan set(0) binding(0) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %1 = hal.interface.binding.subspan set(0) binding(1) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %2 = hal.interface.binding.subspan set(0) binding(2) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> %3 = hal.interface.binding.subspan set(0) binding(3) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %4 = hal.interface.binding.subspan set(0) binding(4) type(storage_buffer) alignment(64) offset(%c0) flags(ReadOnly) : !flow.dispatch.tensor<readonly:tensor<512xf32>> %5 = hal.interface.binding.subspan set(0) binding(5) type(storage_buffer) alignment(64) offset(%c0) : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> %6 = hal.interface.binding.subspan set(0) binding(6) type(storage_buffer) alignment(64) offset(%c786432) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %7 = hal.interface.binding.subspan set(0) binding(7) type(storage_buffer) alignment(64) offset(%c1572864) : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> %workgroup_id_x = hal.interface.workgroup.id[0] : index %workgroup_count_x = hal.interface.workgroup.count[0] : index %workgroup_id_y = hal.interface.workgroup.id[1] : index %workgroup_count_y = hal.interface.workgroup.count[1] : index %8 = affine.apply #map()[%workgroup_id_y] %9 = affine.apply #map()[%workgroup_count_y] %10 = affine.apply #map1()[%workgroup_id_x] %11 = affine.apply #map1()[%workgroup_count_x] %12 = tensor.empty() : tensor<128x64xf32> scf.for %arg0 = %8 to %c48 step %9 { %13 = affine.apply #map2(%arg0) scf.for %arg1 = %10 to %c256 step %11 { %14 = flow.dispatch.tensor.load %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> -> tensor<16x32x8x2xf32> %15 = affine.apply #map3(%arg1) %16 = flow.dispatch.tensor.load %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> -> tensor<128x64xf32> %17 = flow.dispatch.tensor.load %0, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %18 = flow.dispatch.tensor.load %1, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %19 = flow.dispatch.tensor.load %2, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : !flow.dispatch.tensor<readonly:tensor<384x512xf32>> -> tensor<128x64xf32> %20 = flow.dispatch.tensor.load %3, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %21 = flow.dispatch.tensor.load %4, offsets = [%15], sizes = [64], strides = [1] : !flow.dispatch.tensor<readonly:tensor<512xf32>> -> tensor<64xf32> %22:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%17, %18, %19, %20, %21 : tensor<64xf32>, tensor<128x64xf32>, tensor<128x64xf32>, tensor<64xf32>, tensor<64xf32>) outs(%12, %16 : tensor<128x64xf32>, tensor<128x64xf32>) attrs = {lowering_config = #config} { ^bb0(%in: f32, %in_0: f32, %in_1: f32, %in_2: f32, %in_3: f32, %out: f32, %out_4: f32): %24 = arith.addf %in, %in_0 : f32 %25 = arith.addf %24, %in_1 : f32 %26 = arith.mulf %25, %in_2 : f32 %27 = arith.addf %26, %in_3 : f32 linalg.yield %25, %27 : f32, f32 } -> (tensor<128x64xf32>, tensor<128x64xf32>) %23:2 = scf.for %arg2 = %c0 to %c16 step %c1 iter_args(%arg3 = %14, %arg4 = %12) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) { %24 = affine.apply #map2(%arg2) %25:2 = scf.for %arg5 = %c0 to %c32 step %c16 iter_args(%arg6 = %arg3, %arg7 = %arg4) -> (tensor<16x32x8x2xf32>, tensor<128x64xf32>) { %26 = affine.apply #map3(%arg5) %extracted_slice = tensor.extract_slice %17[%26] [32] [1] : tensor<64xf32> to tensor<32xf32> %extracted_slice_0 = tensor.extract_slice %18[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %extracted_slice_1 = tensor.extract_slice %19[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %extracted_slice_2 = tensor.extract_slice %20[%26] [32] [1] : tensor<64xf32> to tensor<32xf32> %extracted_slice_3 = tensor.extract_slice %21[%26] [32] [1] : tensor<64xf32> to tensor<32xf32> %extracted_slice_4 = tensor.extract_slice %arg7[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %extracted_slice_5 = tensor.extract_slice %16[%24, %26] [8, 32] [1, 1] : tensor<128x64xf32> to tensor<8x32xf32> %27:2 = linalg.generic {indexing_maps = [#map4, #map5, #map5, #map4, #map4, #map5, #map5], iterator_types = ["parallel", "parallel"]} ins(%extracted_slice, %extracted_slice_0, %extracted_slice_1, %extracted_slice_2, %extracted_slice_3 : tensor<32xf32>, tensor<8x32xf32>, tensor<8x32xf32>, tensor<32xf32>, tensor<32xf32>) outs(%extracted_slice_4, %extracted_slice_5 : tensor<8x32xf32>, tensor<8x32xf32>) attrs = {lowering_config = #config} { ^bb0(%in: f32, %in_8: f32, %in_9: f32, %in_10: f32, %in_11: f32, %out: f32, %out_12: f32): %28 = arith.addf %in, %in_8 : f32 %29 = arith.addf %28, %in_9 : f32 %30 = arith.mulf %29, %in_10 : f32 %31 = arith.addf %30, %in_11 : f32 linalg.yield %29, %31 : f32, f32 } -> (tensor<8x32xf32>, tensor<8x32xf32>) %extracted_slice_6 = tensor.extract_slice %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<16x32x8x2xf32> to tensor<1x16x8x2xf32> %pack = tensor.pack %27#0 inner_dims_pos = [0, 1] inner_tiles = [8, 2] into %extracted_slice_6 {__internal_linalg_transform__ = "1", lowering_config = #config} : tensor<8x32xf32> -> tensor<1x16x8x2xf32> %inserted_slice = tensor.insert_slice %pack into %arg6[%arg2, %arg5, 0, 0] [1, 16, 8, 2] [1, 1, 1, 1] : tensor<1x16x8x2xf32> into tensor<16x32x8x2xf32> %inserted_slice_7 = tensor.insert_slice %27#0 into %arg7[%24, %26] [8, 32] [1, 1] : tensor<8x32xf32> into tensor<128x64xf32> scf.yield %inserted_slice, %inserted_slice_7 : tensor<16x32x8x2xf32>, tensor<128x64xf32> } scf.yield %25#0, %25#1 : tensor<16x32x8x2xf32>, tensor<128x64xf32> } flow.dispatch.tensor.store %23#0, %5, offsets = [%arg0, %arg1, 0, 0], sizes = [16, 32, 8, 2], strides = [1, 1, 1, 1] : tensor<16x32x8x2xf32> -> !flow.dispatch.tensor<writeonly:tensor<48x256x8x2xf32>> flow.dispatch.tensor.store %23#1, %6, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> flow.dispatch.tensor.store %22#1, %7, offsets = [%13, %15], sizes = [128, 64], strides = [1, 1] : tensor<128x64xf32> -> !flow.dispatch.tensor<writeonly:tensor<384x512xf32>> } } return } } ``` What I'm expecting is that the `%22` is removed and the scf.for loop should return three values. This drops the performance a lot.
non_main
scf tileandfuse produces bad ir for multi result generic op pack op cases i m working on pack op fusion i used aggressive fusion to get more data point and found that the scf tile and fuse does not work well for some cases if we apply tileandfuse on multi result generic ops and tensor pack ops it will remain a generic op outside the scf for loop e g input ir mlir func func main dispatch generic dispatch generic arith constant index arith constant index arith constant index arith constant index arith constant index hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor workgroup id x hal interface workgroup id index workgroup count x hal interface workgroup count index workgroup id y hal interface workgroup id index workgroup count y hal interface workgroup count index affine apply affine map affine apply affine map scf for to step affine apply affine map affine apply affine map scf for to step flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor affine apply affine map affine apply affine map flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor tensor empty tensor linalg generic indexing maps iterator types ins tensor tensor tensor tensor tensor outs tensor tensor attrs lowering config iree codegen lowering config in in in in in out out arith addf in in arith addf in arith mulf in arith addf in linalg yield tensor tensor pack tensor pack inner dims pos inner tiles into lowering config iree codegen lowering config tensor tensor flow dispatch tensor store pack offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor return after running iree opt linalg fuse tile sizes repro mlir mlir config iree codegen lowering config map affine map affine map affine map affine map affine map affine map module func func main dispatch generic dispatch generic arith constant index arith constant index arith constant index arith constant index arith constant index arith constant index arith constant index arith constant index hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flags readonly flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor hal interface binding subspan set binding type storage buffer alignment offset flow dispatch tensor workgroup id x hal interface workgroup id index workgroup count x hal interface workgroup count index workgroup id y hal interface workgroup id index workgroup count y hal interface workgroup count index affine apply map affine apply map affine apply affine apply tensor empty tensor scf for to step affine apply scf for to step flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor affine apply flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor flow dispatch tensor load offsets sizes strides flow dispatch tensor tensor linalg generic indexing maps iterator types ins tensor tensor tensor tensor tensor outs tensor tensor attrs lowering config config in in in in in out out arith addf in in arith addf in arith mulf in arith addf in linalg yield tensor tensor scf for to step iter args tensor tensor affine apply scf for to step iter args tensor tensor affine apply extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor extracted slice tensor extract slice tensor to tensor linalg generic indexing maps iterator types ins extracted slice extracted slice extracted slice extracted slice extracted slice tensor tensor tensor tensor tensor outs extracted slice extracted slice tensor tensor attrs lowering config config in in in in in out out arith addf in in arith addf in arith mulf in arith addf in linalg yield tensor tensor extracted slice tensor extract slice tensor to tensor pack tensor pack inner dims pos inner tiles into extracted slice internal linalg transform lowering config config tensor tensor inserted slice tensor insert slice pack into tensor into tensor inserted slice tensor insert slice into tensor into tensor scf yield inserted slice inserted slice tensor tensor scf yield tensor tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor flow dispatch tensor store offsets sizes strides tensor flow dispatch tensor return what i m expecting is that the is removed and the scf for loop should return three values this drops the performance a lot
0
390,964
26,876,255,619
IssuesEvent
2023-02-05 03:26:24
RodrigoTeran/syntactically-awesome-react-app
https://api.github.com/repos/RodrigoTeran/syntactically-awesome-react-app
closed
Add strict node and git dependencies to use create-sara-project
documentation priority:low
# ✳️ Add strict node and git dependencies to use create-sara-project Dependencies: - ``` npm@8.9.0 ``` - ``` node@v16.17.1 ``` - ``` git@2.28.0 ``` Add them to the [Readme.md](https://github.com/RodrigoTeran/syntactically-awesome-react-app#readme)
1.0
Add strict node and git dependencies to use create-sara-project - # ✳️ Add strict node and git dependencies to use create-sara-project Dependencies: - ``` npm@8.9.0 ``` - ``` node@v16.17.1 ``` - ``` git@2.28.0 ``` Add them to the [Readme.md](https://github.com/RodrigoTeran/syntactically-awesome-react-app#readme)
non_main
add strict node and git dependencies to use create sara project ✳️ add strict node and git dependencies to use create sara project dependencies npm node git add them to the
0
1,387
6,015,202,230
IssuesEvent
2017-06-07 00:57:20
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_robocopy module don't handle all posible exit codes
affects_2.2 bug_report waiting_on_maintainer windows
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> win_robocopy module ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` 2.2.0.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> ##### SUMMARY <!--- Explain the problem briefly --> module handle only 0,1,2,4,8,16 robocopy exit codes but if custom flags used exit codes may eq combined value: These can be combined, giving a few extra exit codes: 0Γ—03 3 (2+1) Some files were copied. Additional files were present. No failure was encountered. 0Γ—05 5 (4+1) Some files were copied. Some files were mismatched. No failure was encountered. 0Γ—06 6 (4+2) Additional files and mismatched files exist. No files were copied and no failures were encountered. This means that the files already exist in the destination directory 0Γ—07 7 (4+1+2) Files were copied, a file mismatch was present, and additional files were present. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` ```
True
win_robocopy module don't handle all posible exit codes - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> win_robocopy module ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` 2.2.0.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> ##### SUMMARY <!--- Explain the problem briefly --> module handle only 0,1,2,4,8,16 robocopy exit codes but if custom flags used exit codes may eq combined value: These can be combined, giving a few extra exit codes: 0Γ—03 3 (2+1) Some files were copied. Additional files were present. No failure was encountered. 0Γ—05 5 (4+1) Some files were copied. Some files were mismatched. No failure was encountered. 0Γ—06 6 (4+2) Additional files and mismatched files exist. No files were copied and no failures were encountered. This means that the files already exist in the destination directory 0Γ—07 7 (4+1+2) Files were copied, a file mismatch was present, and additional files were present. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` ```
main
win robocopy module don t handle all posible exit codes issue type bug report component name win robocopy module ansible version configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say β€œn a” for anything that is not platform specific summary module handle only robocopy exit codes but if custom flags used exit codes may eq combined value these can be combined giving a few extra exit codes Γ— some files were copied additional files were present no failure was encountered Γ— some files were copied some files were mismatched no failure was encountered Γ— additional files and mismatched files exist no files were copied and no failures were encountered this means that the files already exist in the destination directory Γ— files were copied a file mismatch was present and additional files were present steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used expected results actual results
1
77,948
22,049,719,371
IssuesEvent
2022-05-30 07:31:14
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
opened
Code coverage error should not fail a build
area/build kind/engineering
Failures like this one: https://github.com/pulumi/pulumi/runs/6614209710?check_suite_focus=true#step:40:25 ![image](https://user-images.githubusercontent.com/1454008/170940387-0138eb65-e612-4c5a-9d84-5b0690860227.png) should not fail the whole build.
1.0
Code coverage error should not fail a build - Failures like this one: https://github.com/pulumi/pulumi/runs/6614209710?check_suite_focus=true#step:40:25 ![image](https://user-images.githubusercontent.com/1454008/170940387-0138eb65-e612-4c5a-9d84-5b0690860227.png) should not fail the whole build.
non_main
code coverage error should not fail a build failures like this one should not fail the whole build
0
4,491
23,387,863,868
IssuesEvent
2022-08-11 15:07:52
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Improve BreadcrumbSelector component
type: enhancement work: frontend status: ready restricted: new maintainers
## Current behavior - This is the `BreadcrumbSelector` component: ![image](https://user-images.githubusercontent.com/42411/184004827-4bab2c6c-fc23-4768-b50b-1ffd07aa8c03.png) - It is used both for selecting a Schema within the current Database and for selecting a Table or Exploration within the current Schema. ## Desired behavior - It should look more like this: ![mockup](https://user-images.githubusercontent.com/42411/180277150-e68cf483-9ce5-436f-b347-714937d22295.png) Specifically... - A search input should exist and be focused when the BreadcrumbSelector opens. Search queries should filter the entries across all categories. - Entries should highlight the substring of their label which matches the search query. - (Not shown in the mockup) An "Add New" option should appear within each category. To implement this, I'd recommend starting by changing... ```ts export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorEntry[]>; ``` to... ```ts interface BreadcrumbSelectorAddNewViaLink { type: 'link'; href: string; } interface BreadcrumbSelectorAddNewViaButton { type: 'button'; onSubmit: () => Promise<void>; } interface BreadcrumbSelectorSection { entries: BreadcrumbSelectorEntry[]; addNew?: BreadcrumbSelectorAddNewViaLink | BreadcrumbSelectorAddNewViaButton; } export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorSection>; ``` - If the URL for the entry matches _the start_ of the router's current URL, then the entry should visually indicate that it's active. (It's important to match the start because we want to show the active schema when we're on the Table Page, for example.) - Vertical scrolling should not happen so easily. We'll need to increase `max-height` somewhere. A good value might be something like `calc(100vh - 5em)`.
True
Improve BreadcrumbSelector component - ## Current behavior - This is the `BreadcrumbSelector` component: ![image](https://user-images.githubusercontent.com/42411/184004827-4bab2c6c-fc23-4768-b50b-1ffd07aa8c03.png) - It is used both for selecting a Schema within the current Database and for selecting a Table or Exploration within the current Schema. ## Desired behavior - It should look more like this: ![mockup](https://user-images.githubusercontent.com/42411/180277150-e68cf483-9ce5-436f-b347-714937d22295.png) Specifically... - A search input should exist and be focused when the BreadcrumbSelector opens. Search queries should filter the entries across all categories. - Entries should highlight the substring of their label which matches the search query. - (Not shown in the mockup) An "Add New" option should appear within each category. To implement this, I'd recommend starting by changing... ```ts export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorEntry[]>; ``` to... ```ts interface BreadcrumbSelectorAddNewViaLink { type: 'link'; href: string; } interface BreadcrumbSelectorAddNewViaButton { type: 'button'; onSubmit: () => Promise<void>; } interface BreadcrumbSelectorSection { entries: BreadcrumbSelectorEntry[]; addNew?: BreadcrumbSelectorAddNewViaLink | BreadcrumbSelectorAddNewViaButton; } export type BreadcrumbSelectorData = Map<string, BreadcrumbSelectorSection>; ``` - If the URL for the entry matches _the start_ of the router's current URL, then the entry should visually indicate that it's active. (It's important to match the start because we want to show the active schema when we're on the Table Page, for example.) - Vertical scrolling should not happen so easily. We'll need to increase `max-height` somewhere. A good value might be something like `calc(100vh - 5em)`.
main
improve breadcrumbselector component current behavior this is the breadcrumbselector component it is used both for selecting a schema within the current database and for selecting a table or exploration within the current schema desired behavior it should look more like this specifically a search input should exist and be focused when the breadcrumbselector opens search queries should filter the entries across all categories entries should highlight the substring of their label which matches the search query not shown in the mockup an add new option should appear within each category to implement this i d recommend starting by changing ts export type breadcrumbselectordata map to ts interface breadcrumbselectoraddnewvialink type link href string interface breadcrumbselectoraddnewviabutton type button onsubmit promise interface breadcrumbselectorsection entries breadcrumbselectorentry addnew breadcrumbselectoraddnewvialink breadcrumbselectoraddnewviabutton export type breadcrumbselectordata map if the url for the entry matches the start of the router s current url then the entry should visually indicate that it s active it s important to match the start because we want to show the active schema when we re on the table page for example vertical scrolling should not happen so easily we ll need to increase max height somewhere a good value might be something like calc
1
2,939
10,548,952,294
IssuesEvent
2019-10-03 07:30:41
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Huawei CloudEngine modules do not work with Ansible 2.8.5
affects_2.8 bug module needs_maintainer needs_triage source_control support:community traceback
##### SUMMARY No ce_* module work with Ansible 2.8.5 I just try your example command : ` ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=192.168.1.1 port=22 username=huawei password=huawei123" localhost --connection local ` I have the same issue with all modules 'ce_' The first time I was this error : ` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/ce.py", line 35, in <module> from ansible.module_utils.network_common import to_list, ComplexList ImportError: No module named network_common ` I replaced in file `from ansible.module_utils.network_common import to_list, ComplexList` by `from ansible.module_utils.network.common.utils import to_list, ComplexList` Now I have an other error message : ` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__ raise AssertionError('socket_path must be a value') AssertionError: socket_path must be a value ` ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ce_ modules not working in Ansible 2.8.5 ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ``` ansible 2.8.5 config file = /etc/ansible/P3/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/dist-packages/ansible executable location = /usr/bin/ansible python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ``` ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local ``` ##### OS / ENVIRONMENT Debian 10 ##### STEPS TO REPRODUCE ``` ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local ``` ##### EXPECTED RESULTS Vlan List ##### ACTUAL RESULTS Error Message : **socket_path must be a value** ``` ansible 2.8.5 config file = /etc/ansible/P3/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/dist-packages/ansible executable location = /usr/bin/ansible python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0] Using /etc/ansible/P3/ansible.cfg as config file setting up inventory plugins host_list declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method script declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method auto declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method Parsed /etc/ansible/P3/hosts inventory source with ini plugin Loading callback plugin minimal of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/minimal.pyc META: ran handlers <127.0.0.1> connection transport is cli <10.10.10.4> using connection plugin network_cli <10.10.10.4> socket_path: /root/.ansible/pc/cb0005b4c1 <10.10.10.4> exec_command(), socket_path=None The full traceback is: Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 145, in run res = self._execute() File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 664, in _execute result = self._handler.run(task_vars=variables) File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/ce.py", line 79, in run rc, out, err = connection.exec_command('open_shell()') File "/usr/lib/python2.7/dist-packages/ansible/plugins/connection/persistent.py", line 52, in exec_command connection = SocketConnection(self.socket_path) File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__ raise AssertionError('socket_path must be a value') AssertionError: socket_path must be a value localhost | FAILED! => { "msg": "Unexpected failure during module execution.", "stdout": "" } ```
True
Huawei CloudEngine modules do not work with Ansible 2.8.5 - ##### SUMMARY No ce_* module work with Ansible 2.8.5 I just try your example command : ` ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=192.168.1.1 port=22 username=huawei password=huawei123" localhost --connection local ` I have the same issue with all modules 'ce_' The first time I was this error : ` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/ce.py", line 35, in <module> from ansible.module_utils.network_common import to_list, ComplexList ImportError: No module named network_common ` I replaced in file `from ansible.module_utils.network_common import to_list, ComplexList` by `from ansible.module_utils.network.common.utils import to_list, ComplexList` Now I have an other error message : ` File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__ raise AssertionError('socket_path must be a value') AssertionError: socket_path must be a value ` ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ce_ modules not working in Ansible 2.8.5 ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ``` ansible 2.8.5 config file = /etc/ansible/P3/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/dist-packages/ansible executable location = /usr/bin/ansible python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ``` ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local ``` ##### OS / ENVIRONMENT Debian 10 ##### STEPS TO REPRODUCE ``` ansible -m ce_command -a "commands='display vlan summary' transport='cli' host=10.10.10.4 port=22 username=admin password=admin" localhost --connection local ``` ##### EXPECTED RESULTS Vlan List ##### ACTUAL RESULTS Error Message : **socket_path must be a value** ``` ansible 2.8.5 config file = /etc/ansible/P3/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/dist-packages/ansible executable location = /usr/bin/ansible python version = 2.7.16 (default, Apr 6 2019, 01:42:57) [GCC 8.3.0] Using /etc/ansible/P3/ansible.cfg as config file setting up inventory plugins host_list declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method script declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method auto declined parsing /etc/ansible/P3/hosts as it did not pass it's verify_file() method Parsed /etc/ansible/P3/hosts inventory source with ini plugin Loading callback plugin minimal of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/minimal.pyc META: ran handlers <127.0.0.1> connection transport is cli <10.10.10.4> using connection plugin network_cli <10.10.10.4> socket_path: /root/.ansible/pc/cb0005b4c1 <10.10.10.4> exec_command(), socket_path=None The full traceback is: Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 145, in run res = self._execute() File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 664, in _execute result = self._handler.run(task_vars=variables) File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/ce.py", line 79, in run rc, out, err = connection.exec_command('open_shell()') File "/usr/lib/python2.7/dist-packages/ansible/plugins/connection/persistent.py", line 52, in exec_command connection = SocketConnection(self.socket_path) File "/usr/lib/python2.7/dist-packages/ansible/module_utils/connection.py", line 122, in __init__ raise AssertionError('socket_path must be a value') AssertionError: socket_path must be a value localhost | FAILED! => { "msg": "Unexpected failure during module execution.", "stdout": "" } ```
main
huawei cloudengine modules do not work with ansible summary no ce module work with ansible i just try your example command ansible m ce command a commands display vlan summary transport cli host port username huawei password localhost connection local i have the same issue with all modules ce the first time i was this error file usr lib dist packages ansible module utils ce py line in from ansible module utils network common import to list complexlist importerror no module named network common i replaced in file from ansible module utils network common import to list complexlist by from ansible module utils network common utils import to list complexlist now i have an other error message file usr lib dist packages ansible module utils connection py line in init raise assertionerror socket path must be a value assertionerror socket path must be a value issue type bug report component name ce modules not working in ansible ansible version ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib dist packages ansible executable location usr bin ansible python version default apr configuration ansible m ce command a commands display vlan summary transport cli host port username admin password admin localhost connection local os environment debian steps to reproduce ansible m ce command a commands display vlan summary transport cli host port username admin password admin localhost connection local expected results vlan list actual results error message socket path must be a value ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib dist packages ansible executable location usr bin ansible python version default apr using etc ansible ansible cfg as config file setting up inventory plugins host list declined parsing etc ansible hosts as it did not pass it s verify file method script declined parsing etc ansible hosts as it did not pass it s verify file method auto declined parsing etc ansible hosts as it did not pass it s verify file method parsed etc ansible hosts inventory source with ini plugin loading callback plugin minimal of type stdout from usr lib dist packages ansible plugins callback minimal pyc meta ran handlers connection transport is cli using connection plugin network cli socket path root ansible pc exec command socket path none the full traceback is traceback most recent call last file usr lib dist packages ansible executor task executor py line in run res self execute file usr lib dist packages ansible executor task executor py line in execute result self handler run task vars variables file usr lib dist packages ansible plugins action ce py line in run rc out err connection exec command open shell file usr lib dist packages ansible plugins connection persistent py line in exec command connection socketconnection self socket path file usr lib dist packages ansible module utils connection py line in init raise assertionerror socket path must be a value assertionerror socket path must be a value localhost failed msg unexpected failure during module execution stdout
1
146,397
19,403,573,001
IssuesEvent
2021-12-19 16:08:14
victorlmneves/fed-pug-boilerplate-v2
https://api.github.com/repos/victorlmneves/fed-pug-boilerplate-v2
closed
CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz - autoclosed
security vulnerability
## CVE-2018-20821 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: fed-pug-boilerplate-v2/package.json</p> <p>Path to vulnerable library: fed-pug-boilerplate-v2/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - node-sass-middleware-0.11.0.tgz (Root Library) - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/victorlmneves/fed-pug-boilerplate-v2/commit/473ea3597a89ac9b7c4f4d251f4b4c119b4643eb">473ea3597a89ac9b7c4f4d251f4b4c119b4643eb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p> <p>Release Date: 2019-04-23</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20821 (Medium) detected in opennmsopennms-source-26.0.0-1, node-sass-4.14.1.tgz - autoclosed - ## CVE-2018-20821 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>opennmsopennms-source-26.0.0-1</b>, <b>node-sass-4.14.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: fed-pug-boilerplate-v2/package.json</p> <p>Path to vulnerable library: fed-pug-boilerplate-v2/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - node-sass-middleware-0.11.0.tgz (Root Library) - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/victorlmneves/fed-pug-boilerplate-v2/commit/473ea3597a89ac9b7c4f4d251f4b4c119b4643eb">473ea3597a89ac9b7c4f4d251f4b4c119b4643eb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsing component in LibSass through 3.5.5 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Parser::parse_css_variable_value in parser.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20821>CVE-2018-20821</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20821</a></p> <p>Release Date: 2019-04-23</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in opennmsopennms source node sass tgz autoclosed cve medium severity vulnerability vulnerable libraries opennmsopennms source node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file fed pug boilerplate package json path to vulnerable library fed pug boilerplate node modules node sass package json dependency hierarchy node sass middleware tgz root library x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details the parsing component in libsass through allows attackers to cause a denial of service uncontrolled recursion in sass parser parse css variable value in parser cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
9,812
3,321,781,139
IssuesEvent
2015-11-09 10:53:55
interactivethings/catalog
https://api.github.com/repos/interactivethings/catalog
opened
Specimen Documentation
documentation
All specimen documentation pages should follow the same structure: - A brief description - Markdown API (options & JSON configuration) - React API (don't worry right now) - Examples with copy-n-pasteable code snippets
1.0
Specimen Documentation - All specimen documentation pages should follow the same structure: - A brief description - Markdown API (options & JSON configuration) - React API (don't worry right now) - Examples with copy-n-pasteable code snippets
non_main
specimen documentation all specimen documentation pages should follow the same structure a brief description markdown api options json configuration react api don t worry right now examples with copy n pasteable code snippets
0
374,108
11,072,081,075
IssuesEvent
2019-12-12 09:36:01
aspnetboilerplate/aspnetboilerplate
https://api.github.com/repos/aspnetboilerplate/aspnetboilerplate
closed
Does UnitTest Project Support EF Core Lazy Loading?
priority:high problem
I want to let Unit Test project support lazy loading feature which introduced in EF Core 2.1. Ref : https://docs.microsoft.com/en-us/ef/core/querying/related-data#lazy-loading So, I Performed following steps try to reach the goal: **Step1.** Install-Package Microsoft.EntityFrameworkCore.Proxies on Unit Test Project. **Step2.** Add services.AddEntityFrameworkProxies(); statement in function Register(IIocManager iocManager) of ServiceCollectionRegistrar.cs **Step3.** In same file, add use lazy loading proxy like this : builder.UseLazyLoadingProxies().UseInMemoryDatabase(Guid.NewGuid().ToString()).UseInternalServiceProvider(serviceProvider); ABP package version: 4.0.1 Base on .Net Core When I ran the test case, I got this exception message: Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method. Any idea or suggestion about this problem? Thanks in advance! PS: ```` Result StackTrace: at Castle.MicroKernel.ComponentActivator.FactoryMethodActivator`1.Instantiate(CreationContext context) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context) at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated) at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21 at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden) at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired) at Castle.MicroKernel.Resolvers.DefaultDependencyResolver.Resolve(CreationContext context, ISubDependencyResolver contextHandlerResolver, ComponentModel model, DependencyModel dependency) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateConstructorArguments(ConstructorCandidate constructor, CreationContext context) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.Instantiate(CreationContext context) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context) at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated) at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21 at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden) at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired) at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy) at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.ResolveAll(Type service, IDictionary arguments, IReleasePolicy policy) at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.ResolveInstanceOrNull(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 86 at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.GetServiceInternal(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 55 at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetService[T](IServiceProvider provider) at Microsoft.EntityFrameworkCore.Proxies.Internal.ProxiesOptionsExtension.Validate(IDbContextOptions options) at Microsoft.EntityFrameworkCore.Internal.ServiceProviderCache.GetOrAdd(IDbContextOptions options, Boolean providerRequired) at Microsoft.EntityFrameworkCore.DbContext..ctor(DbContextOptions options) at Abp.Zero.EntityFrameworkCore.AbpZeroDbContext`4..ctor(DbContextOptions`1 options) in D:\Github\aspnetboilerplate\src\Abp.ZeroCore.EntityFrameworkCore\Zero\EntityFrameworkCore\AbpZeroDbContext.cs:line 68 at AbpTest.EntityFrameworkCore.AbpTestDbContext..ctor(DbContextOptions`1 options) in C:\Workspace\Lab\4.0.1\aspnet-core\src\AbpTest.EntityFrameworkCore\EntityFrameworkCore\AbpTestDbContext.cs:line 14 at lambda_method(Closure , Object[] ) at Castle.Core.Internal.ReflectionUtil.Instantiate[TBase](Type subtypeofTBase, Object[] ctorArgs) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstanceCore(ConstructorCandidate constructor, Object[] arguments, Type implType) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstance(CreationContext context, ConstructorCandidate constructor, Object[] arguments) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context) at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden) at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired) at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy) at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.Resolve(Type service, IDictionary arguments, IReleasePolicy policy) at Castle.Windsor.WindsorContainer.Resolve[T]() at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Nullable`1 tenantId, Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 84 at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 62 at AbpTest.Tests.AbpTestTestBase..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 33 at AbpTest.Tests.Sessions.SessionAppService_Tests..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\Sessions\SessionAppService_Tests.cs:line 12 Result Message: Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method. ````
1.0
Does UnitTest Project Support EF Core Lazy Loading? - I want to let Unit Test project support lazy loading feature which introduced in EF Core 2.1. Ref : https://docs.microsoft.com/en-us/ef/core/querying/related-data#lazy-loading So, I Performed following steps try to reach the goal: **Step1.** Install-Package Microsoft.EntityFrameworkCore.Proxies on Unit Test Project. **Step2.** Add services.AddEntityFrameworkProxies(); statement in function Register(IIocManager iocManager) of ServiceCollectionRegistrar.cs **Step3.** In same file, add use lazy loading proxy like this : builder.UseLazyLoadingProxies().UseInMemoryDatabase(Guid.NewGuid().ToString()).UseInternalServiceProvider(serviceProvider); ABP package version: 4.0.1 Base on .Net Core When I ran the test case, I got this exception message: Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method. Any idea or suggestion about this problem? Thanks in advance! PS: ```` Result StackTrace: at Castle.MicroKernel.ComponentActivator.FactoryMethodActivator`1.Instantiate(CreationContext context) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context) at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated) at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21 at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden) at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired) at Castle.MicroKernel.Resolvers.DefaultDependencyResolver.Resolve(CreationContext context, ISubDependencyResolver contextHandlerResolver, ComponentModel model, DependencyModel dependency) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateConstructorArguments(ConstructorCandidate constructor, CreationContext context) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.Instantiate(CreationContext context) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context) at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.<>c__DisplayClass4_0.<Resolve>b__0(Action`1 afterCreated) at Castle.MicroKernel.Lifestyle.Scoped.DefaultLifetimeScope.GetCachedInstance(ComponentModel model, ScopedInstanceActivationCallback createInstance) at Castle.MicroKernel.Lifestyle.ScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) at Castle.Windsor.MsDependencyInjection.MsScopedLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\MsScopedLifestyleManager.cs:line 21 at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden) at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired) at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy) at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.ResolveAll(Type service, IDictionary arguments, IReleasePolicy policy) at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.ResolveInstanceOrNull(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 86 at Castle.Windsor.MsDependencyInjection.ScopedWindsorServiceProvider.GetServiceInternal(Type serviceType, Boolean isOptional) in D:\Github\castle-windsor-ms-adapter\src\Castle.Windsor.MsDependencyInjection\ScopedWindsorServiceProvider.cs:line 55 at Microsoft.Extensions.DependencyInjection.ServiceProviderServiceExtensions.GetService[T](IServiceProvider provider) at Microsoft.EntityFrameworkCore.Proxies.Internal.ProxiesOptionsExtension.Validate(IDbContextOptions options) at Microsoft.EntityFrameworkCore.Internal.ServiceProviderCache.GetOrAdd(IDbContextOptions options, Boolean providerRequired) at Microsoft.EntityFrameworkCore.DbContext..ctor(DbContextOptions options) at Abp.Zero.EntityFrameworkCore.AbpZeroDbContext`4..ctor(DbContextOptions`1 options) in D:\Github\aspnetboilerplate\src\Abp.ZeroCore.EntityFrameworkCore\Zero\EntityFrameworkCore\AbpZeroDbContext.cs:line 68 at AbpTest.EntityFrameworkCore.AbpTestDbContext..ctor(DbContextOptions`1 options) in C:\Workspace\Lab\4.0.1\aspnet-core\src\AbpTest.EntityFrameworkCore\EntityFrameworkCore\AbpTestDbContext.cs:line 14 at lambda_method(Closure , Object[] ) at Castle.Core.Internal.ReflectionUtil.Instantiate[TBase](Type subtypeofTBase, Object[] ctorArgs) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstanceCore(ConstructorCandidate constructor, Object[] arguments, Type implType) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.CreateInstance(CreationContext context, ConstructorCandidate constructor, Object[] arguments) at Castle.MicroKernel.ComponentActivator.DefaultComponentActivator.InternalCreate(CreationContext context) at Castle.MicroKernel.ComponentActivator.AbstractComponentActivator.Create(CreationContext context, Burden burden) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.CreateInstance(CreationContext context, Boolean trackedExternally) at Castle.MicroKernel.Lifestyle.AbstractLifestyleManager.Resolve(CreationContext context, IReleasePolicy releasePolicy) at Castle.MicroKernel.Handlers.DefaultHandler.ResolveCore(CreationContext context, Boolean requiresDecommission, Boolean instanceRequired, Burden& burden) at Castle.MicroKernel.Handlers.DefaultHandler.Resolve(CreationContext context, Boolean instanceRequired) at Castle.MicroKernel.DefaultKernel.ResolveComponent(IHandler handler, Type service, IDictionary additionalArguments, IReleasePolicy policy) at Castle.MicroKernel.DefaultKernel.Castle.MicroKernel.IKernelInternal.Resolve(Type service, IDictionary arguments, IReleasePolicy policy) at Castle.Windsor.WindsorContainer.Resolve[T]() at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Nullable`1 tenantId, Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 84 at AbpTest.Tests.AbpTestTestBase.UsingDbContext(Action`1 action) in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 62 at AbpTest.Tests.AbpTestTestBase..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\AbpTestTestBase.cs:line 33 at AbpTest.Tests.Sessions.SessionAppService_Tests..ctor() in C:\Workspace\Lab\4.0.1\aspnet-core\test\AbpTest.Tests\Sessions\SessionAppService_Tests.cs:line 12 Result Message: Castle.MicroKernel.ComponentActivator.ComponentActivatorException : Factory method creating instances of component 'Microsoft.EntityFrameworkCore.Infrastructure.IDbContextOptions_d4040eea-3d91-4e91-aee4-454219d15a92' returned null. This is not allowed and most likely a bug in the factory method. ````
non_main
does unittest project support ef core lazy loading i want to let unit test project support lazy loading feature which introduced in ef core ref so i performed following steps try to reach the goal install package microsoft entityframeworkcore proxies on unit test project add services addentityframeworkproxies statement in function register iiocmanager iocmanager of servicecollectionregistrar cs in same file add use lazy loading proxy like this builder uselazyloadingproxies useinmemorydatabase guid newguid tostring useinternalserviceprovider serviceprovider abp package version base on net core when i ran the test case i got this exception message castle microkernel componentactivator componentactivatorexception factory method creating instances of component microsoft entityframeworkcore infrastructure idbcontextoptions returned null this is not allowed and most likely a bug in the factory method any idea or suggestion about this problem? thanks in advance! ps result stacktrace at castle microkernel componentactivator factorymethodactivator instantiate creationcontext context at castle microkernel componentactivator defaultcomponentactivator internalcreate creationcontext context at castle microkernel componentactivator abstractcomponentactivator create creationcontext context burden burden at castle microkernel lifestyle abstractlifestylemanager createinstance creationcontext context boolean trackedexternally at castle microkernel lifestyle scopedlifestylemanager c b action aftercreated at castle microkernel lifestyle scoped defaultlifetimescope getcachedinstance componentmodel model scopedinstanceactivationcallback createinstance at castle microkernel lifestyle scopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy at castle windsor msdependencyinjection msscopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy in d github castle windsor ms adapter src castle windsor msdependencyinjection msscopedlifestylemanager cs line at castle microkernel handlers defaulthandler resolvecore creationcontext context boolean requiresdecommission boolean instancerequired burden burden at castle microkernel handlers defaulthandler resolve creationcontext context boolean instancerequired at castle microkernel resolvers defaultdependencyresolver resolve creationcontext context isubdependencyresolver contexthandlerresolver componentmodel model dependencymodel dependency at castle microkernel componentactivator defaultcomponentactivator createconstructorarguments constructorcandidate constructor creationcontext context at castle microkernel componentactivator defaultcomponentactivator instantiate creationcontext context at castle microkernel componentactivator defaultcomponentactivator internalcreate creationcontext context at castle microkernel componentactivator abstractcomponentactivator create creationcontext context burden burden at castle microkernel lifestyle abstractlifestylemanager createinstance creationcontext context boolean trackedexternally at castle microkernel lifestyle scopedlifestylemanager c b action aftercreated at castle microkernel lifestyle scoped defaultlifetimescope getcachedinstance componentmodel model scopedinstanceactivationcallback createinstance at castle microkernel lifestyle scopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy at castle windsor msdependencyinjection msscopedlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy in d github castle windsor ms adapter src castle windsor msdependencyinjection msscopedlifestylemanager cs line at castle microkernel handlers defaulthandler resolvecore creationcontext context boolean requiresdecommission boolean instancerequired burden burden at castle microkernel handlers defaulthandler resolve creationcontext context boolean instancerequired at castle microkernel defaultkernel resolvecomponent ihandler handler type service idictionary additionalarguments ireleasepolicy policy at castle microkernel defaultkernel castle microkernel ikernelinternal resolveall type service idictionary arguments ireleasepolicy policy at castle windsor msdependencyinjection scopedwindsorserviceprovider resolveinstanceornull type servicetype boolean isoptional in d github castle windsor ms adapter src castle windsor msdependencyinjection scopedwindsorserviceprovider cs line at castle windsor msdependencyinjection scopedwindsorserviceprovider getserviceinternal type servicetype boolean isoptional in d github castle windsor ms adapter src castle windsor msdependencyinjection scopedwindsorserviceprovider cs line at microsoft extensions dependencyinjection serviceproviderserviceextensions getservice iserviceprovider provider at microsoft entityframeworkcore proxies internal proxiesoptionsextension validate idbcontextoptions options at microsoft entityframeworkcore internal serviceprovidercache getoradd idbcontextoptions options boolean providerrequired at microsoft entityframeworkcore dbcontext ctor dbcontextoptions options at abp zero entityframeworkcore abpzerodbcontext ctor dbcontextoptions options in d github aspnetboilerplate src abp zerocore entityframeworkcore zero entityframeworkcore abpzerodbcontext cs line at abptest entityframeworkcore abptestdbcontext ctor dbcontextoptions options in c workspace lab aspnet core src abptest entityframeworkcore entityframeworkcore abptestdbcontext cs line at lambda method closure object at castle core internal reflectionutil instantiate type subtypeoftbase object ctorargs at castle microkernel componentactivator defaultcomponentactivator createinstancecore constructorcandidate constructor object arguments type impltype at castle microkernel componentactivator defaultcomponentactivator createinstance creationcontext context constructorcandidate constructor object arguments at castle microkernel componentactivator defaultcomponentactivator internalcreate creationcontext context at castle microkernel componentactivator abstractcomponentactivator create creationcontext context burden burden at castle microkernel lifestyle abstractlifestylemanager createinstance creationcontext context boolean trackedexternally at castle microkernel lifestyle abstractlifestylemanager resolve creationcontext context ireleasepolicy releasepolicy at castle microkernel handlers defaulthandler resolvecore creationcontext context boolean requiresdecommission boolean instancerequired burden burden at castle microkernel handlers defaulthandler resolve creationcontext context boolean instancerequired at castle microkernel defaultkernel resolvecomponent ihandler handler type service idictionary additionalarguments ireleasepolicy policy at castle microkernel defaultkernel castle microkernel ikernelinternal resolve type service idictionary arguments ireleasepolicy policy at castle windsor windsorcontainer resolve at abptest tests abptesttestbase usingdbcontext nullable tenantid action action in c workspace lab aspnet core test abptest tests abptesttestbase cs line at abptest tests abptesttestbase usingdbcontext action action in c workspace lab aspnet core test abptest tests abptesttestbase cs line at abptest tests abptesttestbase ctor in c workspace lab aspnet core test abptest tests abptesttestbase cs line at abptest tests sessions sessionappservice tests ctor in c workspace lab aspnet core test abptest tests sessions sessionappservice tests cs line result message castle microkernel componentactivator componentactivatorexception factory method creating instances of component microsoft entityframeworkcore infrastructure idbcontextoptions returned null this is not allowed and most likely a bug in the factory method
0
773,550
27,161,536,993
IssuesEvent
2023-02-17 12:17:18
JamieMason/syncpack
https://api.github.com/repos/JamieMason/syncpack
closed
Output semver/version groups in the order they're defined
Priority: Medium Type: Feat good first issue
## Description Every command that lists output [has a `.reverse()` call](https://github.com/JamieMason/syncpack/blob/d397ea7d986815e3b5a79bcb017db83670839661/src/bin-list/list.ts#L8) – this is a hangover from a previous implementation before I recently refactored the code – looping in reverse just made it easier because it meant the first item could be known to be the default group. This makes it harder to debug because the groups don't align with the order they're defined in config. ## Suggested Solution Remove the `.reverse()` calls from each command and update the expected numbers in the tests. ## Help Needed This _should_ be an easy one, find and delete the calls, run `yarn jest --watch --no-coverage`, and fix the expected numbers in the log output in the failing tests.
1.0
Output semver/version groups in the order they're defined - ## Description Every command that lists output [has a `.reverse()` call](https://github.com/JamieMason/syncpack/blob/d397ea7d986815e3b5a79bcb017db83670839661/src/bin-list/list.ts#L8) – this is a hangover from a previous implementation before I recently refactored the code – looping in reverse just made it easier because it meant the first item could be known to be the default group. This makes it harder to debug because the groups don't align with the order they're defined in config. ## Suggested Solution Remove the `.reverse()` calls from each command and update the expected numbers in the tests. ## Help Needed This _should_ be an easy one, find and delete the calls, run `yarn jest --watch --no-coverage`, and fix the expected numbers in the log output in the failing tests.
non_main
output semver version groups in the order they re defined description every command that lists output – this is a hangover from a previous implementation before i recently refactored the code – looping in reverse just made it easier because it meant the first item could be known to be the default group this makes it harder to debug because the groups don t align with the order they re defined in config suggested solution remove the reverse calls from each command and update the expected numbers in the tests help needed this should be an easy one find and delete the calls run yarn jest watch no coverage and fix the expected numbers in the log output in the failing tests
0
4,174
20,015,665,939
IssuesEvent
2022-02-01 11:48:11
MetaCell/cloud-harness
https://api.github.com/repos/MetaCell/cloud-harness
closed
Add pull request template
maintainance
The pull request template shall include some self sanity checks to help the reviewer and the maintainer contextualize and evaluate the pull request
True
Add pull request template - The pull request template shall include some self sanity checks to help the reviewer and the maintainer contextualize and evaluate the pull request
main
add pull request template the pull request template shall include some self sanity checks to help the reviewer and the maintainer contextualize and evaluate the pull request
1
116,590
9,856,618,180
IssuesEvent
2019-06-19 22:50:53
livinglab/webwork-for-wordpress
https://api.github.com/repos/livinglab/webwork-for-wordpress
closed
Additional style adjustments
enhancement testing-needed
@boonebgorges, I've been able to identify a few additional style adjustments. I _think_ they are small and won't affect other elements but should tighten things up and get us a little closer to the original mockups. Feel free to push to "future" if there are cascade issues I'm not aware of. .item-stats.problem-stats { font-size: 1rem; } h3.ww-header { font-size: 1.4rem; } .ww-ask-question-form label { font-size: 1.4rem; } .ww-question-gloss { font-size: 1.1rem; } .anonymous-toggle label { line-height: 1.2rem; font-size: .85rem; } .index-intro > p { line-height: 2rem; font-size: 1.4rem; } .ww-author-name { font-size: 1.5rem; }
1.0
Additional style adjustments - @boonebgorges, I've been able to identify a few additional style adjustments. I _think_ they are small and won't affect other elements but should tighten things up and get us a little closer to the original mockups. Feel free to push to "future" if there are cascade issues I'm not aware of. .item-stats.problem-stats { font-size: 1rem; } h3.ww-header { font-size: 1.4rem; } .ww-ask-question-form label { font-size: 1.4rem; } .ww-question-gloss { font-size: 1.1rem; } .anonymous-toggle label { line-height: 1.2rem; font-size: .85rem; } .index-intro > p { line-height: 2rem; font-size: 1.4rem; } .ww-author-name { font-size: 1.5rem; }
non_main
additional style adjustments boonebgorges i ve been able to identify a few additional style adjustments i think they are small and won t affect other elements but should tighten things up and get us a little closer to the original mockups feel free to push to future if there are cascade issues i m not aware of item stats problem stats font size ww header font size ww ask question form label font size ww question gloss font size anonymous toggle label line height font size index intro p line height font size ww author name font size
0
118,458
4,745,601,847
IssuesEvent
2016-10-21 08:02:42
kubernetes/dashboard
https://api.github.com/repos/kubernetes/dashboard
closed
Pod state not correct
area/api kind/bug priority/P3
#### Issue details Pod state is not correct in UI ##### Environment ``` Dashboard version: master (12. Aug) Kubernetes version: v1.4.0-alpha.2 ``` ##### Steps to reproduce ``` $ kubectl run test --image=debian -- echo hello $ kubectl get pods NAME READY STATUS RESTARTS AGE test-2172490464-elqbi 0/1 CrashLoopBackOff 18 1h ``` ##### Observed result UI claims pod is in "Running" state ##### Expected result "Crashed"
1.0
Pod state not correct - #### Issue details Pod state is not correct in UI ##### Environment ``` Dashboard version: master (12. Aug) Kubernetes version: v1.4.0-alpha.2 ``` ##### Steps to reproduce ``` $ kubectl run test --image=debian -- echo hello $ kubectl get pods NAME READY STATUS RESTARTS AGE test-2172490464-elqbi 0/1 CrashLoopBackOff 18 1h ``` ##### Observed result UI claims pod is in "Running" state ##### Expected result "Crashed"
non_main
pod state not correct issue details pod state is not correct in ui environment dashboard version master aug kubernetes version alpha steps to reproduce kubectl run test image debian echo hello kubectl get pods name ready status restarts age test elqbi crashloopbackoff observed result ui claims pod is in running state expected result crashed
0
54,069
3,059,291,851
IssuesEvent
2015-08-14 14:16:03
phetsims/tasks
https://api.github.com/repos/phetsims/tasks
closed
Test Bending Light Performance on iPad2 for a few scenarios
High Priority QA
Please test the performance for Bending Light in the scenarios given in: https://github.com/phetsims/bending-light/issues/153 There is a link there for testing and full details in the issue. Please report your results in https://github.com/phetsims/bending-light/issues/153 @ariel-phet please delegate.
1.0
Test Bending Light Performance on iPad2 for a few scenarios - Please test the performance for Bending Light in the scenarios given in: https://github.com/phetsims/bending-light/issues/153 There is a link there for testing and full details in the issue. Please report your results in https://github.com/phetsims/bending-light/issues/153 @ariel-phet please delegate.
non_main
test bending light performance on for a few scenarios please test the performance for bending light in the scenarios given in there is a link there for testing and full details in the issue please report your results in ariel phet please delegate
0
1,174
5,096,305,393
IssuesEvent
2017-01-03 17:46:22
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
macports module: support for state=latest (upgrade) and for variants
affects_2.0 bug_report feature_idea waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME macports ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Orchestrator: Ubuntu trusty Target: Macos 10.11 ##### SUMMARY from http://docs.ansible.com/ansible/macports_module.html no state=latest is supported while commands `port upgrade {{ port }}` and `port outdated | grep {{ port }}` allow to upgrade and check if upgrade available. A `port sync` might be needed as equivalent of update_cache Macports also allow to have different packages/variant for some compile/install options. For example ``` port install lftp +ssl ``` a way to support it is needed. can have multiple +variant1 +variant2 ##### STEPS TO REPRODUCE ``` - name: Darwin | macports upgrade macports: name=* present=latest update_cache=yes when: ansible_os_family == "Darwin" - name: Darwin | variant install macports: name=lftp variants=ssl,variant2 when: ansible_os_family == "Darwin" ``` ##### EXPECTED RESULTS Macports is upgraded Macports has port installed with corresponding variants. ##### ACTUAL RESULTS Above actions are currently not possible within module and must be executed as command losing idempotency.
True
macports module: support for state=latest (upgrade) and for variants - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME macports ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Orchestrator: Ubuntu trusty Target: Macos 10.11 ##### SUMMARY from http://docs.ansible.com/ansible/macports_module.html no state=latest is supported while commands `port upgrade {{ port }}` and `port outdated | grep {{ port }}` allow to upgrade and check if upgrade available. A `port sync` might be needed as equivalent of update_cache Macports also allow to have different packages/variant for some compile/install options. For example ``` port install lftp +ssl ``` a way to support it is needed. can have multiple +variant1 +variant2 ##### STEPS TO REPRODUCE ``` - name: Darwin | macports upgrade macports: name=* present=latest update_cache=yes when: ansible_os_family == "Darwin" - name: Darwin | variant install macports: name=lftp variants=ssl,variant2 when: ansible_os_family == "Darwin" ``` ##### EXPECTED RESULTS Macports is upgraded Macports has port installed with corresponding variants. ##### ACTUAL RESULTS Above actions are currently not possible within module and must be executed as command losing idempotency.
main
macports module support for state latest upgrade and for variants issue type bug report component name macports ansible version ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration n a os environment orchestrator ubuntu trusty target macos summary from no state latest is supported while commands port upgrade port and port outdated grep port allow to upgrade and check if upgrade available a port sync might be needed as equivalent of update cache macports also allow to have different packages variant for some compile install options for example port install lftp ssl a way to support it is needed can have multiple steps to reproduce name darwin macports upgrade macports name present latest update cache yes when ansible os family darwin name darwin variant install macports name lftp variants ssl when ansible os family darwin expected results macports is upgraded macports has port installed with corresponding variants actual results above actions are currently not possible within module and must be executed as command losing idempotency
1
2,786
9,985,230,884
IssuesEvent
2019-07-10 16:04:27
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
opened
Move duplicated CHECK defines in tests to client_tools.h
Maintainability good first issue
We have 15 different tests defining a CHECK macro. We should move all of those to a single define in client_tools.h.
True
Move duplicated CHECK defines in tests to client_tools.h - We have 15 different tests defining a CHECK macro. We should move all of those to a single define in client_tools.h.
main
move duplicated check defines in tests to client tools h we have different tests defining a check macro we should move all of those to a single define in client tools h
1
152,868
5,871,404,869
IssuesEvent
2017-05-15 08:37:44
PX4/Firmware
https://api.github.com/repos/PX4/Firmware
closed
Losing GPS does not trigger fail safe
bug priority-critical
On latest master, failsafe when losing GPS does not work. A few test cases: * When I unclick `use GPS` for the EKF2_AID_MASK (in air), the only thing appearing is `WARN [navigator] global position timeout` after a few seconds but nothing happens. The quad just starts to drift. * When I stop sending GPS (in air), nothing happens. The quad just starts to drift. * When the home position was set but has no GPS anymore (on ground), I can still takeoff. Then the quad drifts I guess the EKF reports local position as still valid?
1.0
Losing GPS does not trigger fail safe - On latest master, failsafe when losing GPS does not work. A few test cases: * When I unclick `use GPS` for the EKF2_AID_MASK (in air), the only thing appearing is `WARN [navigator] global position timeout` after a few seconds but nothing happens. The quad just starts to drift. * When I stop sending GPS (in air), nothing happens. The quad just starts to drift. * When the home position was set but has no GPS anymore (on ground), I can still takeoff. Then the quad drifts I guess the EKF reports local position as still valid?
non_main
losing gps does not trigger fail safe on latest master failsafe when losing gps does not work a few test cases when i unclick use gps for the aid mask in air the only thing appearing is warn global position timeout after a few seconds but nothing happens the quad just starts to drift when i stop sending gps in air nothing happens the quad just starts to drift when the home position was set but has no gps anymore on ground i can still takeoff then the quad drifts i guess the ekf reports local position as still valid
0
4,737
24,456,719,413
IssuesEvent
2022-10-07 07:29:27
NaluKit/nalu
https://api.github.com/repos/NaluKit/nalu
closed
Update plugin & dependency versions
maintainance
There are some newer versions of plugins and dependencies. Especially to avoid the JUnit 5 deprecate warnings
True
Update plugin & dependency versions - There are some newer versions of plugins and dependencies. Especially to avoid the JUnit 5 deprecate warnings
main
update plugin dependency versions there are some newer versions of plugins and dependencies especially to avoid the junit deprecate warnings
1
435,154
30,488,378,132
IssuesEvent
2023-07-18 05:27:55
hwchase17/langchain
https://api.github.com/repos/hwchase17/langchain
closed
DOC: SupabaseVectorStore.from_documents read operation timed out.
area: vector store auto:question auto:documentation
### Issue with current documentation: ``` # We're using the default `documents` table here. You can modify this by passing in a `table_name` argument to the `from_documents` method. vector_store = SupabaseVectorStore.from_documents(docs, embeddings, client=supabase) ``` ### Idea or request for content: throw error: httpx.ReadTimeout: The read operation timed out Is it because the documents are too large? Is there a way to change the timeout?
1.0
DOC: SupabaseVectorStore.from_documents read operation timed out. - ### Issue with current documentation: ``` # We're using the default `documents` table here. You can modify this by passing in a `table_name` argument to the `from_documents` method. vector_store = SupabaseVectorStore.from_documents(docs, embeddings, client=supabase) ``` ### Idea or request for content: throw error: httpx.ReadTimeout: The read operation timed out Is it because the documents are too large? Is there a way to change the timeout?
non_main
doc supabasevectorstore from documents read operation timed out issue with current documentation we re using the default documents table here you can modify this by passing in a table name argument to the from documents method vector store supabasevectorstore from documents docs embeddings client supabase idea or request for content throw error httpx readtimeout the read operation timed out is it because the documents are too large is there a way to change the timeout
0
637
4,155,107,931
IssuesEvent
2016-06-16 14:00:10
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Notepad_plus_plus Cheat Sheet:
Maintainer Submitted
In producing this CS, I made the mistake of incorrectly specifying single key combinations with square brackets, e.g. "[F7]" instead of "F7" ------ IA Page: http://duck.co/ia/view/notepad_plus_plus_cheat_sheet [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @biscuitNinja
True
Notepad_plus_plus Cheat Sheet: - In producing this CS, I made the mistake of incorrectly specifying single key combinations with square brackets, e.g. "[F7]" instead of "F7" ------ IA Page: http://duck.co/ia/view/notepad_plus_plus_cheat_sheet [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @biscuitNinja
main
notepad plus plus cheat sheet in producing this cs i made the mistake of incorrectly specifying single key combinations with square brackets e g instead of ia page biscuitninja
1
1,982
6,694,201,269
IssuesEvent
2017-10-10 00:16:50
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Astrobin Apod: not triggering
Maintainer Input Requested
the example query "astronomy picture of the day" isn't triggering for me. --- IA Page: http://duck.co/ia/view/apod [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @killerfish
True
Astrobin Apod: not triggering - the example query "astronomy picture of the day" isn't triggering for me. --- IA Page: http://duck.co/ia/view/apod [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @killerfish
main
astrobin apod not triggering the example query astronomy picture of the day isn t triggering for me ia page killerfish
1
64,510
15,896,497,500
IssuesEvent
2021-04-11 17:38:32
haskell/text
https://api.github.com/repos/haskell/text
closed
Can't build benchmarks
build failure
When trying to build benchmarks, I get a linker error, as the `cbits.c` from the `text` i have installed clash with the `cbits.c` from the `text` source code i'm currently working on and which i'm trying to benchmark. ``` Linking dist/build/text-benchmarks/text-benchmarks ... /home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_memcmp': (.text+0x20): multiple definition of `_hs_text_memcmp' dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x0): first defined here /home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8_state': (.text+0xe0): multiple definition of `_hs_text_decode_utf8_state' dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x20): first defined here /home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8': (.text+0x2a0): multiple definition of `_hs_text_decode_utf8' dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x150): first defined here collect2: error: ld returned 1 exit status ``` The problem is, that `criterion` has `text` as a dependency. Currently, as a workaround, I just rename the functions in `cbits.c`, then `cabal build` just works. * Is there a better solution? * If not, I can send a PR adding a compilation flag and do the renaming, so `cabal build` just works. * Is this related to https://github.com/haskell/cabal/issues/1575 ?
1.0
Can't build benchmarks - When trying to build benchmarks, I get a linker error, as the `cbits.c` from the `text` i have installed clash with the `cbits.c` from the `text` source code i'm currently working on and which i'm trying to benchmark. ``` Linking dist/build/text-benchmarks/text-benchmarks ... /home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_memcmp': (.text+0x20): multiple definition of `_hs_text_memcmp' dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x0): first defined here /home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8_state': (.text+0xe0): multiple definition of `_hs_text_decode_utf8_state' dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x20): first defined here /home/kuko/.cabal/lib/x86_64-linux-ghc-8.0.2/text-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh/libHStext-1.2.2.2-KC7dWoG09dA1F6jKj5GSqh.a(cbits.o): In function `_hs_text_decode_utf8': (.text+0x2a0): multiple definition of `_hs_text_decode_utf8' dist/build/text-benchmarks/text-benchmarks-tmp/../cbits/cbits.o:cbits.c:(.text+0x150): first defined here collect2: error: ld returned 1 exit status ``` The problem is, that `criterion` has `text` as a dependency. Currently, as a workaround, I just rename the functions in `cbits.c`, then `cabal build` just works. * Is there a better solution? * If not, I can send a PR adding a compilation flag and do the renaming, so `cabal build` just works. * Is this related to https://github.com/haskell/cabal/issues/1575 ?
non_main
can t build benchmarks when trying to build benchmarks i get a linker error as the cbits c from the text i have installed clash with the cbits c from the text source code i m currently working on and which i m trying to benchmark linking dist build text benchmarks text benchmarks home kuko cabal lib linux ghc text libhstext a cbits o in function hs text memcmp text multiple definition of hs text memcmp dist build text benchmarks text benchmarks tmp cbits cbits o cbits c text first defined here home kuko cabal lib linux ghc text libhstext a cbits o in function hs text decode state text multiple definition of hs text decode state dist build text benchmarks text benchmarks tmp cbits cbits o cbits c text first defined here home kuko cabal lib linux ghc text libhstext a cbits o in function hs text decode text multiple definition of hs text decode dist build text benchmarks text benchmarks tmp cbits cbits o cbits c text first defined here error ld returned exit status the problem is that criterion has text as a dependency currently as a workaround i just rename the functions in cbits c then cabal build just works is there a better solution if not i can send a pr adding a compilation flag and do the renaming so cabal build just works is this related to
0
94,781
10,854,066,752
IssuesEvent
2019-11-13 15:47:51
dotnet/reactive
https://api.github.com/repos/dotnet/reactive
closed
Make the Documentation available on the Microsoft Docs website
documentation
Rx is an extremely powerful (and in many cases, essential) framework to build complex event-driven / reactive applications or to deal with complex asynchronous workflows. As such an important library, I think it should definitely be available on the [Microsoft Docs](https://docs.microsoft.com) website. There is a documentation available on [MSDN](https://msdn.microsoft.com/en-us/library/hh242985(v=vs.103).aspx) with API Reference, but I think might be outdated. It could have its place somewhere alongside LINQ in the [.NET Guide](https://docs.microsoft.com/en-us/dotnet/standard/index) section, or under the [Asynchronous Programming Patterns](https://docs.microsoft.com/en-us/dotnet/standard/asynchronous-programming-patterns/) section, maybe?
1.0
Make the Documentation available on the Microsoft Docs website - Rx is an extremely powerful (and in many cases, essential) framework to build complex event-driven / reactive applications or to deal with complex asynchronous workflows. As such an important library, I think it should definitely be available on the [Microsoft Docs](https://docs.microsoft.com) website. There is a documentation available on [MSDN](https://msdn.microsoft.com/en-us/library/hh242985(v=vs.103).aspx) with API Reference, but I think might be outdated. It could have its place somewhere alongside LINQ in the [.NET Guide](https://docs.microsoft.com/en-us/dotnet/standard/index) section, or under the [Asynchronous Programming Patterns](https://docs.microsoft.com/en-us/dotnet/standard/asynchronous-programming-patterns/) section, maybe?
non_main
make the documentation available on the microsoft docs website rx is an extremely powerful and in many cases essential framework to build complex event driven reactive applications or to deal with complex asynchronous workflows as such an important library i think it should definitely be available on the website there is a documentation available on with api reference but i think might be outdated it could have its place somewhere alongside linq in the section or under the section maybe
0
3,157
12,198,047,592
IssuesEvent
2020-04-29 21:59:12
short-d/short
https://api.github.com/repos/short-d/short
closed
[Refactor] Move envconfig into app framework
maintainability
Current [envconfig](https://github.com/short-d/short/tree/master/backend/envconfig) in located in Short's repo. However, it's reusable across all microservices. Define `EnvConfig` interface in [fw](https://github.com/short-d/app/tree/master/fw) and create implementation here: https://github.com/short-d/app/tree/master/modern
True
[Refactor] Move envconfig into app framework - Current [envconfig](https://github.com/short-d/short/tree/master/backend/envconfig) in located in Short's repo. However, it's reusable across all microservices. Define `EnvConfig` interface in [fw](https://github.com/short-d/app/tree/master/fw) and create implementation here: https://github.com/short-d/app/tree/master/modern
main
move envconfig into app framework current in located in short s repo however it s reusable across all microservices define envconfig interface in and create implementation here
1
3,673
15,036,008,410
IssuesEvent
2021-02-02 14:47:15
IITIDIDX597/sp_2021_team1
https://api.github.com/repos/IITIDIDX597/sp_2021_team1
opened
Sorting information in categories
Epic: 5 Maintaining the system Story Week 3
**Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care. **Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform. **Sub-Hill Statements:** 1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments. ### **Story Details:** As an: administrator I want: to be able to add information in their respective categories So that: I can keep the system updated
True
Sorting information in categories - **Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care. **Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform. **Sub-Hill Statements:** 1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments. ### **Story Details:** As an: administrator I want: to be able to add information in their respective categories So that: I can keep the system updated
main
sorting information in categories project goal s lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way while at the same time foster deeper learning experiences in order to deliver better abilitylab patient care hill statement individual clinicians can reference relevant continuously evolving information for their patient s therapy needs to self manage their approach patient care plan development in a single platform sub hill statements the learning platform will be routinely updated with s lab s own research advancements as well as outside discoveries and best practices developed for rehabilitation treatments story details as an administrator i want to be able to add information in their respective categories so that i can keep the system updated
1
2,816
10,102,930,772
IssuesEvent
2019-07-29 12:29:32
luckyariane/arthas-bot
https://api.github.com/repos/luckyariane/arthas-bot
closed
Fix recent subscriber parsing.
bug maintain
The loss of twitch alert local file creation means I need to find a new source for this data.
True
Fix recent subscriber parsing. - The loss of twitch alert local file creation means I need to find a new source for this data.
main
fix recent subscriber parsing the loss of twitch alert local file creation means i need to find a new source for this data
1
4,607
23,855,859,644
IssuesEvent
2022-09-06 23:24:56
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[a11y]: Violations found on UI Shell w/ SideNav component in v10 and v11
severity: 2 type: a11y β™Ώ component: ui-shell status: waiting for maintainer response πŸ’¬
### Package @carbon/react ### Browser Chrome, Firefox ### Operating System MacOS ### Package version v10, v11 ### React version _No response_ ### Automated testing tool and ruleset IBM Accessibility Checker - Latest Deployment ### Assistive technology _No response_ ### Description I used IBM Accessibility Checker to scan a newly developed page based on React which uses the UI Shell w/ SideNav component. When I scan this component independently (the page on storybook), violations are reported. ### WCAG 2.1 Violation 1.3.1 Info and Relationships, 2.4.1 Bypass Blocks, 4.1.2 Name, Role, Value (v10) and 1.3.1 Info and Relationships, 4.1.2 Name, Role, Value (v11). ### Reproduction/example Violations are found on Carbon-React storybook for v10 and v11 ### Steps to reproduce To reproduce violations on v10: 1. Go to https://v10-react.carbondesignsystem.com/iframe.html?id=components-ui-shell--header-base-w-side-nav&viewMode=story&args= 2. Scan this page using IBM Accessibility Checkers 3. 8 violations of 4.1.2 Name, Role, Value will be found: ![Screenshot 2022-08-22 at 14 52 58](https://user-images.githubusercontent.com/81176980/185926786-b433cc4a-f2fc-47ab-8370-3fb18eafe35c.png) To reproduce violations on v11: 1. Go to https://react.carbondesignsystem.com/iframe.html?viewMode=story&id=components-ui-shell--header-base-w-side-nav&args= 2. Scan this page using IBM Accessibility Checkers 3. 5 violation ![Screenshot 2022-08-22 at 14 53 02](https://user-images.githubusercontent.com/81176980/185926815-bd056c57-0c40-4456-88e7-6bedad02da3e.png) s of 2.4.1 Bypass Blocks will be found ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
True
[a11y]: Violations found on UI Shell w/ SideNav component in v10 and v11 - ### Package @carbon/react ### Browser Chrome, Firefox ### Operating System MacOS ### Package version v10, v11 ### React version _No response_ ### Automated testing tool and ruleset IBM Accessibility Checker - Latest Deployment ### Assistive technology _No response_ ### Description I used IBM Accessibility Checker to scan a newly developed page based on React which uses the UI Shell w/ SideNav component. When I scan this component independently (the page on storybook), violations are reported. ### WCAG 2.1 Violation 1.3.1 Info and Relationships, 2.4.1 Bypass Blocks, 4.1.2 Name, Role, Value (v10) and 1.3.1 Info and Relationships, 4.1.2 Name, Role, Value (v11). ### Reproduction/example Violations are found on Carbon-React storybook for v10 and v11 ### Steps to reproduce To reproduce violations on v10: 1. Go to https://v10-react.carbondesignsystem.com/iframe.html?id=components-ui-shell--header-base-w-side-nav&viewMode=story&args= 2. Scan this page using IBM Accessibility Checkers 3. 8 violations of 4.1.2 Name, Role, Value will be found: ![Screenshot 2022-08-22 at 14 52 58](https://user-images.githubusercontent.com/81176980/185926786-b433cc4a-f2fc-47ab-8370-3fb18eafe35c.png) To reproduce violations on v11: 1. Go to https://react.carbondesignsystem.com/iframe.html?viewMode=story&id=components-ui-shell--header-base-w-side-nav&args= 2. Scan this page using IBM Accessibility Checkers 3. 5 violation ![Screenshot 2022-08-22 at 14 53 02](https://user-images.githubusercontent.com/81176980/185926815-bd056c57-0c40-4456-88e7-6bedad02da3e.png) s of 2.4.1 Bypass Blocks will be found ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
main
violations found on ui shell w sidenav component in and package carbon react browser chrome firefox operating system macos package version react version no response automated testing tool and ruleset ibm accessibility checker latest deployment assistive technology no response description i used ibm accessibility checker to scan a newly developed page based on react which uses the ui shell w sidenav component when i scan this component independently the page on storybook violations are reported wcag violation info and relationships bypass blocks name role value and info and relationships name role value reproduction example violations are found on carbon react storybook for and steps to reproduce to reproduce violations on go to scan this page using ibm accessibility checkers violations of name role value will be found to reproduce violations on go to scan this page using ibm accessibility checkers violation s of bypass blocks will be found code of conduct i agree to follow this project s i checked the for duplicate problems
1
1,124
4,995,646,808
IssuesEvent
2016-12-09 10:53:16
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
vmware_guest is very slow
affects_2.2 bug_report cloud vmware waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME vmware_guest.py ##### ANSIBLE VERSION ``` 2.2.0 ``` ##### SUMMARY I have a lot of VMs and folders in my vcenter and _build_folder_map function take approximately 3 minutes to run. Have you an idea to accelerate it ?
True
vmware_guest is very slow - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME vmware_guest.py ##### ANSIBLE VERSION ``` 2.2.0 ``` ##### SUMMARY I have a lot of VMs and folders in my vcenter and _build_folder_map function take approximately 3 minutes to run. Have you an idea to accelerate it ?
main
vmware guest is very slow issue type bug report component name vmware guest py ansible version summary i have a lot of vms and folders in my vcenter and build folder map function take approximately minutes to run have you an idea to accelerate it
1
1,392
6,025,276,083
IssuesEvent
2017-06-08 08:16:04
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_iis_webbinding fails to assign specified certificate_hash
affects_2.1 bug_report waiting_on_maintainer windows
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME win_iis_webbinding ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.1.0 (detached HEAD 35da6ba9d1) last updated 2016/07/25 09:38:23 (GMT +000) lib/ansible/modules/core: (detached HEAD 45128c8bab) last updated 2016/07/25 09:38:23 (GMT +000) lib/ansible/modules/extras: (detached HEAD 511752e53a) last updated 2016/07/25 09:38:23 (GMT +000) config file = /tmp/***/ansible/ansible.cfg configured module search path = ['modules'] ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> [defaults] roles_path = roles library = modules ##### OS / ENVIRONMENT Server 2012 R2 ##### SUMMARY win_iis_webbinding module is not assigning the certificate specified in the certificate_hash variable. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> - Install certificate on to IIS server (we've tested this with a PS script and installing manually, it's installed properly). - Paste the certificate hash in to the win_iis_webbinding module - After the play, the IIS website should have a HTTPS web binding attached with the specified certificate - it doesn't. <!--- Paste example playbooks or commands between quotes below --> ``` - name: IIS | Add HTTPS binding with cert for Default Website win_iis_webbinding: name: "Default Web Site" protocol: https port: 443 state: present certificate_hash: β€Ž"***VALID CERT HASH OF INSTALLED CERT*****" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> HTTPS web binding created with specified certificate used. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> A web binding is created successfully for HTTPS, but the certificate is not attached properly. <!--- Paste verbatim command output between quotes below --> ``` changed: [ec2-**-***-***-***.eu-west-1.compute.amazonaws.com] => {"added": [{"bindingInformation": "*:443:", "certificateHash": "", "certificateStoreName": "", "isDsMapperEnabled": false, "protocol": "https", "sslFlags": 0}], "changed": true, "invocation": {"module_name": "win_iis_webbinding"}, "ip": "0.0.0.0", "matched": [], "parameters": {"Name": "Default Web Site", "Port": 443, "Protocol": "https"}, "port": 443, "removed": []} ```
True
win_iis_webbinding fails to assign specified certificate_hash - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME win_iis_webbinding ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.1.0 (detached HEAD 35da6ba9d1) last updated 2016/07/25 09:38:23 (GMT +000) lib/ansible/modules/core: (detached HEAD 45128c8bab) last updated 2016/07/25 09:38:23 (GMT +000) lib/ansible/modules/extras: (detached HEAD 511752e53a) last updated 2016/07/25 09:38:23 (GMT +000) config file = /tmp/***/ansible/ansible.cfg configured module search path = ['modules'] ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> [defaults] roles_path = roles library = modules ##### OS / ENVIRONMENT Server 2012 R2 ##### SUMMARY win_iis_webbinding module is not assigning the certificate specified in the certificate_hash variable. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> - Install certificate on to IIS server (we've tested this with a PS script and installing manually, it's installed properly). - Paste the certificate hash in to the win_iis_webbinding module - After the play, the IIS website should have a HTTPS web binding attached with the specified certificate - it doesn't. <!--- Paste example playbooks or commands between quotes below --> ``` - name: IIS | Add HTTPS binding with cert for Default Website win_iis_webbinding: name: "Default Web Site" protocol: https port: 443 state: present certificate_hash: β€Ž"***VALID CERT HASH OF INSTALLED CERT*****" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> HTTPS web binding created with specified certificate used. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> A web binding is created successfully for HTTPS, but the certificate is not attached properly. <!--- Paste verbatim command output between quotes below --> ``` changed: [ec2-**-***-***-***.eu-west-1.compute.amazonaws.com] => {"added": [{"bindingInformation": "*:443:", "certificateHash": "", "certificateStoreName": "", "isDsMapperEnabled": false, "protocol": "https", "sslFlags": 0}], "changed": true, "invocation": {"module_name": "win_iis_webbinding"}, "ip": "0.0.0.0", "matched": [], "parameters": {"Name": "Default Web Site", "Port": 443, "Protocol": "https"}, "port": 443, "removed": []} ```
main
win iis webbinding fails to assign specified certificate hash issue type bug report component name win iis webbinding ansible version ansible detached head last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file tmp ansible ansible cfg configured module search path configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables roles path roles library modules os environment server summary win iis webbinding module is not assigning the certificate specified in the certificate hash variable steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used install certificate on to iis server we ve tested this with a ps script and installing manually it s installed properly paste the certificate hash in to the win iis webbinding module after the play the iis website should have a https web binding attached with the specified certificate it doesn t name iis add https binding with cert for default website win iis webbinding name default web site protocol https port state present certificate hash β€Ž valid cert hash of installed cert expected results https web binding created with specified certificate used actual results a web binding is created successfully for https but the certificate is not attached properly changed added changed true invocation module name win iis webbinding ip matched parameters name default web site port protocol https port removed
1
26,409
7,835,342,476
IssuesEvent
2018-06-17 03:53:10
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
closed
Build-JDK8-win_x86 i686-w64-mingw32-g++: Command not found
comp:build
https://ci.eclipse.org/openj9/job/Build-JDK8-win_x86/3/console I assume a machine setup issue. win2012r2-x86-1 The build has previously worked on win2012r2-x86-3 ``` 07:29:22 make[3]: i686-w64-mingw32-g++: Command not found 07:29:22 make[3]: i686-w64-mingw32-g++: Command not found 07:29:22 make[3]: *** [../makelib/targets.mk:438: BytecodeInterpreter.obj] Error 127 07:29:22 make[3]: *** Waiting for unfinished jobs.... 07:29:22 make[3]: *** [../makelib/targets.mk:441: DebugBytecodeInterpreter.obj] Error 127 ```
1.0
Build-JDK8-win_x86 i686-w64-mingw32-g++: Command not found - https://ci.eclipse.org/openj9/job/Build-JDK8-win_x86/3/console I assume a machine setup issue. win2012r2-x86-1 The build has previously worked on win2012r2-x86-3 ``` 07:29:22 make[3]: i686-w64-mingw32-g++: Command not found 07:29:22 make[3]: i686-w64-mingw32-g++: Command not found 07:29:22 make[3]: *** [../makelib/targets.mk:438: BytecodeInterpreter.obj] Error 127 07:29:22 make[3]: *** Waiting for unfinished jobs.... 07:29:22 make[3]: *** [../makelib/targets.mk:441: DebugBytecodeInterpreter.obj] Error 127 ```
non_main
build win g command not found i assume a machine setup issue the build has previously worked on make g command not found make g command not found make error make waiting for unfinished jobs make error
0
3,074
11,642,592,238
IssuesEvent
2020-02-29 08:01:06
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
It would be very cool if there is a ftp net_tool.
affects_2.9 deprecated feature infoblox module needs_maintainer needs_triage net_tools networking support:certified support:community support:core
<!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> [get_url](https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/net_tools/basics/get_url.py]) which can download files from HTTP, HTTPS, or FTP to the remote server. How about uploading files? It seems that there is no module to upload files to HTTP, HTTPS, or FTP to the remote server.Sometimes ftp is a good net tool for upload/download files. Can ansible have a module for that ? Thanks for your attention. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> ansible/lib/ansible/modules/net_tools/basics/ftp.py ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> <!--- Paste example playbooks or commands between quotes below --> ```yaml ``` <!--- HINT: You can also paste gist.github.com links for larger files -->
True
It would be very cool if there is a ftp net_tool. - <!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> [get_url](https://github.com/ansible/ansible/blob/devel/lib/ansible/modules/net_tools/basics/get_url.py]) which can download files from HTTP, HTTPS, or FTP to the remote server. How about uploading files? It seems that there is no module to upload files to HTTP, HTTPS, or FTP to the remote server.Sometimes ftp is a good net tool for upload/download files. Can ansible have a module for that ? Thanks for your attention. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> ansible/lib/ansible/modules/net_tools/basics/ftp.py ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> <!--- Paste example playbooks or commands between quotes below --> ```yaml ``` <!--- HINT: You can also paste gist.github.com links for larger files -->
main
it would be very cool if there is a ftp net tool summary which can download files from http https or ftp to the remote server how about uploading files it seems that there is no module to upload files to http https or ftp to the remote server sometimes ftp is a good net tool for upload download files can ansible have a module for that thanks for your attention issue type feature idea component name ansible lib ansible modules net tools basics ftp py additional information yaml
1
1,439
6,254,208,027
IssuesEvent
2017-07-14 01:03:28
Microsoft/DirectXMath
https://api.github.com/repos/Microsoft/DirectXMath
opened
Remove VS 2013 compiler support
maintainence
At some point it will make sense to retire VS 2013 compiler support. This will enable the ability to remove the ``constexpr`` workaround (assuming I also drop support for VS 2015 RTM): #if defined(_MSC_VER) && (_MSC_FULL_VER < 190023506) #define XM_CONST const #define XM_CONSTEXPR #else #define XM_CONST constexpr #define XM_CONSTEXPR constexpr #endif
True
Remove VS 2013 compiler support - At some point it will make sense to retire VS 2013 compiler support. This will enable the ability to remove the ``constexpr`` workaround (assuming I also drop support for VS 2015 RTM): #if defined(_MSC_VER) && (_MSC_FULL_VER < 190023506) #define XM_CONST const #define XM_CONSTEXPR #else #define XM_CONST constexpr #define XM_CONSTEXPR constexpr #endif
main
remove vs compiler support at some point it will make sense to retire vs compiler support this will enable the ability to remove the constexpr workaround assuming i also drop support for vs rtm if defined msc ver msc full ver define xm const const define xm constexpr else define xm const constexpr define xm constexpr constexpr endif
1
4,379
22,287,203,914
IssuesEvent
2022-06-11 20:35:00
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFM - privoxy
Status: Available For Maintainer(s) Embeddable
## I DON'T Want To Become The Maintainer - [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package; - [x] There is no existing open maintainer request for this package; ## Checklist - [x] Issue title starts with 'RFM - ' ## Existing Package Details Package URL: https://chocolatey.org/packages/privoxy Package source URL: https://github.com/dtgm/chocolatey-packages/tree/master/automatic/privoxy/ Date the maintainer was contacted (in YYYY-MM-DD): 2021-12-13 How the maintainer was contacted: Contacted via Email
True
RFM - privoxy - ## I DON'T Want To Become The Maintainer - [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package; - [x] There is no existing open maintainer request for this package; ## Checklist - [x] Issue title starts with 'RFM - ' ## Existing Package Details Package URL: https://chocolatey.org/packages/privoxy Package source URL: https://github.com/dtgm/chocolatey-packages/tree/master/automatic/privoxy/ Date the maintainer was contacted (in YYYY-MM-DD): 2021-12-13 How the maintainer was contacted: Contacted via Email
main
rfm privoxy i don t want to become the maintainer i have followed the package triage process and i do not want to become maintainer of the package there is no existing open maintainer request for this package checklist issue title starts with rfm existing package details package url package source url date the maintainer was contacted in yyyy mm dd how the maintainer was contacted contacted via email
1
2,346
8,392,686,410
IssuesEvent
2018-10-09 18:20:42
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Archive module doesn't work with python 2.7 and LZMA package, asks for backport.
affects_2.7 bug module needs_maintainer python3 support:community traceback
<!--- Verify first that your issue/request is not already reported on GitHub. THIS FORM WILL BE READ BY A MACHINE, COMPLETE ALL SECTIONS AS DESCRIBED. Also test if the latest release, and devel branch are affected too. ALWAYS add information AFTER (OUTSIDE) these html comments. Otherwise it may end up being automatically closed by our bot. --> ##### SUMMARY <!--- Explain the problem briefly --> I've tried using the xz archive extension from ansible and can't seem to make it work. Whenever i try to execute the script ansible errors out with the message: ` lzma or backports.lzma is required when using xz format.` ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Insert, BELOW THIS COMMENT, the name of the module, plugin, task or feature. Do not include extra details here, e.g. "vyos_command" not "the network module vyos_command" or the full path--> archive ##### ANSIBLE VERSION <!--- Paste, BELOW THIS COMMENT, verbatim output from "ansible --version" between quotes below --> ``` ansible 2.7.0.dev0 config file = /root/.ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python2.7/dist-packages/ansible-2.7.0.dev0-py2.7.egg/ansible executable location = /usr/local/bin/ansible python version = 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION <!--- If using Ansible 2.4 or above, paste, BELOW THIS COMMENT, the results of "ansible-config dump --only-changed" Otherwise, mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables).--> ``` DEFAULT_CALLBACK_WHITELIST(/root/.ansible.cfg) = [u'time'] DEFAULT_REMOTE_USER(/root/.ansible.cfg) = root DEFAULT_STDOUT_CALLBACK(/root/.ansible.cfg) = debug HOST_KEY_CHECKING(/root/.ansible.cfg) = False RETRY_FILES_SAVE_PATH(/root/.ansible.cfg) = /root/.ansible/retry ``` ##### OS / ENVIRONMENT <!--- Mention, BELOW THIS COMMENT, the OS you are running Ansible from, and the OS you are managing, or say "N/A" for anything that is not platform-specific. Also mention the specific version of what you are trying to control, e.g. if this is a network bug the version of firmware on the network device.--> Linux 4.4.0-124-generic #148-Ubuntu SMP Wed May 2 13:00:18 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem, using a minimal test-case. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ```yaml --- - name: "-- proof of concept" hosts: localhost - name: "-- archive something" archive: format: xz path: "./something" dest: "./something.xz" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Ansible creates package something.xz ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> ``` MSG: lzma or backports.lzma is required when using xz format. ``` --- ##### Further testing and other debug information <!--- Paste verbatim command output between quotes below --> ``` 13:56$ > python Python 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609] on linux2 Type "help", "copyright", "credits" or "license" for more information. >>> from backports import lzma Traceback (most recent call last): File "<stdin>", line 1, in <module> ImportError: cannot import name lzma >>> import lzma >>> ``` --- ``` temporary-build-server @ root: ~ 13:56$ > python3 Python 3.5.2 (default, Nov 23 2017, 16:37:01) [GCC 5.4.0 20160609] on linux Type "help", "copyright", "credits" or "license" for more information. >>> from backports import lzma Traceback (most recent call last): File "<stdin>", line 1, in <module> ImportError: No module named 'backports' >>> import lzma >>> ``` --- ``` 13:57$ > dpkg -l |grep lzma ii liblzma5:amd64 5.1.1alpha+20120614-2ubuntu2 amd64 XZ-format compression library ii lzma 9.22-2ubuntu2 amd64 Compression and decompression in the LZMA format - command line utility ii python-lzma 0.5.3-3 amd64 Python bindings for liblzma ``` --- ##### Source code modifications __this section seem to be the culprit__ https://github.com/ansible/ansible/blob/9ff20521d1ada2acc64b623875b1d8e51809e0f9/lib/ansible/modules/files/archive.py#L150-L161 Since python is identifying my installation as python 2.7 it defaults to asking for backports, even tho the actual lzma import is there. This might be caused by lzma being installed through `apt` and not `pip` In my humble opinion the code should check for both lzma and backports.lzma. --- ##### Temporary fix Installing `backports.lzma` through pip and pip3 seems to fix the issue on my machine.
True
Archive module doesn't work with python 2.7 and LZMA package, asks for backport. - <!--- Verify first that your issue/request is not already reported on GitHub. THIS FORM WILL BE READ BY A MACHINE, COMPLETE ALL SECTIONS AS DESCRIBED. Also test if the latest release, and devel branch are affected too. ALWAYS add information AFTER (OUTSIDE) these html comments. Otherwise it may end up being automatically closed by our bot. --> ##### SUMMARY <!--- Explain the problem briefly --> I've tried using the xz archive extension from ansible and can't seem to make it work. Whenever i try to execute the script ansible errors out with the message: ` lzma or backports.lzma is required when using xz format.` ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Insert, BELOW THIS COMMENT, the name of the module, plugin, task or feature. Do not include extra details here, e.g. "vyos_command" not "the network module vyos_command" or the full path--> archive ##### ANSIBLE VERSION <!--- Paste, BELOW THIS COMMENT, verbatim output from "ansible --version" between quotes below --> ``` ansible 2.7.0.dev0 config file = /root/.ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python2.7/dist-packages/ansible-2.7.0.dev0-py2.7.egg/ansible executable location = /usr/local/bin/ansible python version = 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION <!--- If using Ansible 2.4 or above, paste, BELOW THIS COMMENT, the results of "ansible-config dump --only-changed" Otherwise, mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables).--> ``` DEFAULT_CALLBACK_WHITELIST(/root/.ansible.cfg) = [u'time'] DEFAULT_REMOTE_USER(/root/.ansible.cfg) = root DEFAULT_STDOUT_CALLBACK(/root/.ansible.cfg) = debug HOST_KEY_CHECKING(/root/.ansible.cfg) = False RETRY_FILES_SAVE_PATH(/root/.ansible.cfg) = /root/.ansible/retry ``` ##### OS / ENVIRONMENT <!--- Mention, BELOW THIS COMMENT, the OS you are running Ansible from, and the OS you are managing, or say "N/A" for anything that is not platform-specific. Also mention the specific version of what you are trying to control, e.g. if this is a network bug the version of firmware on the network device.--> Linux 4.4.0-124-generic #148-Ubuntu SMP Wed May 2 13:00:18 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem, using a minimal test-case. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ```yaml --- - name: "-- proof of concept" hosts: localhost - name: "-- archive something" archive: format: xz path: "./something" dest: "./something.xz" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Ansible creates package something.xz ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> ``` MSG: lzma or backports.lzma is required when using xz format. ``` --- ##### Further testing and other debug information <!--- Paste verbatim command output between quotes below --> ``` 13:56$ > python Python 2.7.12 (default, Dec 4 2017, 14:50:18) [GCC 5.4.0 20160609] on linux2 Type "help", "copyright", "credits" or "license" for more information. >>> from backports import lzma Traceback (most recent call last): File "<stdin>", line 1, in <module> ImportError: cannot import name lzma >>> import lzma >>> ``` --- ``` temporary-build-server @ root: ~ 13:56$ > python3 Python 3.5.2 (default, Nov 23 2017, 16:37:01) [GCC 5.4.0 20160609] on linux Type "help", "copyright", "credits" or "license" for more information. >>> from backports import lzma Traceback (most recent call last): File "<stdin>", line 1, in <module> ImportError: No module named 'backports' >>> import lzma >>> ``` --- ``` 13:57$ > dpkg -l |grep lzma ii liblzma5:amd64 5.1.1alpha+20120614-2ubuntu2 amd64 XZ-format compression library ii lzma 9.22-2ubuntu2 amd64 Compression and decompression in the LZMA format - command line utility ii python-lzma 0.5.3-3 amd64 Python bindings for liblzma ``` --- ##### Source code modifications __this section seem to be the culprit__ https://github.com/ansible/ansible/blob/9ff20521d1ada2acc64b623875b1d8e51809e0f9/lib/ansible/modules/files/archive.py#L150-L161 Since python is identifying my installation as python 2.7 it defaults to asking for backports, even tho the actual lzma import is there. This might be caused by lzma being installed through `apt` and not `pip` In my humble opinion the code should check for both lzma and backports.lzma. --- ##### Temporary fix Installing `backports.lzma` through pip and pip3 seems to fix the issue on my machine.
main
archive module doesn t work with python and lzma package asks for backport verify first that your issue request is not already reported on github this form will be read by a machine complete all sections as described also test if the latest release and devel branch are affected too always add information after outside these html comments otherwise it may end up being automatically closed by our bot summary i ve tried using the xz archive extension from ansible and can t seem to make it work whenever i try to execute the script ansible errors out with the message lzma or backports lzma is required when using xz format issue type bug report component name insert below this comment the name of the module plugin task or feature do not include extra details here e g vyos command not the network module vyos command or the full path archive ansible version ansible config file root ansible cfg configured module search path ansible python module location usr local lib dist packages ansible egg ansible executable location usr local bin ansible python version default dec configuration if using ansible or above paste below this comment the results of ansible config dump only changed otherwise mention any settings you have changed added removed in ansible cfg or using the ansible environment variables default callback whitelist root ansible cfg default remote user root ansible cfg root default stdout callback root ansible cfg debug host key checking root ansible cfg false retry files save path root ansible cfg root ansible retry os environment mention below this comment the os you are running ansible from and the os you are managing or say n a for anything that is not platform specific also mention the specific version of what you are trying to control e g if this is a network bug the version of firmware on the network device linux generic ubuntu smp wed may utc gnu linux steps to reproduce for bugs show exactly how to reproduce the problem using a minimal test case for new features show how the feature would be used yaml name proof of concept hosts localhost name archive something archive format xz path something dest something xz expected results ansible creates package something xz actual results msg lzma or backports lzma is required when using xz format further testing and other debug information python python default dec on type help copyright credits or license for more information from backports import lzma traceback most recent call last file line in importerror cannot import name lzma import lzma temporary build server root python default nov on linux type help copyright credits or license for more information from backports import lzma traceback most recent call last file line in importerror no module named backports import lzma dpkg l grep lzma ii xz format compression library ii lzma compression and decompression in the lzma format command line utility ii python lzma python bindings for liblzma source code modifications this section seem to be the culprit since python is identifying my installation as python it defaults to asking for backports even tho the actual lzma import is there this might be caused by lzma being installed through apt and not pip in my humble opinion the code should check for both lzma and backports lzma temporary fix installing backports lzma through pip and seems to fix the issue on my machine
1
49,475
12,345,381,157
IssuesEvent
2020-05-15 08:51:10
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
opened
There is no synchronization between replicas, neither in readonly state nor in synchronization
build
2020.05.15 16:50:06.800472 [ 84 ] {} <Error> k19_test.replica_shard (ReplicatedMergeTreeRestartingThread): void DB::ReplicatedMergeTreeRestartingThread::run(): Code: 27, e.displayText() = DB::Exception: Cannot parse input: expected format version: at end of stream., Stack trace (when copying this message, always include the lines below): 0. 0x100ac1bc Poco::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse 1. 0x8e74849 DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse 2. 0x8eaacd5 ? in /usr/bin/clickhouse 3. 0x8ea8caa DB::assertString(char const*, DB::ReadBuffer&) in /usr/bin/clickhouse 4. 0xd78dd1b DB::ReplicatedMergeTreeLogEntryData::readText(DB::ReadBuffer&) in /usr/bin/clickhouse 5. 0xd78f04b DB::ReplicatedMergeTreeLogEntry::parse(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, Coordination::Stat const&) in /usr/bin/clickhouse 6. 0xd7b49c1 DB::ReplicatedMergeTreeQueue::load(std::__1::shared_ptr<zkutil::ZooKeeper>) in /usr/bin/clickhouse 7. 0xd7d7393 DB::ReplicatedMergeTreeRestartingThread::tryStartup() in /usr/bin/clickhouse 8. 0xd7d7cf8 DB::ReplicatedMergeTreeRestartingThread::run() in /usr/bin/clickhouse 9. 0xcd939f1 DB::BackgroundSchedulePoolTaskInfo::execute() in /usr/bin/clickhouse 10. 0xcd93fca DB::BackgroundSchedulePool::threadFunction() in /usr/bin/clickhouse 11. 0xcd94100 ? in /usr/bin/clickhouse 12. 0x8e97347 ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) in /usr/bin/clickhouse 13. 0x8e9580f ? in /usr/bin/clickhouse 14. 0x7e25 start_thread in /usr/lib64/libpthread-2.17.so 15. 0xfebad __clone in /usr/lib64/libc-2.17.so (version 20.1.6.30 (official build))
1.0
There is no synchronization between replicas, neither in readonly state nor in synchronization - 2020.05.15 16:50:06.800472 [ 84 ] {} <Error> k19_test.replica_shard (ReplicatedMergeTreeRestartingThread): void DB::ReplicatedMergeTreeRestartingThread::run(): Code: 27, e.displayText() = DB::Exception: Cannot parse input: expected format version: at end of stream., Stack trace (when copying this message, always include the lines below): 0. 0x100ac1bc Poco::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse 1. 0x8e74849 DB::Exception::Exception(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, int) in /usr/bin/clickhouse 2. 0x8eaacd5 ? in /usr/bin/clickhouse 3. 0x8ea8caa DB::assertString(char const*, DB::ReadBuffer&) in /usr/bin/clickhouse 4. 0xd78dd1b DB::ReplicatedMergeTreeLogEntryData::readText(DB::ReadBuffer&) in /usr/bin/clickhouse 5. 0xd78f04b DB::ReplicatedMergeTreeLogEntry::parse(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, Coordination::Stat const&) in /usr/bin/clickhouse 6. 0xd7b49c1 DB::ReplicatedMergeTreeQueue::load(std::__1::shared_ptr<zkutil::ZooKeeper>) in /usr/bin/clickhouse 7. 0xd7d7393 DB::ReplicatedMergeTreeRestartingThread::tryStartup() in /usr/bin/clickhouse 8. 0xd7d7cf8 DB::ReplicatedMergeTreeRestartingThread::run() in /usr/bin/clickhouse 9. 0xcd939f1 DB::BackgroundSchedulePoolTaskInfo::execute() in /usr/bin/clickhouse 10. 0xcd93fca DB::BackgroundSchedulePool::threadFunction() in /usr/bin/clickhouse 11. 0xcd94100 ? in /usr/bin/clickhouse 12. 0x8e97347 ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) in /usr/bin/clickhouse 13. 0x8e9580f ? in /usr/bin/clickhouse 14. 0x7e25 start_thread in /usr/lib64/libpthread-2.17.so 15. 0xfebad __clone in /usr/lib64/libc-2.17.so (version 20.1.6.30 (official build))
non_main
there is no synchronization between replicas neither in readonly state nor in synchronization test replica shard replicatedmergetreerestartingthread void db replicatedmergetreerestartingthread run code e displaytext db exception cannot parse input expected format version at end of stream stack trace when copying this message always include the lines below poco exception exception std basic string std allocator const int in usr bin clickhouse db exception exception std basic string std allocator const int in usr bin clickhouse in usr bin clickhouse db assertstring char const db readbuffer in usr bin clickhouse db replicatedmergetreelogentrydata readtext db readbuffer in usr bin clickhouse db replicatedmergetreelogentry parse std basic string std allocator const coordination stat const in usr bin clickhouse db replicatedmergetreequeue load std shared ptr in usr bin clickhouse db replicatedmergetreerestartingthread trystartup in usr bin clickhouse db replicatedmergetreerestartingthread run in usr bin clickhouse db backgroundschedulepooltaskinfo execute in usr bin clickhouse db backgroundschedulepool threadfunction in usr bin clickhouse in usr bin clickhouse threadpoolimpl worker std list iterator in usr bin clickhouse in usr bin clickhouse start thread in usr libpthread so clone in usr libc so version official build
0
625,902
19,770,029,514
IssuesEvent
2022-01-17 09:06:16
scipp/scipp
https://api.github.com/repos/scipp/scipp
closed
Consider wider build matrix options
discussion priority:low
Solving regressions resulting from changes in compiler flags and linking options is time consuming. Would it be better to build with some of these options set to exercise the functionality and find regressions earlier. A few more recent aspect have made me raise this question. - With #1670 we had a test segfault, that actually displayed as an assert error in debug mode. Would have been easier found if this was visible in the logs. - Recent work on dynamic libs for scipp (optional) - Recent introduction of precombiled headers (optional) - A thought on packaging -It may actually be desirable to package and distribute via conda Release with Debug information. Would allow users to swap packages on beamlines, re-run and attach debugger We will have to strike a balance between resource use and time (of CI) and resource use an time of developers. **Questions** - What options would we exercise? - How frequently should we run with these options? Related issue: https://github.com/scipp/scipp/issues/1244
1.0
Consider wider build matrix options - Solving regressions resulting from changes in compiler flags and linking options is time consuming. Would it be better to build with some of these options set to exercise the functionality and find regressions earlier. A few more recent aspect have made me raise this question. - With #1670 we had a test segfault, that actually displayed as an assert error in debug mode. Would have been easier found if this was visible in the logs. - Recent work on dynamic libs for scipp (optional) - Recent introduction of precombiled headers (optional) - A thought on packaging -It may actually be desirable to package and distribute via conda Release with Debug information. Would allow users to swap packages on beamlines, re-run and attach debugger We will have to strike a balance between resource use and time (of CI) and resource use an time of developers. **Questions** - What options would we exercise? - How frequently should we run with these options? Related issue: https://github.com/scipp/scipp/issues/1244
non_main
consider wider build matrix options solving regressions resulting from changes in compiler flags and linking options is time consuming would it be better to build with some of these options set to exercise the functionality and find regressions earlier a few more recent aspect have made me raise this question with we had a test segfault that actually displayed as an assert error in debug mode would have been easier found if this was visible in the logs recent work on dynamic libs for scipp optional recent introduction of precombiled headers optional a thought on packaging it may actually be desirable to package and distribute via conda release with debug information would allow users to swap packages on beamlines re run and attach debugger we will have to strike a balance between resource use and time of ci and resource use an time of developers questions what options would we exercise how frequently should we run with these options related issue
0
55,181
23,408,153,120
IssuesEvent
2022-08-12 14:44:02
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Expressions: Improve test coverage
Feature:ExpressionLanguage loe:week Team:AppServicesSv impact:medium
Improve test coverage in Expressions plugin `src/plugins/expressions`. - [ ] Add unit tests for everything - [ ] Add few integration tests for the whole `expressions` plugin - [ ] Set-up test coverage reporting - [ ] Look into [moving interpreter functional snapshots to use the unified snapshot interface](https://github.com/elastic/kibana/issues/83955) Parent issue: https://github.com/elastic/kibana/issues/46909
1.0
Expressions: Improve test coverage - Improve test coverage in Expressions plugin `src/plugins/expressions`. - [ ] Add unit tests for everything - [ ] Add few integration tests for the whole `expressions` plugin - [ ] Set-up test coverage reporting - [ ] Look into [moving interpreter functional snapshots to use the unified snapshot interface](https://github.com/elastic/kibana/issues/83955) Parent issue: https://github.com/elastic/kibana/issues/46909
non_main
expressions improve test coverage improve test coverage in expressions plugin src plugins expressions add unit tests for everything add few integration tests for the whole expressions plugin set up test coverage reporting look into parent issue
0
443,405
30,888,764,199
IssuesEvent
2023-08-04 01:47:51
psf/black
https://api.github.com/repos/psf/black
closed
pre-commit hook fails when using `required-version`
T: documentation C: integrations C: configuration
This is attempting to use the pre-commit hook as described under [version control docs](https://black.readthedocs.io/en/stable/integrations/source_version_control.html) along with setting [--required-version](https://black.readthedocs.io/en/stable/usage_and_configuration/the_basics.html#versions) in a configuration file. The pre-commit hook fails with: ``` Oh no! πŸ’₯ πŸ’” πŸ’₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`! ``` **To Reproduce** First, a temporary folder/venv to test things in: ``` mkdir black-pre-commit-test cd black-pre-commit-test `which python3.9` -m venv venv source venv/bin/activate ``` Then this script: ``` git init pip install pre-commit pre-commit install cat > .pre-commit-config.yaml << EOF repos: - repo: https://github.com/psf/black rev: 21.8b0 hooks: - id: black language_version: python3 EOF cat > pyproject.toml << EOF [tool.black] required-version = '21.8b0' EOF git add .pre-commit-config.yaml touch test.py pre-commit run black --files test.py ``` **Expected output** ``` black....................................................................Passed ``` **Actual output** ``` black....................................................................Failed - hook id: black - exit code: 1 Oh no! πŸ’₯ πŸ’” πŸ’₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`! ``` The same happens if doing a normal "git commit" - it is just easier to run the hook via `pre-commit run` directly. To test with hook: ``` git commit .pre-commit-config.yaml -m "pre-commit" git add test.py git commit test.py -m "test" ``` **Environment:** - Version: 21.8b0 - OS: Linux, - Python: 3.9.5 (i.e. `which python3.9` used in the beginning of the instructions above) - However, the version installed by pre-commit in `~/.cache/pre-commit/repo8_9vcuko/py_env-python3/bin/python3` is different - it is whatever is pointed to by `which python3`, which was Python 3.8.10 first time I ran it. By changing `language_version` in `.pre-commit-config.yaml` I can get a different version, e.g. Python 3.9.5, but it made no difference. **Does this bug also happen on main?** Haven't tried, because this is specific to pinning to published versions. **Notes** I can see the incorrect Black version by doing this: ``` ~/.cache/pre-commit/repo8_9vcuko/py_env-python3.9/bin/black --version ``` However, I haven't been able to further debug because I don't know how this line works: https://github.com/psf/black/blob/41e670064063e3e221b1c3c40db5aaae784b5231/src/black/__init__.py#L67 - it appears to be doing some magical import.
1.0
pre-commit hook fails when using `required-version` - This is attempting to use the pre-commit hook as described under [version control docs](https://black.readthedocs.io/en/stable/integrations/source_version_control.html) along with setting [--required-version](https://black.readthedocs.io/en/stable/usage_and_configuration/the_basics.html#versions) in a configuration file. The pre-commit hook fails with: ``` Oh no! πŸ’₯ πŸ’” πŸ’₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`! ``` **To Reproduce** First, a temporary folder/venv to test things in: ``` mkdir black-pre-commit-test cd black-pre-commit-test `which python3.9` -m venv venv source venv/bin/activate ``` Then this script: ``` git init pip install pre-commit pre-commit install cat > .pre-commit-config.yaml << EOF repos: - repo: https://github.com/psf/black rev: 21.8b0 hooks: - id: black language_version: python3 EOF cat > pyproject.toml << EOF [tool.black] required-version = '21.8b0' EOF git add .pre-commit-config.yaml touch test.py pre-commit run black --files test.py ``` **Expected output** ``` black....................................................................Passed ``` **Actual output** ``` black....................................................................Failed - hook id: black - exit code: 1 Oh no! πŸ’₯ πŸ’” πŸ’₯ The required version `21.8b0` does not match the running version `0.1.dev1+ga8b4665`! ``` The same happens if doing a normal "git commit" - it is just easier to run the hook via `pre-commit run` directly. To test with hook: ``` git commit .pre-commit-config.yaml -m "pre-commit" git add test.py git commit test.py -m "test" ``` **Environment:** - Version: 21.8b0 - OS: Linux, - Python: 3.9.5 (i.e. `which python3.9` used in the beginning of the instructions above) - However, the version installed by pre-commit in `~/.cache/pre-commit/repo8_9vcuko/py_env-python3/bin/python3` is different - it is whatever is pointed to by `which python3`, which was Python 3.8.10 first time I ran it. By changing `language_version` in `.pre-commit-config.yaml` I can get a different version, e.g. Python 3.9.5, but it made no difference. **Does this bug also happen on main?** Haven't tried, because this is specific to pinning to published versions. **Notes** I can see the incorrect Black version by doing this: ``` ~/.cache/pre-commit/repo8_9vcuko/py_env-python3.9/bin/black --version ``` However, I haven't been able to further debug because I don't know how this line works: https://github.com/psf/black/blob/41e670064063e3e221b1c3c40db5aaae784b5231/src/black/__init__.py#L67 - it appears to be doing some magical import.
non_main
pre commit hook fails when using required version this is attempting to use the pre commit hook as described under along with setting in a configuration file the pre commit hook fails with oh no πŸ’₯ πŸ’” πŸ’₯ the required version does not match the running version to reproduce first a temporary folder venv to test things in mkdir black pre commit test cd black pre commit test which m venv venv source venv bin activate then this script git init pip install pre commit pre commit install cat pre commit config yaml eof repos repo rev hooks id black language version eof cat pyproject toml eof required version eof git add pre commit config yaml touch test py pre commit run black files test py expected output black passed actual output black failed hook id black exit code oh no πŸ’₯ πŸ’” πŸ’₯ the required version does not match the running version the same happens if doing a normal git commit it is just easier to run the hook via pre commit run directly to test with hook git commit pre commit config yaml m pre commit git add test py git commit test py m test environment version os linux python i e which used in the beginning of the instructions above however the version installed by pre commit in cache pre commit py env bin is different it is whatever is pointed to by which which was python first time i ran it by changing language version in pre commit config yaml i can get a different version e g python but it made no difference does this bug also happen on main haven t tried because this is specific to pinning to published versions notes i can see the incorrect black version by doing this cache pre commit py env bin black version however i haven t been able to further debug because i don t know how this line works it appears to be doing some magical import
0
2,299
8,221,956,555
IssuesEvent
2018-09-06 05:11:51
TravisSpark/spark-website
https://api.github.com/repos/TravisSpark/spark-website
closed
Edit Readme
maintainence
### Checklist - [X] Searched for, and did not find, duplicate [issue](https://github.com/TravisSpark/spark-website/issues) - [X] Indicated whether the issue is a bug or a feature - [X] Focused on one specific bug/feature - [X] Gave a concise and relevant name - [X] Created clear and concise description - [X] Outlined which components are affected - [X] Assigned issue to project, appropriate contributors, and relevant labels <!-- Edit as Appropriate --> ### Issue Type: Feature ### Description The descriptions in the read me lack impact and clarity. The phrasing should be edited. Use the homepage of travisspark.org as a template. ### Affected Components .readme
True
Edit Readme - ### Checklist - [X] Searched for, and did not find, duplicate [issue](https://github.com/TravisSpark/spark-website/issues) - [X] Indicated whether the issue is a bug or a feature - [X] Focused on one specific bug/feature - [X] Gave a concise and relevant name - [X] Created clear and concise description - [X] Outlined which components are affected - [X] Assigned issue to project, appropriate contributors, and relevant labels <!-- Edit as Appropriate --> ### Issue Type: Feature ### Description The descriptions in the read me lack impact and clarity. The phrasing should be edited. Use the homepage of travisspark.org as a template. ### Affected Components .readme
main
edit readme checklist searched for and did not find duplicate indicated whether the issue is a bug or a feature focused on one specific bug feature gave a concise and relevant name created clear and concise description outlined which components are affected assigned issue to project appropriate contributors and relevant labels issue type feature description the descriptions in the read me lack impact and clarity the phrasing should be edited use the homepage of travisspark org as a template affected components readme
1
108,663
23,644,045,693
IssuesEvent
2022-08-25 20:02:18
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] No way to change the site name during installation
No Code Attached Yet J4 Issue
### Steps to reproduce the issue Start an installation. Enter the site name, go to "setup login data" (and mybe to the database setup) Did you enter the site name right? Do you want to update the site name? ### Expected result You can see the site name you have entered in the first step. You can change the site name ### Actual result The sie name is not visible and there is no way back to change the site name (except restarting the installation)
1.0
[4.0] No way to change the site name during installation - ### Steps to reproduce the issue Start an installation. Enter the site name, go to "setup login data" (and mybe to the database setup) Did you enter the site name right? Do you want to update the site name? ### Expected result You can see the site name you have entered in the first step. You can change the site name ### Actual result The sie name is not visible and there is no way back to change the site name (except restarting the installation)
non_main
no way to change the site name during installation steps to reproduce the issue start an installation enter the site name go to setup login data and mybe to the database setup did you enter the site name right do you want to update the site name expected result you can see the site name you have entered in the first step you can change the site name actual result the sie name is not visible and there is no way back to change the site name except restarting the installation
0
16,547
2,914,792,045
IssuesEvent
2015-06-23 08:23:07
beefproject/beef
https://api.github.com/repos/beefproject/beef
closed
beef.net.request callbacks aren't being called
Core Defect Priority Low
I think I found an odd issue (Whilst working on Issue #1083) .. it appears that outside of the use of beef.net.request in core/main/client/updater.js (https://github.com/beefproject/beef/blob/master/core/main/client/updater.js#L60) the use of callbacks in beef.net.request() method calls do not actually get used. I found this in the integration tests this module didn't work properly (https://github.com/beefproject/beef/blob/master/test/integration/tc_debug_modules.rb#L141) And then subsequently, I couldn't get the coldfusion_dir_traversal_exploit module to work as it relies on this pattern too. I'm unsure why this isn't working, whether the variables don't get passed into the callback function, or perhaps the module's use is slightly incorrect. Will investigate further.
1.0
beef.net.request callbacks aren't being called - I think I found an odd issue (Whilst working on Issue #1083) .. it appears that outside of the use of beef.net.request in core/main/client/updater.js (https://github.com/beefproject/beef/blob/master/core/main/client/updater.js#L60) the use of callbacks in beef.net.request() method calls do not actually get used. I found this in the integration tests this module didn't work properly (https://github.com/beefproject/beef/blob/master/test/integration/tc_debug_modules.rb#L141) And then subsequently, I couldn't get the coldfusion_dir_traversal_exploit module to work as it relies on this pattern too. I'm unsure why this isn't working, whether the variables don't get passed into the callback function, or perhaps the module's use is slightly incorrect. Will investigate further.
non_main
beef net request callbacks aren t being called i think i found an odd issue whilst working on issue it appears that outside of the use of beef net request in core main client updater js the use of callbacks in beef net request method calls do not actually get used i found this in the integration tests this module didn t work properly and then subsequently i couldn t get the coldfusion dir traversal exploit module to work as it relies on this pattern too i m unsure why this isn t working whether the variables don t get passed into the callback function or perhaps the module s use is slightly incorrect will investigate further
0
703
4,281,278,518
IssuesEvent
2016-07-15 01:45:48
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Counter-Strike Go Cheat Sheet: Add aliases
Maintainer Input Requested Suggestion
This could probably benefit from a few aliases, for example: "counter-strike go", "counter strike go", "cs:go", "counter-strike global offensive", "counter strike global offensive" ------ IA Page: http://duck.co/ia/view/csgo_cheat_sheet [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @aiyankovil
True
Counter-Strike Go Cheat Sheet: Add aliases - This could probably benefit from a few aliases, for example: "counter-strike go", "counter strike go", "cs:go", "counter-strike global offensive", "counter strike global offensive" ------ IA Page: http://duck.co/ia/view/csgo_cheat_sheet [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @aiyankovil
main
counter strike go cheat sheet add aliases this could probably benefit from a few aliases for example counter strike go counter strike go cs go counter strike global offensive counter strike global offensive ia page aiyankovil
1
4,993
25,708,406,639
IssuesEvent
2022-12-07 03:35:40
aws/serverless-application-model
https://api.github.com/repos/aws/serverless-application-model
closed
Accept boolean values in CloudWatchEvent patterns
type/feature maintainer/need-response
I have a Lambda that is triggered by CloudWatchEvents. I'm trying to match events that have "campaign" in their details as in [this example of the documentation](https://docs.aws.amazon.com/eventbridge/latest/userguide/content-filtering-with-event-patterns.html#filtering-exists-matching). This is the code in the lambda's Events property of my SAM template: ``` feedPost: Type: CloudWatchEvent Properties: EventBusName: my_bus Pattern: source: - mySource detail: campaign: - exists: true ``` But when it is deploying, in CloudFormation the following error appears: > Event pattern is not valid. Reason: exists match pattern must be either true or false. at [Source: (String)"{ "source":["mySource"],"detail":{"campaign":[{"exists":"true"}]}}"; So apparently it is taking the true value as a String and fails to deploy.
True
Accept boolean values in CloudWatchEvent patterns - I have a Lambda that is triggered by CloudWatchEvents. I'm trying to match events that have "campaign" in their details as in [this example of the documentation](https://docs.aws.amazon.com/eventbridge/latest/userguide/content-filtering-with-event-patterns.html#filtering-exists-matching). This is the code in the lambda's Events property of my SAM template: ``` feedPost: Type: CloudWatchEvent Properties: EventBusName: my_bus Pattern: source: - mySource detail: campaign: - exists: true ``` But when it is deploying, in CloudFormation the following error appears: > Event pattern is not valid. Reason: exists match pattern must be either true or false. at [Source: (String)"{ "source":["mySource"],"detail":{"campaign":[{"exists":"true"}]}}"; So apparently it is taking the true value as a String and fails to deploy.
main
accept boolean values in cloudwatchevent patterns i have a lambda that is triggered by cloudwatchevents i m trying to match events that have campaign in their details as in this is the code in the lambda s events property of my sam template feedpost type cloudwatchevent properties eventbusname my bus pattern source mysource detail campaign exists true but when it is deploying in cloudformation the following error appears event pattern is not valid reason exists match pattern must be either true or false at detail campaign so apparently it is taking the true value as a string and fails to deploy
1
5,680
29,833,373,332
IssuesEvent
2023-06-18 14:40:37
Windham-High-School/CubeServer
https://api.github.com/repos/Windham-High-School/CubeServer
closed
Refactor scoring code
maintainability
The current scoring code is ridiculously hard to follow and read, and is written in the same file as a database object mapping model class!
True
Refactor scoring code - The current scoring code is ridiculously hard to follow and read, and is written in the same file as a database object mapping model class!
main
refactor scoring code the current scoring code is ridiculously hard to follow and read and is written in the same file as a database object mapping model class
1
1,089
4,939,643,278
IssuesEvent
2016-11-29 14:55:51
numbbo/coco
https://api.github.com/repos/numbbo/coco
opened
data archives folder name convention
Maintainability question Usability
This is how it looks like currently: <img width="381" alt="screen shot 2016-11-29 at 15 53 46" src="https://cloud.githubusercontent.com/assets/7316439/20714360/177deb12-b64c-11e6-9225-e7a61a78c48c.png"> Shouldn't we rename `bbob-biobj-2016` to `2016-bbob-biobj`? Do we have dependencies which would brake? Even after this, the naming convention is inconsistent, as we have noiseless and noisy testbeds mixed. How about to have for each year/event also a subfolder for each testbed?
True
data archives folder name convention - This is how it looks like currently: <img width="381" alt="screen shot 2016-11-29 at 15 53 46" src="https://cloud.githubusercontent.com/assets/7316439/20714360/177deb12-b64c-11e6-9225-e7a61a78c48c.png"> Shouldn't we rename `bbob-biobj-2016` to `2016-bbob-biobj`? Do we have dependencies which would brake? Even after this, the naming convention is inconsistent, as we have noiseless and noisy testbeds mixed. How about to have for each year/event also a subfolder for each testbed?
main
data archives folder name convention this is how it looks like currently img width alt screen shot at src shouldn t we rename bbob biobj to bbob biobj do we have dependencies which would brake even after this the naming convention is inconsistent as we have noiseless and noisy testbeds mixed how about to have for each year event also a subfolder for each testbed
1
4,659
24,097,560,643
IssuesEvent
2022-09-19 20:15:13
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Sort sam init runtimes alphanumerically
type/ux area/init maintainer/need-followup
### Describe your idea/feature/enhancement ![image](https://user-images.githubusercontent.com/39839589/74185353-53070700-4bfd-11ea-9e5b-6d57667f4543.png) The list of runtimes shown in `sam init` are hard to look through. For example, if I was interested in a node app, I have to hunt through the list to find one up top, and one in the middle (7 in the image). I think two improvements could help improve time to locate the runtime of interest: * alphanumeric-sort the runtimes * right-align the numbers (1-13 in the image)
True
Sort sam init runtimes alphanumerically - ### Describe your idea/feature/enhancement ![image](https://user-images.githubusercontent.com/39839589/74185353-53070700-4bfd-11ea-9e5b-6d57667f4543.png) The list of runtimes shown in `sam init` are hard to look through. For example, if I was interested in a node app, I have to hunt through the list to find one up top, and one in the middle (7 in the image). I think two improvements could help improve time to locate the runtime of interest: * alphanumeric-sort the runtimes * right-align the numbers (1-13 in the image)
main
sort sam init runtimes alphanumerically describe your idea feature enhancement the list of runtimes shown in sam init are hard to look through for example if i was interested in a node app i have to hunt through the list to find one up top and one in the middle in the image i think two improvements could help improve time to locate the runtime of interest alphanumeric sort the runtimes right align the numbers in the image
1
6,706
9,815,575,684
IssuesEvent
2019-06-13 12:58:01
linnovate/root
https://api.github.com/repos/linnovate/root
closed
in tasks and projects, "no select" option when selecting an assignee changes the status to assigned
2.0.7 Fixed Process bug Projects Tasks
create a new task/ project click on select assignee click on "no select" the status is changed to assigned ![image](https://user-images.githubusercontent.com/38312178/50585726-0f30c280-0e7f-11e9-952e-544495e9a4b7.png)
1.0
in tasks and projects, "no select" option when selecting an assignee changes the status to assigned - create a new task/ project click on select assignee click on "no select" the status is changed to assigned ![image](https://user-images.githubusercontent.com/38312178/50585726-0f30c280-0e7f-11e9-952e-544495e9a4b7.png)
non_main
in tasks and projects no select option when selecting an assignee changes the status to assigned create a new task project click on select assignee click on no select the status is changed to assigned
0
814,793
30,522,160,222
IssuesEvent
2023-07-19 08:49:06
owncloud/web
https://api.github.com/repos/owncloud/web
closed
Search location filter is modified after opening a file in the text editor
Type:Bug Priority:p3-medium
steps: Precondition: user has files with the same "content" in different places (personal, project, shares jail) Steps: - user searches files using fullTextSearch and `all files` filter - open file from personal space in the text editor - close text editor Expected: `all files` filter in the search result. Actual: `current folder` filter in the search result. https://github.com/owncloud/web/assets/84779829/0a735e01-be71-42e6-94f7-f317f2247fbc
1.0
Search location filter is modified after opening a file in the text editor - steps: Precondition: user has files with the same "content" in different places (personal, project, shares jail) Steps: - user searches files using fullTextSearch and `all files` filter - open file from personal space in the text editor - close text editor Expected: `all files` filter in the search result. Actual: `current folder` filter in the search result. https://github.com/owncloud/web/assets/84779829/0a735e01-be71-42e6-94f7-f317f2247fbc
non_main
search location filter is modified after opening a file in the text editor steps precondition user has files with the same content in different places personal project shares jail steps user searches files using fulltextsearch and all files filter open file from personal space in the text editor close text editor expected all files filter in the search result actual current folder filter in the search result
0
1,411
6,130,961,657
IssuesEvent
2017-06-24 11:01:17
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
Remove TimeSeries
API Component-Core Format-DCD maintainability
We discussed recently to remove the TimeSeries objects from MDAnalysis. As far as I know none of the core devs is using it. We also note that it is unlikely we will ever implement TimeSeries for any other readers besides dcd and memory and most of it's function can be replaced by the Analysis Framework. The only part in the library that uses TimeSeries is encore. So it would be useful get some input from @wouterboomsma and @mtiberti before we go ahead with this. As a first step I would remove timeseries from DCD during the cython transition. In a later step we would remove timeseries from memory reader and remove the core TimeSeries class.
True
Remove TimeSeries - We discussed recently to remove the TimeSeries objects from MDAnalysis. As far as I know none of the core devs is using it. We also note that it is unlikely we will ever implement TimeSeries for any other readers besides dcd and memory and most of it's function can be replaced by the Analysis Framework. The only part in the library that uses TimeSeries is encore. So it would be useful get some input from @wouterboomsma and @mtiberti before we go ahead with this. As a first step I would remove timeseries from DCD during the cython transition. In a later step we would remove timeseries from memory reader and remove the core TimeSeries class.
main
remove timeseries we discussed recently to remove the timeseries objects from mdanalysis as far as i know none of the core devs is using it we also note that it is unlikely we will ever implement timeseries for any other readers besides dcd and memory and most of it s function can be replaced by the analysis framework the only part in the library that uses timeseries is encore so it would be useful get some input from wouterboomsma and mtiberti before we go ahead with this as a first step i would remove timeseries from dcd during the cython transition in a later step we would remove timeseries from memory reader and remove the core timeseries class
1
4,261
21,261,293,875
IssuesEvent
2022-04-13 04:45:56
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
ERPNEXT 13 Installation Easy install error URGENT !!!!
blocked/close-if-inactive maintainer/need-followup
I have tried to install ERPnext 13 more than 20 times , using all the tutorials but no good result , when I use Easy install [ sudo python3 install.py --verbose --production --user [USER] --mariadb-version 10.5 --frappe-branch version-13 --erpnext-branch version-13 ] I got error , return none zero status 2 Could you please help me out the whole result fatal: [localhost]: FAILED! => { "changed": true, "cmd": [ "bench", "init", "/home/frappe/frappe-bench", "--frappe-path", "https://github.com/frappe/frappe", "--frappe-branch", "version-13", "--python", "python3" ], "delta": "0:00:00.378029", "end": "2022-02-20 10:45:29.953302", "invocation": { "module_args": { "_raw_params": "bench init /home/frappe/frappe-bench --frappe-path https://github.com/frappe/frappe --frappe-branch version-13 --python python3", "_uses_shell": false, "argv": null, "chdir": null, "creates": "/home/frappe/frappe-bench", "executable": null, "removes": null, "stdin": null, "stdin_add_newline": true, "strip_empty_ends": true, "warn": true } }, "msg": "non-zero return code", "rc": 1, "start": "2022-02-20 10:45:29.575273", "stderr": "Traceback (most recent call last):\n File "/usr/bin/bench", line 33, in \n sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())\n File "/usr/bin/bench", line 22, in importlib_load_entry_point\n for entry_point in distribution(dist_name).entry_points\n File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution\n return Distribution.from_name(distribution_name)\n File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name\n raise PackageNotFoundError(name)\nimportlib.metadata.PackageNotFoundError: frappe-bench", "stderr_lines": [ "Traceback (most recent call last):", " File "/usr/bin/bench", line 33, in ", " sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())", " File "/usr/bin/bench", line 22, in importlib_load_entry_point", " for entry_point in distribution(dist_name).entry_points", " File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution", " return Distribution.from_name(distribution_name)", " File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name", " raise PackageNotFoundError(name)", "importlib.metadata.PackageNotFoundError: frappe-bench" ], "stdout": "", "stdout_lines": [] } PLAY RECAP ***************************************************************************************************************************************************************** localhost : ok=67 changed=17 unreachable=0 failed=1 skipped=62 rescued=0 ignored=0 Traceback (most recent call last): File "install.py", line 497, in install_bench(args) File "install.py", line 278, in install_bench run_playbook('site.yml', sudo=True, extra_vars=extra_vars) File "install.py", line 413, in run_playbook success = subprocess.check_call(args, cwd=playbooks_folder, stdout=log_stream, stderr=sys.stderr) File "/usr/lib/python3.8/subprocess.py", line 364, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['ansible-playbook', '-c', 'local', 'site.yml', '-vvvv', '-e', '@/tmp/extra_vars.json', '--become', '--become-user=frappe']' returned non-zero exit status 2.
True
ERPNEXT 13 Installation Easy install error URGENT !!!! - I have tried to install ERPnext 13 more than 20 times , using all the tutorials but no good result , when I use Easy install [ sudo python3 install.py --verbose --production --user [USER] --mariadb-version 10.5 --frappe-branch version-13 --erpnext-branch version-13 ] I got error , return none zero status 2 Could you please help me out the whole result fatal: [localhost]: FAILED! => { "changed": true, "cmd": [ "bench", "init", "/home/frappe/frappe-bench", "--frappe-path", "https://github.com/frappe/frappe", "--frappe-branch", "version-13", "--python", "python3" ], "delta": "0:00:00.378029", "end": "2022-02-20 10:45:29.953302", "invocation": { "module_args": { "_raw_params": "bench init /home/frappe/frappe-bench --frappe-path https://github.com/frappe/frappe --frappe-branch version-13 --python python3", "_uses_shell": false, "argv": null, "chdir": null, "creates": "/home/frappe/frappe-bench", "executable": null, "removes": null, "stdin": null, "stdin_add_newline": true, "strip_empty_ends": true, "warn": true } }, "msg": "non-zero return code", "rc": 1, "start": "2022-02-20 10:45:29.575273", "stderr": "Traceback (most recent call last):\n File "/usr/bin/bench", line 33, in \n sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())\n File "/usr/bin/bench", line 22, in importlib_load_entry_point\n for entry_point in distribution(dist_name).entry_points\n File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution\n return Distribution.from_name(distribution_name)\n File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name\n raise PackageNotFoundError(name)\nimportlib.metadata.PackageNotFoundError: frappe-bench", "stderr_lines": [ "Traceback (most recent call last):", " File "/usr/bin/bench", line 33, in ", " sys.exit(load_entry_point('frappe-bench', 'console_scripts', 'bench')())", " File "/usr/bin/bench", line 22, in importlib_load_entry_point", " for entry_point in distribution(dist_name).entry_points", " File "/usr/lib/python3.8/importlib/metadata.py", line 503, in distribution", " return Distribution.from_name(distribution_name)", " File "/usr/lib/python3.8/importlib/metadata.py", line 177, in from_name", " raise PackageNotFoundError(name)", "importlib.metadata.PackageNotFoundError: frappe-bench" ], "stdout": "", "stdout_lines": [] } PLAY RECAP ***************************************************************************************************************************************************************** localhost : ok=67 changed=17 unreachable=0 failed=1 skipped=62 rescued=0 ignored=0 Traceback (most recent call last): File "install.py", line 497, in install_bench(args) File "install.py", line 278, in install_bench run_playbook('site.yml', sudo=True, extra_vars=extra_vars) File "install.py", line 413, in run_playbook success = subprocess.check_call(args, cwd=playbooks_folder, stdout=log_stream, stderr=sys.stderr) File "/usr/lib/python3.8/subprocess.py", line 364, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['ansible-playbook', '-c', 'local', 'site.yml', '-vvvv', '-e', '@/tmp/extra_vars.json', '--become', '--become-user=frappe']' returned non-zero exit status 2.
main
erpnext installation easy install error urgent i have tried to install erpnext more than times using all the tutorials but no good result when i use easy install mariadb version frappe branch version erpnext branch version i got error return none zero status could you please help me out the whole result fatal failed changed true cmd bench init home frappe frappe bench frappe path frappe branch version python delta end invocation module args raw params bench init home frappe frappe bench frappe path frappe branch version python uses shell false argv null chdir null creates home frappe frappe bench executable null removes null stdin null stdin add newline true strip empty ends true warn true msg non zero return code rc start stderr traceback most recent call last n file usr bin bench line in n sys exit load entry point frappe bench console scripts bench n file usr bin bench line in importlib load entry point n for entry point in distribution dist name entry points n file usr lib importlib metadata py line in distribution n return distribution from name distribution name n file usr lib importlib metadata py line in from name n raise packagenotfounderror name nimportlib metadata packagenotfounderror frappe bench stderr lines traceback most recent call last file usr bin bench line in sys exit load entry point frappe bench console scripts bench file usr bin bench line in importlib load entry point for entry point in distribution dist name entry points file usr lib importlib metadata py line in distribution return distribution from name distribution name file usr lib importlib metadata py line in from name raise packagenotfounderror name importlib metadata packagenotfounderror frappe bench stdout stdout lines play recap localhost ok changed unreachable failed skipped rescued ignored traceback most recent call last file install py line in install bench args file install py line in install bench run playbook site yml sudo true extra vars extra vars file install py line in run playbook success subprocess check call args cwd playbooks folder stdout log stream stderr sys stderr file usr lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status
1
60,407
14,543,819,136
IssuesEvent
2020-12-15 17:20:58
bitbar/testdroid-api
https://api.github.com/repos/bitbar/testdroid-api
opened
WS-2019-0379 (Medium) detected in commons-codec-1.10.jar
security vulnerability
## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: testdroid-api/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p> <p> Dependency Hierarchy: - httpclient-4.5.6.jar (Root Library) - :x: **commons-codec-1.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bitbar/testdroid-api/commit/e27a23c3fc46b3858353a24f6550013c097bb0c4">e27a23c3fc46b3858353a24f6550013c097bb0c4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version β€œcommons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.10","isTransitiveDependency":true,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.6;commons-codec:commons-codec:1.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version β€œcommons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
WS-2019-0379 (Medium) detected in commons-codec-1.10.jar - ## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: testdroid-api/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p> <p> Dependency Hierarchy: - httpclient-4.5.6.jar (Root Library) - :x: **commons-codec-1.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/bitbar/testdroid-api/commit/e27a23c3fc46b3858353a24f6550013c097bb0c4">e27a23c3fc46b3858353a24f6550013c097bb0c4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version β€œcommons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"commons-codec","packageName":"commons-codec","packageVersion":"1.10","isTransitiveDependency":true,"dependencyTree":"org.apache.httpcomponents:httpclient:4.5.6;commons-codec:commons-codec:1.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.13-RC1"}],"vulnerabilityIdentifier":"WS-2019-0379","vulnerabilityDetails":"Apache commons-codec before version β€œcommons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.","vulnerabilityUrl":"https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_main
ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file testdroid api pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar dependency hierarchy httpclient jar root library x commons codec jar vulnerable library found in head commit a href found in base branch master vulnerability details apache commons codec before version β€œcommons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails apache commons codec before version β€œcommons codec ” is vulnerable to information disclosure due to improper input validation vulnerabilityurl
0
2,996
10,885,066,939
IssuesEvent
2019-11-18 09:39:46
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFP - AudioFlyout
Status: Available For Maintainer(s)
<!-- * Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search. * Please ensure there is no existing open package request. * Please ensure the issue title starts with 'RFP - ' - for example 'RFP - Adobe Reader' * Please ensure you have both the Software Project URL and the Software Download URL before continuing. NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM. --> ## Checklist - [X] The package I am requesting does not already exist on https://chocolatey.org/packages; - [X] There is no open issue for this package; - [X] The issue title starts 'RFP - '; - [X] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://github.com/ADeltaX/AudioFlyout Direct download URL for the software / installer : https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyout.zip https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyoutUA.zip Software summary / short description: Replace the Volume/SMTC UI with a custom one. fluent design style and more available additional controls. <!-- ## Package Expectations Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop -->
True
RFP - AudioFlyout - <!-- * Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search. * Please ensure there is no existing open package request. * Please ensure the issue title starts with 'RFP - ' - for example 'RFP - Adobe Reader' * Please ensure you have both the Software Project URL and the Software Download URL before continuing. NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM. --> ## Checklist - [X] The package I am requesting does not already exist on https://chocolatey.org/packages; - [X] There is no open issue for this package; - [X] The issue title starts 'RFP - '; - [X] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://github.com/ADeltaX/AudioFlyout Direct download URL for the software / installer : https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyout.zip https://github.com/ADeltaX/AudioFlyout/releases/download/0.9.2.0/AudioFlyoutUA.zip Software summary / short description: Replace the Volume/SMTC UI with a custom one. fluent design style and more available additional controls. <!-- ## Package Expectations Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop -->
main
rfp audioflyout please ensure the package does not already exist in the chocolatey community repository by using a relevant search please ensure there is no existing open package request please ensure the issue title starts with rfp for example rfp adobe reader please ensure you have both the software project url and the software download url before continuing note keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the chocolatey community please remove all comments once you have read them checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description replace the volume smtc ui with a custom one fluent design style and more available additional controls package expectations here you can make suggestions on what you would expect the package to do outside of installing eg adding icons to the desktop
1
3,669
14,997,077,640
IssuesEvent
2021-01-29 16:25:07
exercism/python
https://api.github.com/repos/exercism/python
opened
[CI] Update tooling for v3
maintainer action required:grey_exclamation:
## Changed components - Exercises are no longer in just one directory (exercises/). They are in either exercises/concept/ or exercises/practice/ - Example solutions are no longer in the exercise directory itself (two-fer/example.py). They have been/should be moved to .meta/exemplar.py (two-fer/.meta/exemplar.py) - config.json's `exercises` field is no longer an array of exercises. It is an object with two properties: `concept` and `practice`. These each have the same format as the old `exercises` property. - Exercise READMEs are no longer stored as a single, generated file in track repos. They can be removed from exercises if present ## Scripts dependent on changes components: - [ ] bin/template_status.py - [ ] bin/generate_tests.py - [ ] test/check-exercises.py ## Obsolete scripts The following scripts are no longer needed: - [ ] bin/check-readmes.sh - [ ] bin/check-test-version.py (versions no longer used in problem-specifications) ## Ambiguity @ErikSchierboom, we have these scripts. Neither are needed for CI, but which (if either) should be kept for contributors/maintainers who need to run the new v3 configlet locally - [ ] bin/fetch-configlet - [ ] bin/fetch-canonical_data_syncer
True
[CI] Update tooling for v3 - ## Changed components - Exercises are no longer in just one directory (exercises/). They are in either exercises/concept/ or exercises/practice/ - Example solutions are no longer in the exercise directory itself (two-fer/example.py). They have been/should be moved to .meta/exemplar.py (two-fer/.meta/exemplar.py) - config.json's `exercises` field is no longer an array of exercises. It is an object with two properties: `concept` and `practice`. These each have the same format as the old `exercises` property. - Exercise READMEs are no longer stored as a single, generated file in track repos. They can be removed from exercises if present ## Scripts dependent on changes components: - [ ] bin/template_status.py - [ ] bin/generate_tests.py - [ ] test/check-exercises.py ## Obsolete scripts The following scripts are no longer needed: - [ ] bin/check-readmes.sh - [ ] bin/check-test-version.py (versions no longer used in problem-specifications) ## Ambiguity @ErikSchierboom, we have these scripts. Neither are needed for CI, but which (if either) should be kept for contributors/maintainers who need to run the new v3 configlet locally - [ ] bin/fetch-configlet - [ ] bin/fetch-canonical_data_syncer
main
update tooling for changed components exercises are no longer in just one directory exercises they are in either exercises concept or exercises practice example solutions are no longer in the exercise directory itself two fer example py they have been should be moved to meta exemplar py two fer meta exemplar py config json s exercises field is no longer an array of exercises it is an object with two properties concept and practice these each have the same format as the old exercises property exercise readmes are no longer stored as a single generated file in track repos they can be removed from exercises if present scripts dependent on changes components bin template status py bin generate tests py test check exercises py obsolete scripts the following scripts are no longer needed bin check readmes sh bin check test version py versions no longer used in problem specifications ambiguity erikschierboom we have these scripts neither are needed for ci but which if either should be kept for contributors maintainers who need to run the new configlet locally bin fetch configlet bin fetch canonical data syncer
1
38,616
8,517,113,048
IssuesEvent
2018-11-01 06:28:56
virtual-labs/colloid-and-surface-chemistry-iiith
https://api.github.com/repos/virtual-labs/colloid-and-surface-chemistry-iiith
opened
Missing vendor-prefixed CSS gradients for Webkit (Safari 5+, Chrome), Opera 11.1+.
2018-Open static-code-analysis
CSS gradients in a cross-browser way requires using many different vendor-prefixed versions. There are currently five different vendor-prefixed versions of CSS gradient: -ms-linear-gradient and -ms-radial-gradient for Internet Explorer 10+ -moz-linear-gradient and -moz-radial-gradient for Firefox 3.6+ -o-linear-gradient and -o-radial-gradient for Opera 11.10+ -webkit-linear-gradient and -webkit-radial-gradient for Safari 5+ and Chrome -webkit-gradient for Safari 4+ and Chrome (aka "Old WebKit") Meaning a simple two-color gradient that works across all browsers must look like this: ``` background: -moz-linear-gradient(...); /* FF3.6+ */ background: -webkit-gradient(...); /* Chrome,Safari4+ */ background: -webkit-linear-gradient(...); /* Chrome10+,Safari5.1+ */ background: -o-linear-gradient(...); /* Opera 11.10+ */ background: -ms-linear-gradient(...); /* IE10+ */ ``` It's easy to forget one or more gradient definitions with all of the various vendor prefix gradients available. **Please refer to the following link to fix similar issues.** https://app.codacy.com/app/BSravanthi/colloid-and-surface-chemistry-iiith/issues?&filters=W3siaWQiOiJMYW5ndWFnZSIsInZhbHVlcyI6W251bGxdfSx7ImlkIjoiQ2F0ZWdvcnkiLCJ2YWx1ZXMiOlsiQ29tcGF0aWJpbGl0eSJdfSx7ImlkIjoiTGV2ZWwiLCJ2YWx1ZXMiOltudWxsXX0seyJpZCI6IlBhdHRlcm4iLCJ2YWx1ZXMiOltudWxsXX1d
1.0
Missing vendor-prefixed CSS gradients for Webkit (Safari 5+, Chrome), Opera 11.1+. - CSS gradients in a cross-browser way requires using many different vendor-prefixed versions. There are currently five different vendor-prefixed versions of CSS gradient: -ms-linear-gradient and -ms-radial-gradient for Internet Explorer 10+ -moz-linear-gradient and -moz-radial-gradient for Firefox 3.6+ -o-linear-gradient and -o-radial-gradient for Opera 11.10+ -webkit-linear-gradient and -webkit-radial-gradient for Safari 5+ and Chrome -webkit-gradient for Safari 4+ and Chrome (aka "Old WebKit") Meaning a simple two-color gradient that works across all browsers must look like this: ``` background: -moz-linear-gradient(...); /* FF3.6+ */ background: -webkit-gradient(...); /* Chrome,Safari4+ */ background: -webkit-linear-gradient(...); /* Chrome10+,Safari5.1+ */ background: -o-linear-gradient(...); /* Opera 11.10+ */ background: -ms-linear-gradient(...); /* IE10+ */ ``` It's easy to forget one or more gradient definitions with all of the various vendor prefix gradients available. **Please refer to the following link to fix similar issues.** https://app.codacy.com/app/BSravanthi/colloid-and-surface-chemistry-iiith/issues?&filters=W3siaWQiOiJMYW5ndWFnZSIsInZhbHVlcyI6W251bGxdfSx7ImlkIjoiQ2F0ZWdvcnkiLCJ2YWx1ZXMiOlsiQ29tcGF0aWJpbGl0eSJdfSx7ImlkIjoiTGV2ZWwiLCJ2YWx1ZXMiOltudWxsXX0seyJpZCI6IlBhdHRlcm4iLCJ2YWx1ZXMiOltudWxsXX1d
non_main
missing vendor prefixed css gradients for webkit safari chrome opera css gradients in a cross browser way requires using many different vendor prefixed versions there are currently five different vendor prefixed versions of css gradient ms linear gradient and ms radial gradient for internet explorer moz linear gradient and moz radial gradient for firefox o linear gradient and o radial gradient for opera webkit linear gradient and webkit radial gradient for safari and chrome webkit gradient for safari and chrome aka old webkit meaning a simple two color gradient that works across all browsers must look like this background moz linear gradient background webkit gradient chrome background webkit linear gradient background o linear gradient opera background ms linear gradient it s easy to forget one or more gradient definitions with all of the various vendor prefix gradients available please refer to the following link to fix similar issues
0
24,754
17,692,928,112
IssuesEvent
2021-08-24 12:20:15
tarantool/tarantool
https://api.github.com/repos/tarantool/tarantool
closed
new release policy: adjust download.tarantool.org infrastructure
infrastructure teamQ
The main discussion about the new release policy is [here](https://github.com/tarantool/tarantool/discussions/6182). Since we [plan](https://github.com/tarantool/tarantool/issues/6185) to add repositories with a bit different layout (`series-2`, `pre-release`), we should adjust download.tarantool.org redirections and ensure that everything works good. NB: Don't forget about source tarballs.
1.0
new release policy: adjust download.tarantool.org infrastructure - The main discussion about the new release policy is [here](https://github.com/tarantool/tarantool/discussions/6182). Since we [plan](https://github.com/tarantool/tarantool/issues/6185) to add repositories with a bit different layout (`series-2`, `pre-release`), we should adjust download.tarantool.org redirections and ensure that everything works good. NB: Don't forget about source tarballs.
non_main
new release policy adjust download tarantool org infrastructure the main discussion about the new release policy is since we to add repositories with a bit different layout series pre release we should adjust download tarantool org redirections and ensure that everything works good nb don t forget about source tarballs
0
635,851
20,510,795,378
IssuesEvent
2022-03-01 06:08:04
AY2122S2-CS2103T-T13-1/tp
https://api.github.com/repos/AY2122S2-CS2103T-T13-1/tp
opened
As a pet daycare owner I can get a list of pets which will be staying overnight in the daycare
type.Story priority.Medium
... so that I can arrange the necessary manpower support required.
1.0
As a pet daycare owner I can get a list of pets which will be staying overnight in the daycare - ... so that I can arrange the necessary manpower support required.
non_main
as a pet daycare owner i can get a list of pets which will be staying overnight in the daycare so that i can arrange the necessary manpower support required
0
157,506
13,691,125,176
IssuesEvent
2020-09-30 15:10:41
codesankalp/dsalgo
https://api.github.com/repos/codesankalp/dsalgo
opened
PR Protocol
documentation enhancement
required submission protocols for pull request. type: google doc, word , ppt must include: format for pull request, test,quality assurance and other neccessary checks
1.0
PR Protocol - required submission protocols for pull request. type: google doc, word , ppt must include: format for pull request, test,quality assurance and other neccessary checks
non_main
pr protocol required submission protocols for pull request type google doc word ppt must include format for pull request test quality assurance and other neccessary checks
0
299,678
22,617,917,411
IssuesEvent
2022-06-30 01:22:41
sonr-io/sonr
https://api.github.com/repos/sonr-io/sonr
closed
Integrate Stripe for Client Side Payment
documentation
We need to do some R&D and handle Fiat to ATOM to SNR conversion. Some questions that need to be answered: * How do Coins operate in the Cosmos Network? * Do we need to have an ATOM -> SNR conversion? * Can we have a direct Fiat -> SNR NFT purchase? https://dashboard.moonpay.com/getting\_started ┆Issue is synchronized with this [Asana task](https://app.asana.com/0/1202528620406598/1202529507214139) by [Unito](https://www.unito.io)
1.0
Integrate Stripe for Client Side Payment - We need to do some R&D and handle Fiat to ATOM to SNR conversion. Some questions that need to be answered: * How do Coins operate in the Cosmos Network? * Do we need to have an ATOM -> SNR conversion? * Can we have a direct Fiat -> SNR NFT purchase? https://dashboard.moonpay.com/getting\_started ┆Issue is synchronized with this [Asana task](https://app.asana.com/0/1202528620406598/1202529507214139) by [Unito](https://www.unito.io)
non_main
integrate stripe for client side payment we need to do some r d and handle fiat to atom to snr conversion some questions that need to be answered how do coins operate in the cosmos network do we need to have an atom snr conversion can we have a direct fiat snr nft purchase ┆issue is synchronized with this by
0
528,068
15,359,403,350
IssuesEvent
2021-03-01 15:49:06
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.pornhubpremium.com - video or audio doesn't play
browser-firefox-mobile browser-firefox-reality engine-gecko priority-normal type-webvr
<!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/67701 --> **URL**: https://www.pornhubpremium.com/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 7.1.1 **Tested Another Browser**: Yes Other **Problem type**: Video or audio doesn't play **Description**: There is no video **Steps to Reproduce**: oculus questfirefox reality when you try to watch something in webr its just a black screen <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❀️_
1.0
www.pornhubpremium.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.1; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/67701 --> **URL**: https://www.pornhubpremium.com/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android 7.1.1 **Tested Another Browser**: Yes Other **Problem type**: Video or audio doesn't play **Description**: There is no video **Steps to Reproduce**: oculus questfirefox reality when you try to watch something in webr its just a black screen <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❀️_
non_main
video or audio doesn t play url browser version firefox mobile operating system android tested another browser yes other problem type video or audio doesn t play description there is no video steps to reproduce oculus questfirefox reality when you try to watch something in webr its just a black screen browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❀️
0
390,102
11,524,808,568
IssuesEvent
2020-02-15 03:06:23
HW-PlayersPatch/Development
https://api.github.com/repos/HW-PlayersPatch/Development
opened
Salvage Crash Fix
Priority3: Low Status2: Research Needed Type1: Feature Type2: Bug
Build 11: "When salvage corvettes in a formation are given multiple different salvage targets, they often crash the game. To prevent this crash, salvage corvettes will no longer be able to join formations/strikegroups." https://github.com/HW-PlayersPatch/Development/commit/3b66a70e9b40f5a3b67c48d6b99f339d8561afba Try to make this an option.
1.0
Salvage Crash Fix - Build 11: "When salvage corvettes in a formation are given multiple different salvage targets, they often crash the game. To prevent this crash, salvage corvettes will no longer be able to join formations/strikegroups." https://github.com/HW-PlayersPatch/Development/commit/3b66a70e9b40f5a3b67c48d6b99f339d8561afba Try to make this an option.
non_main
salvage crash fix build when salvage corvettes in a formation are given multiple different salvage targets they often crash the game to prevent this crash salvage corvettes will no longer be able to join formations strikegroups try to make this an option
0
4,832
24,910,996,393
IssuesEvent
2022-10-29 21:24:07
deislabs/spiderlightning
https://api.github.com/repos/deislabs/spiderlightning
closed
Bin releases of `slight`
🚧 maintainer issue
We should start building and publishing releases for `slight`, so that folks don't need to build bins themselves. We need to add a release action that is triggered by a semantic version tag that would build `slight` bins targeting the major OS/Arch combinations.
True
Bin releases of `slight` - We should start building and publishing releases for `slight`, so that folks don't need to build bins themselves. We need to add a release action that is triggered by a semantic version tag that would build `slight` bins targeting the major OS/Arch combinations.
main
bin releases of slight we should start building and publishing releases for slight so that folks don t need to build bins themselves we need to add a release action that is triggered by a semantic version tag that would build slight bins targeting the major os arch combinations
1
88,021
11,018,306,658
IssuesEvent
2019-12-05 10:15:10
statsmodels/statsmodels
https://api.github.com/repos/statsmodels/statsmodels
closed
use httplib2 for caching downloaded data files ?
design
Skipper recommended https://groups.google.com/group/pystatsmodels/browse_thread/thread/5bfe1e45b3336765?hl=en http://pypi.python.org/pypi/httplib2/ 80,000 downloads looks pretty good
1.0
use httplib2 for caching downloaded data files ? - Skipper recommended https://groups.google.com/group/pystatsmodels/browse_thread/thread/5bfe1e45b3336765?hl=en http://pypi.python.org/pypi/httplib2/ 80,000 downloads looks pretty good
non_main
use for caching downloaded data files skipper recommended downloads looks pretty good
0
166,141
12,891,859,896
IssuesEvent
2020-07-13 18:30:47
ReactiveX/RxJava
https://api.github.com/repos/ReactiveX/RxJava
closed
3.x: Flaky GroupBy test
3.x PR welcome Test-Failures good first issue
https://github.com/ReactiveX/RxJava/blob/98acac218cdb04d279b5ac49bb1afc65bc6ec4fe/src/test/java/io/reactivex/rxjava3/internal/operators/flowable/FlowableGroupByTest.java#L2668 ``` io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest > issue6974Part2Case1NoEvict FAILED java.lang.AssertionError: Error(s) present: [io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred. ] (latch = 0, values = 15551, errors = 1, completions = 0) at io.reactivex.rxjava3.observers.BaseTestConsumer.fail(BaseTestConsumer.java:125) at io.reactivex.rxjava3.observers.BaseTestConsumer.assertNoErrors(BaseTestConsumer.java:212) at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974RunPart2NoEvict(FlowableGroupByTest.java:2681) at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974Part2Case1NoEvict(FlowableGroupByTest.java:2693) Caused by: io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred. Caused by: io.reactivex.rxjava3.exceptions.MissingBackpressureException: Unable to emit a new group (#71) due to lack of requests. Please make sure the downstream can always accept a new group as well as each group is consumed in order for the whole operator to be able to proceed. ``` Error is an allowed outcome here.
1.0
3.x: Flaky GroupBy test - https://github.com/ReactiveX/RxJava/blob/98acac218cdb04d279b5ac49bb1afc65bc6ec4fe/src/test/java/io/reactivex/rxjava3/internal/operators/flowable/FlowableGroupByTest.java#L2668 ``` io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest > issue6974Part2Case1NoEvict FAILED java.lang.AssertionError: Error(s) present: [io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred. ] (latch = 0, values = 15551, errors = 1, completions = 0) at io.reactivex.rxjava3.observers.BaseTestConsumer.fail(BaseTestConsumer.java:125) at io.reactivex.rxjava3.observers.BaseTestConsumer.assertNoErrors(BaseTestConsumer.java:212) at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974RunPart2NoEvict(FlowableGroupByTest.java:2681) at io.reactivex.rxjava3.internal.operators.flowable.FlowableGroupByTest.issue6974Part2Case1NoEvict(FlowableGroupByTest.java:2693) Caused by: io.reactivex.rxjava3.exceptions.CompositeException: 1 exceptions occurred. Caused by: io.reactivex.rxjava3.exceptions.MissingBackpressureException: Unable to emit a new group (#71) due to lack of requests. Please make sure the downstream can always accept a new group as well as each group is consumed in order for the whole operator to be able to proceed. ``` Error is an allowed outcome here.
non_main
x flaky groupby test io reactivex internal operators flowable flowablegroupbytest failed java lang assertionerror error s present latch values errors completions at io reactivex observers basetestconsumer fail basetestconsumer java at io reactivex observers basetestconsumer assertnoerrors basetestconsumer java at io reactivex internal operators flowable flowablegroupbytest flowablegroupbytest java at io reactivex internal operators flowable flowablegroupbytest flowablegroupbytest java caused by io reactivex exceptions compositeexception exceptions occurred caused by io reactivex exceptions missingbackpressureexception unable to emit a new group due to lack of requests please make sure the downstream can always accept a new group as well as each group is consumed in order for the whole operator to be able to proceed error is an allowed outcome here
0
2,679
9,219,081,694
IssuesEvent
2019-03-11 14:44:39
precice/precice
https://api.github.com/repos/precice/precice
closed
Integrate a testing coverage tool
maintainability
Since we have quite some tests in preCICE, it would be useful to also have a testing coverage tool to show us where tests are missing etc. It would also be a nice code quality metric which we can track over time. A common tool that several projects use is [code coverage](https://codecov.io/) (e.g. see [Spack](https://codecov.io/gh/spack/spack)). Any other ideas/suggestions?
True
Integrate a testing coverage tool - Since we have quite some tests in preCICE, it would be useful to also have a testing coverage tool to show us where tests are missing etc. It would also be a nice code quality metric which we can track over time. A common tool that several projects use is [code coverage](https://codecov.io/) (e.g. see [Spack](https://codecov.io/gh/spack/spack)). Any other ideas/suggestions?
main
integrate a testing coverage tool since we have quite some tests in precice it would be useful to also have a testing coverage tool to show us where tests are missing etc it would also be a nice code quality metric which we can track over time a common tool that several projects use is e g see any other ideas suggestions
1
5,526
27,632,268,436
IssuesEvent
2023-03-10 11:49:27
jesus2099/konami-command
https://api.github.com/repos/jesus2099/konami-command
opened
Retrieve available localised texts from MBS page
ninja mb_POWER-VOTE minor maintainability
Instead of hardcoding _Yes, No, Abstain, No vote, Approve,_ etc., in several languages, fetch these text from current page.
True
Retrieve available localised texts from MBS page - Instead of hardcoding _Yes, No, Abstain, No vote, Approve,_ etc., in several languages, fetch these text from current page.
main
retrieve available localised texts from mbs page instead of hardcoding yes no abstain no vote approve etc in several languages fetch these text from current page
1
26,387
12,404,876,892
IssuesEvent
2020-05-21 16:16:37
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
opened
TIA Customer Case Management: Customer Permit Details Page
Need: 1-Must Have Product: TIA Module Service: Apps Status: Done Type: Feature Workgroup: TDSD imported-from-csv
As a TIA Customer I'd like a case details page that allows me to see basic information about my case, especially the current status.
1.0
TIA Customer Case Management: Customer Permit Details Page - As a TIA Customer I'd like a case details page that allows me to see basic information about my case, especially the current status.
non_main
tia customer case management customer permit details page as a tia customer i d like a case details page that allows me to see basic information about my case especially the current status
0
493,593
14,235,366,822
IssuesEvent
2020-11-18 14:44:50
telerik/kendo-ui-core
https://api.github.com/repos/telerik/kendo-ui-core
opened
Scheduler's navigate event does not trigger on mobile device
Bug C: Scheduler Kendo2 Priority 5 SEV: Medium Touch Devices
### Bug report The navigate event does not trigger when changing the view on a mobile device. The event triggers back in version 2019.1.220. ### Reproduction of the problem 1. Open [this example](https://dojo.telerik.com/OSOCOgEm) on a mobile device. 2. Change the View. ### Current behavior The navigate event does not trigger when changing the view. ### Expected/desired behavior The navigate event should trigger. ### Environment * **Kendo UI version:** 2020.3.1118 * **Browser:** [all]
1.0
Scheduler's navigate event does not trigger on mobile device - ### Bug report The navigate event does not trigger when changing the view on a mobile device. The event triggers back in version 2019.1.220. ### Reproduction of the problem 1. Open [this example](https://dojo.telerik.com/OSOCOgEm) on a mobile device. 2. Change the View. ### Current behavior The navigate event does not trigger when changing the view. ### Expected/desired behavior The navigate event should trigger. ### Environment * **Kendo UI version:** 2020.3.1118 * **Browser:** [all]
non_main
scheduler s navigate event does not trigger on mobile device bug report the navigate event does not trigger when changing the view on a mobile device the event triggers back in version reproduction of the problem open on a mobile device change the view current behavior the navigate event does not trigger when changing the view expected desired behavior the navigate event should trigger environment kendo ui version browser
0
30,432
11,825,715,288
IssuesEvent
2020-03-21 14:21:50
stefanfreitag/s3_yum_repository_slides
https://api.github.com/repos/stefanfreitag/s3_yum_repository_slides
opened
CVE-2015-9251 (Medium) detected in jquery-1.7.2.min.js
security vulnerability
## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/s3_yum_repository_slides/node_modules/js-base64/test/index.html</p> <p>Path to vulnerable library: /s3_yum_repository_slides/node_modules/js-base64/test/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/s3_yum_repository_slides/commit/a3e7d6e421e87c01267600c253daad2de918d386">a3e7d6e421e87c01267600c253daad2de918d386</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9251 (Medium) detected in jquery-1.7.2.min.js - ## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.2.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.2/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/s3_yum_repository_slides/node_modules/js-base64/test/index.html</p> <p>Path to vulnerable library: /s3_yum_repository_slides/node_modules/js-base64/test/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.7.2.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/stefanfreitag/s3_yum_repository_slides/commit/a3e7d6e421e87c01267600c253daad2de918d386">a3e7d6e421e87c01267600c253daad2de918d386</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm yum repository slides node modules js test index html path to vulnerable library yum repository slides node modules js test index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
182,572
14,917,998,820
IssuesEvent
2021-01-22 20:51:48
ZupIT/ritchie-cli
https://api.github.com/repos/ZupIT/ritchie-cli
closed
Add Bitbucket to provider list for rit add repo
:books: documentation :hammer: improvement :heavy_check_mark: refined :sparkles: feature Hacktoberfest
## What would you like to see - I would like to be able to import a bitbucket repo on Ritchie through the `rit add repo` command. ## Why is it needed - Bitbucket is not supported yet, only Github and Gitlab.
1.0
Add Bitbucket to provider list for rit add repo - ## What would you like to see - I would like to be able to import a bitbucket repo on Ritchie through the `rit add repo` command. ## Why is it needed - Bitbucket is not supported yet, only Github and Gitlab.
non_main
add bitbucket to provider list for rit add repo what would you like to see i would like to be able to import a bitbucket repo on ritchie through the rit add repo command why is it needed bitbucket is not supported yet only github and gitlab
0
99,438
8,700,535,103
IssuesEvent
2018-12-05 09:02:03
SME-Issues/issues
https://api.github.com/repos/SME-Issues/issues
closed
General Comprehension None Tests - 04/12/2018 - 5004
NLP Api pulse_tests
**General Comprehension None Tests** - Total: 23 - Passed: 17 - **Pass: 17 (77%)** - Not Understood: 0 - Error (not understood): 1 - Failed but Understood: 5 (23%)
1.0
General Comprehension None Tests - 04/12/2018 - 5004 - **General Comprehension None Tests** - Total: 23 - Passed: 17 - **Pass: 17 (77%)** - Not Understood: 0 - Error (not understood): 1 - Failed but Understood: 5 (23%)
non_main
general comprehension none tests general comprehension none tests total passed pass not understood error not understood failed but understood
0
1,395
6,025,335,501
IssuesEvent
2017-06-08 08:25:53
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_iis_webapplication: slashes in physical path are not converted to windows style
affects_2.3 bug_report waiting_on_maintainer windows
##### ISSUE TYPE Bug Report ##### COMPONENT NAME win_iis_webapplication module ##### ANSIBLE VERSION N/A ##### SUMMARY win_iis_website converts them. Example: vars somewhere: ``` path: c:/sites/https/develop ``` playbook: ``` - name: create website win_iis_website: name: "{{ name }}" state: started application_pool: "{{ appdomain }}" physical_path: "{{ path }}/web" - name: create rest virtual folder win_iis_webapplication: name: rest state: present site: "{{ name }}" physical_path: "{{ path }}/api" application_pool: "{{ appdomain }}" ``` Website will work ok, webapplication will always fail with 404.
True
win_iis_webapplication: slashes in physical path are not converted to windows style - ##### ISSUE TYPE Bug Report ##### COMPONENT NAME win_iis_webapplication module ##### ANSIBLE VERSION N/A ##### SUMMARY win_iis_website converts them. Example: vars somewhere: ``` path: c:/sites/https/develop ``` playbook: ``` - name: create website win_iis_website: name: "{{ name }}" state: started application_pool: "{{ appdomain }}" physical_path: "{{ path }}/web" - name: create rest virtual folder win_iis_webapplication: name: rest state: present site: "{{ name }}" physical_path: "{{ path }}/api" application_pool: "{{ appdomain }}" ``` Website will work ok, webapplication will always fail with 404.
main
win iis webapplication slashes in physical path are not converted to windows style issue type bug report component name win iis webapplication module ansible version n a summary win iis website converts them example vars somewhere path c sites https develop playbook name create website win iis website name name state started application pool appdomain physical path path web name create rest virtual folder win iis webapplication name rest state present site name physical path path api application pool appdomain website will work ok webapplication will always fail with
1
247,133
18,857,342,416
IssuesEvent
2021-11-12 08:29:05
chongjunwei/pe
https://api.github.com/repos/chongjunwei/pe
opened
Command message ambiguous
severity.Low type.DocumentationBug
![Screenshot 2021-11-12 at 4.25.11 PM.png](https://raw.githubusercontent.com/chongjunwei/pe/main/files/76726e17-ce17-4781-ae92-ad32ea89681b.png) Instead of "show all modules", perhaps "show students from all the modules" or something similar would be clearer. The current wording is slightly ambiguous and may be confused with the existing "lsmod" command, which shows a list of all the modules. <!--session: 1636703307298-1218c5c4-1eba-4ef0-95db-9ff00e3700d4--> <!--Version: Web v3.4.1-->
1.0
Command message ambiguous - ![Screenshot 2021-11-12 at 4.25.11 PM.png](https://raw.githubusercontent.com/chongjunwei/pe/main/files/76726e17-ce17-4781-ae92-ad32ea89681b.png) Instead of "show all modules", perhaps "show students from all the modules" or something similar would be clearer. The current wording is slightly ambiguous and may be confused with the existing "lsmod" command, which shows a list of all the modules. <!--session: 1636703307298-1218c5c4-1eba-4ef0-95db-9ff00e3700d4--> <!--Version: Web v3.4.1-->
non_main
command message ambiguous instead of show all modules perhaps show students from all the modules or something similar would be clearer the current wording is slightly ambiguous and may be confused with the existing lsmod command which shows a list of all the modules
0
9,454
8,639,227,120
IssuesEvent
2018-11-23 17:45:18
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Add information about "Diagnose and solve problems" experience
app-service-web/svc assigned-to-author doc-enhancement triaged
We have a great new experience, but not covered in a documentation --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a14de731-a265-1737-4580-67ca6dede020 * Version Independent ID: a275bb18-dedc-d41c-3541-224b42487b52 * Content: [Slow web app performance in App Service](https://docs.microsoft.com/en-us/azure/app-service/app-service-web-troubleshoot-performance-degradation) * Content Source: [articles/app-service/app-service-web-troubleshoot-performance-degradation.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-web-troubleshoot-performance-degradation.md) * Service: **app-service-web** * GitHub Login: @cephalin * Microsoft Alias: **cephalin**
1.0
Add information about "Diagnose and solve problems" experience - We have a great new experience, but not covered in a documentation --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a14de731-a265-1737-4580-67ca6dede020 * Version Independent ID: a275bb18-dedc-d41c-3541-224b42487b52 * Content: [Slow web app performance in App Service](https://docs.microsoft.com/en-us/azure/app-service/app-service-web-troubleshoot-performance-degradation) * Content Source: [articles/app-service/app-service-web-troubleshoot-performance-degradation.md](https://github.com/Microsoft/azure-docs/blob/master/articles/app-service/app-service-web-troubleshoot-performance-degradation.md) * Service: **app-service-web** * GitHub Login: @cephalin * Microsoft Alias: **cephalin**
non_main
add information about diagnose and solve problems experience we have a great new experience but not covered in a documentation document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id dedc content content source service app service web github login cephalin microsoft alias cephalin
0
4,869
25,020,305,365
IssuesEvent
2022-11-03 23:27:56
aws/serverless-application-model
https://api.github.com/repos/aws/serverless-application-model
closed
"Auth" property do not work with "AWS::Include" DefinitionBody
area/resource/api area/intrinsics maintainer/need-response
**Description:** We need to use some cloudformation functions in our swagger. So the "AWS::Include" transform is used in the DefnitionBody. I tried to move the authorizer declaration out of the swagger using the new "Auth" property. But I received an error of "Unable to add Auth configuration because 'DefinitionBody' does not contain a valid Swagger". Also, I notice the authorizer caching TTL is not an attribute in the "Auth" property? ** Template snippet:** ``` Type: AWS::Serverless::Api Properties: StageName: !Ref "Environment" EndpointConfiguration: REGIONAL Auth: Authorizers: RequestAuth: FunctionPayloadType: REQUEST FunctionArn: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerArn"] FunctionInvokeRole: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerCredentials"] Identity: Headers: - Authorization Context: - httpMethod - path ReauthorizeEvery: 0 # OPTIONAL; Service Default: 300 DefinitionBody: 'Fn::Transform': Name: 'AWS::Include' Parameters: Location: s3://accolade-api-swaggers-test-577121982548/users.yml ```
True
"Auth" property do not work with "AWS::Include" DefinitionBody - **Description:** We need to use some cloudformation functions in our swagger. So the "AWS::Include" transform is used in the DefnitionBody. I tried to move the authorizer declaration out of the swagger using the new "Auth" property. But I received an error of "Unable to add Auth configuration because 'DefinitionBody' does not contain a valid Swagger". Also, I notice the authorizer caching TTL is not an attribute in the "Auth" property? ** Template snippet:** ``` Type: AWS::Serverless::Api Properties: StageName: !Ref "Environment" EndpointConfiguration: REGIONAL Auth: Authorizers: RequestAuth: FunctionPayloadType: REQUEST FunctionArn: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerArn"] FunctionInvokeRole: !FindInMap [EnvironmentDependentParams, !Ref "Environment", "AuthorizerCredentials"] Identity: Headers: - Authorization Context: - httpMethod - path ReauthorizeEvery: 0 # OPTIONAL; Service Default: 300 DefinitionBody: 'Fn::Transform': Name: 'AWS::Include' Parameters: Location: s3://accolade-api-swaggers-test-577121982548/users.yml ```
main
auth property do not work with aws include definitionbody description we need to use some cloudformation functions in our swagger so the aws include transform is used in the defnitionbody i tried to move the authorizer declaration out of the swagger using the new auth property but i received an error of unable to add auth configuration because definitionbody does not contain a valid swagger also i notice the authorizer caching ttl is not an attribute in the auth property template snippet type aws serverless api properties stagename ref environment endpointconfiguration regional auth authorizers requestauth functionpayloadtype request functionarn findinmap functioninvokerole findinmap identity headers authorization context httpmethod path reauthorizeevery optional service default definitionbody fn transform name aws include parameters location accolade api swaggers test users yml
1
255,452
21,926,005,845
IssuesEvent
2022-05-23 04:18:28
stores-cedcommerce/Sungin-Internal--Feb21st-2022
https://api.github.com/repos/stores-cedcommerce/Sungin-Internal--Feb21st-2022
closed
Password Char length verification error issue.
Account pages Desktop Functional / bug Ready to test
Bug - Password char length verification showing different on create account page and on reset password page. Exp - Password char length should be visible same in validation on both pages. Ref Link - https://drive.google.com/file/d/1wQit2JLx4UtunlBWVkgX11DQce7qmysV/view
1.0
Password Char length verification error issue. - Bug - Password char length verification showing different on create account page and on reset password page. Exp - Password char length should be visible same in validation on both pages. Ref Link - https://drive.google.com/file/d/1wQit2JLx4UtunlBWVkgX11DQce7qmysV/view
non_main
password char length verification error issue bug password char length verification showing different on create account page and on reset password page exp password char length should be visible same in validation on both pages ref link
0
91,529
10,722,606,139
IssuesEvent
2019-10-27 13:16:50
riot/riot
https://api.github.com/repos/riot/riot
closed
[Question] How should I migrate mixins to riot 4?
documentation update required good for beginner
5. How would you tag this issue? - [x ] Question I noticed mixins are absent from riot 4 documentation and from what I can tell the code, too. Are there any undocumented features, samples, or ideas for users migrating mixins from riot 3 to riot 4?
1.0
[Question] How should I migrate mixins to riot 4? - 5. How would you tag this issue? - [x ] Question I noticed mixins are absent from riot 4 documentation and from what I can tell the code, too. Are there any undocumented features, samples, or ideas for users migrating mixins from riot 3 to riot 4?
non_main
how should i migrate mixins to riot how would you tag this issue question i noticed mixins are absent from riot documentation and from what i can tell the code too are there any undocumented features samples or ideas for users migrating mixins from riot to riot
0
854
4,513,365,927
IssuesEvent
2016-09-04 07:56:06
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
ec2_eni is not idempotent
aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - ec2_eni ##### ANSIBLE VERSION - devel ##### SUMMARY For the ec2_eni module to be idempotent on the eni creation, it should create the interface on the first execution and just return the interface with changed=False on the subsequent executions. The current behavior is to return an error on the subsequent executions. ##### STEPS TO REPRODUCE ``` $ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost localhost | SUCCESS => { "changed": true, "interface": { "description": "", "groups": { "sg-6b98a613": "default" }, "id": "eni-51zzz311", "mac_address": "0a:ff:2b:b7:0b:91", "owner_id": "...", "private_ip_address": "10.137.0.10", "source_dest_check": true, "status": "pending", "subnet_id": "subnet-1b2c696d", "vpc_id": "vpc-3ee6ff5a" } } $ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost localhost | FAILED! => { "changed": false, "failed": true, "msg": "The specified address is already in use." } ``` ##### EXPECTED RESULTS Second execution would return `changed:false` along with the `interface` data structure
True
ec2_eni is not idempotent - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - ec2_eni ##### ANSIBLE VERSION - devel ##### SUMMARY For the ec2_eni module to be idempotent on the eni creation, it should create the interface on the first execution and just return the interface with changed=False on the subsequent executions. The current behavior is to return an error on the subsequent executions. ##### STEPS TO REPRODUCE ``` $ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost localhost | SUCCESS => { "changed": true, "interface": { "description": "", "groups": { "sg-6b98a613": "default" }, "id": "eni-51zzz311", "mac_address": "0a:ff:2b:b7:0b:91", "owner_id": "...", "private_ip_address": "10.137.0.10", "source_dest_check": true, "status": "pending", "subnet_id": "subnet-1b2c696d", "vpc_id": "vpc-3ee6ff5a" } } $ ansible -i localhost, -c local -m ec2_eni -a 'private_ip_address=10.137.0.10 subnet_id=subnet-zzzzz state=present region=us-east-1' localhost localhost | FAILED! => { "changed": false, "failed": true, "msg": "The specified address is already in use." } ``` ##### EXPECTED RESULTS Second execution would return `changed:false` along with the `interface` data structure
main
eni is not idempotent issue type bug report component name eni ansible version devel summary for the eni module to be idempotent on the eni creation it should create the interface on the first execution and just return the interface with changed false on the subsequent executions the current behavior is to return an error on the subsequent executions steps to reproduce ansible i localhost c local m eni a private ip address subnet id subnet zzzzz state present region us east localhost localhost success changed true interface description groups sg default id eni mac address ff owner id private ip address source dest check true status pending subnet id subnet vpc id vpc ansible i localhost c local m eni a private ip address subnet id subnet zzzzz state present region us east localhost localhost failed changed false failed true msg the specified address is already in use expected results second execution would return changed false along with the interface data structure
1
166,569
12,962,315,342
IssuesEvent
2020-07-20 16:57:44
longhorn/longhorn
https://api.github.com/repos/longhorn/longhorn
closed
Listing backups may fail during backup deletion
area/engine area/test bug reproduce/rare
``` clients = {'longhorn-staging-tests-01': <longhorn.Client object at 0x7f0e589e2cd0>, 'longhorn-staging-tests-02': <longhorn.Client object at 0x7f0e58491e10>, 'longhorn-staging-tests-03': <longhorn.Client object at 0x7f0e589e2f50>} volume_name = 'longhorn-testvol-b4pqea' @pytest.mark.coretest # NOQA def test_backup(clients, volume_name): # NOQA > backup_test(clients, volume_name, SIZE) test_basic.py:396: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_basic.py:430: in backup_test backupstore_test(client, lht_hostId, volume_name, size) test_basic.py:538: in backupstore_test backups = bv.backupList().data longhorn.py:248: in cb *args, **kw) longhorn.py:442: in action return self._post_and_retry(url, *args, **kw) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <longhorn.Client object at 0x7f0e589e2f50> url = 'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea?action=backupList' args = (), kw = {}, retries = 3, last_error = None, i = 0 e = ApiError(ApiError(...), '500 : error listing backups for volume \'longhorn-tes... exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}') def _post_and_retry(self, url, *args, **kw): retries = kw.get('retries', 3) last_error = None for i in range(retries): try: return self._post(url, data=self._to_dict(*args, **kw)) except ApiError as e: if e.error.code == 409: last_error = e time.sleep(.1) else: > raise e E ApiError: (ApiError(...), '500 : error listing backups for volume \'longhorn-testvol-b4pqea\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\n404 \n\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\n AcceptRanges: \\"bytes\\",\\n Body: <nil>,\\n ContentLength: 511,\\n ContentType: \\"application/xml\\",\\n Metadata: {\\n\\n }\\n}" pkg=s3\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n"\n, error exit status 1\n{\'status\': 500, \'code\': 500, \'links\': {\'self\': u\'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea\'}, \'self\': <function cb at 0x7f0e58bd8230>, \'detail\': u\'\', \'actions\': {}, \'message\': u\'error listing backups for volume \\\'longhorn-testvol-b4pqea\\\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n\\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\\\n AcceptRanges: \\\\"bytes\\\\",\\\\n Body: <nil>,\\\\n ContentLength: 511,\\\\n ContentType: \\\\"application/xml\\\\",\\\\n Metadata: {\\\\n\\\\n }\\\\n}" pkg=s3\\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}') ``` When Longhorn tried to list the backup which was being deleted, this error may occur.
1.0
Listing backups may fail during backup deletion - ``` clients = {'longhorn-staging-tests-01': <longhorn.Client object at 0x7f0e589e2cd0>, 'longhorn-staging-tests-02': <longhorn.Client object at 0x7f0e58491e10>, 'longhorn-staging-tests-03': <longhorn.Client object at 0x7f0e589e2f50>} volume_name = 'longhorn-testvol-b4pqea' @pytest.mark.coretest # NOQA def test_backup(clients, volume_name): # NOQA > backup_test(clients, volume_name, SIZE) test_basic.py:396: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_basic.py:430: in backup_test backupstore_test(client, lht_hostId, volume_name, size) test_basic.py:538: in backupstore_test backups = bv.backupList().data longhorn.py:248: in cb *args, **kw) longhorn.py:442: in action return self._post_and_retry(url, *args, **kw) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <longhorn.Client object at 0x7f0e589e2f50> url = 'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea?action=backupList' args = (), kw = {}, retries = 3, last_error = None, i = 0 e = ApiError(ApiError(...), '500 : error listing backups for volume \'longhorn-tes... exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}') def _post_and_retry(self, url, *args, **kw): retries = kw.get('retries', 3) last_error = None for i in range(retries): try: return self._post(url, data=self._to_dict(*args, **kw)) except ApiError as e: if e.error.code == 409: last_error = e time.sleep(.1) else: > raise e E ApiError: (ApiError(...), '500 : error listing backups for volume \'longhorn-testvol-b4pqea\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\n404 \n\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\n AcceptRanges: \\"bytes\\",\\n Body: <nil>,\\n ContentLength: 511,\\n ContentType: \\"application/xml\\",\\n Metadata: {\\n\\n }\\n}" pkg=s3\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n"\n, error exit status 1\n{\'status\': 500, \'code\': 500, \'links\': {\'self\': u\'http://10.42.1.5:9500/v1/backupvolumes/longhorn-testvol-b4pqea\'}, \'self\': <function cb at 0x7f0e58bd8230>, \'detail\': u\'\', \'actions\': {}, \'message\': u\'error listing backups for volume \\\'longhorn-testvol-b4pqea\\\': error listing backups: Failed to execute: /var/lib/rancher/longhorn/engine-binaries/longhornio-longhorn-engine-staging/longhorn [backup ls --volume longhorn-testvol-b4pqea s3://backupbucket@us-east-1/backupstore], output AWS Error: NoSuchKey The specified key does not exist. <nil>\\n404 \\n\\n, stderr, time="2019-09-04T08:54:11Z" level=error msg="{\\\\n AcceptRanges: \\\\"bytes\\\\",\\\\n Body: <nil>,\\\\n ContentLength: 511,\\\\n ContentType: \\\\"application/xml\\\\",\\\\n Metadata: {\\\\n\\\\n }\\\\n}" pkg=s3\\ntime="2019-09-04T08:54:11Z" level=error msg="AWS Error: NoSuchKey The specified key does not exist. <nil>\\\\n404 \\\\n"\\n, error exit status 1\', \'type\': u\'error\'}') ``` When Longhorn tried to list the backup which was being deleted, this error may occur.
non_main
listing backups may fail during backup deletion clients longhorn staging tests longhorn staging tests longhorn staging tests volume name longhorn testvol pytest mark coretest noqa def test backup clients volume name noqa backup test clients volume name size test basic py test basic py in backup test backupstore test client lht hostid volume name size test basic py in backupstore test backups bv backuplist data longhorn py in cb args kw longhorn py in action return self post and retry url args kw self url args kw retries last error none i e apierror apierror error listing backups for volume longhorn tes exist n n error exit status type u error def post and retry self url args kw retries kw get retries last error none for i in range retries try return self post url data self to dict args kw except apierror as e if e error code last error e time sleep else raise e e apierror apierror error listing backups for volume longhorn testvol error listing backups failed to execute var lib rancher longhorn engine binaries longhornio longhorn engine staging longhorn output aws error nosuchkey the specified key does not exist n n stderr time level error msg n acceptranges bytes n body n contentlength n contenttype application xml n metadata n n n pkg ntime level error msg aws error nosuchkey the specified key does not exist n n error exit status n status code links self u self detail u actions message u error listing backups for volume longhorn testvol error listing backups failed to execute var lib rancher longhorn engine binaries longhornio longhorn engine staging longhorn output aws error nosuchkey the specified key does not exist n n stderr time level error msg n acceptranges bytes n body n contentlength n contenttype application xml n metadata n n n pkg ntime level error msg aws error nosuchkey the specified key does not exist n n error exit status type u error when longhorn tried to list the backup which was being deleted this error may occur
0
753
4,351,730,564
IssuesEvent
2016-08-01 01:10:53
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
azure_rm_virtualmachine module fails creating a virtualmachine when the name of vm contains upper-case.
azure bug_report cloud easyfix waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME azure_rm_virtualmachine ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel 221520cbad) last updated 2016/07/13 15:32:29 (GMT +900) lib/ansible/modules/core: (detached HEAD db8af4c5af) last updated 2016/07/13 15:32:38 (GMT +900) lib/ansible/modules/extras: (detached HEAD 482b1a640e) last updated 2016/07/13 15:32:38 (GMT +900) config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY Creating a new azure virtualmachine with upper-cased letter fails without setting a specific storage account because `AzureRMVirtualMachine.create_default_storage_account` try to create a storage account with upper-case. As described in [this document](https://msdn.microsoft.com/en-us/library/azure/hh264518.aspx), the storage account name can only use numbers and lower-case letters. ##### STEPS TO REPRODUCE Here is a sample task. ```yaml - azure_rm_virtualmachine: name: nameWithUpper resource_group: Testing vm_size: Standard_D1 public_ip_allocation_method: Dynamic admin_username: AdminUserName admin_password: AdminP@ssw0rd open_ports: - 3389 - 5986 os_type: Windows image: publisher: MicrosoftWindowsServer offer: WindowsServer sku: Windows-Server-Technical-Preview version: latest ``` ##### EXPECTED RESULTS The module should convert the vm name to lowercase before trying to create a default storage account. ##### ACTUAL RESULTS Creating storage account always fails as below. ``` TASK [azure_rm_virtualmachine] ************************************************* fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to create a unique storage account name for nameWithUpper. Try using a different VM name."} ```
True
azure_rm_virtualmachine module fails creating a virtualmachine when the name of vm contains upper-case. - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME azure_rm_virtualmachine ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel 221520cbad) last updated 2016/07/13 15:32:29 (GMT +900) lib/ansible/modules/core: (detached HEAD db8af4c5af) last updated 2016/07/13 15:32:38 (GMT +900) lib/ansible/modules/extras: (detached HEAD 482b1a640e) last updated 2016/07/13 15:32:38 (GMT +900) config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY Creating a new azure virtualmachine with upper-cased letter fails without setting a specific storage account because `AzureRMVirtualMachine.create_default_storage_account` try to create a storage account with upper-case. As described in [this document](https://msdn.microsoft.com/en-us/library/azure/hh264518.aspx), the storage account name can only use numbers and lower-case letters. ##### STEPS TO REPRODUCE Here is a sample task. ```yaml - azure_rm_virtualmachine: name: nameWithUpper resource_group: Testing vm_size: Standard_D1 public_ip_allocation_method: Dynamic admin_username: AdminUserName admin_password: AdminP@ssw0rd open_ports: - 3389 - 5986 os_type: Windows image: publisher: MicrosoftWindowsServer offer: WindowsServer sku: Windows-Server-Technical-Preview version: latest ``` ##### EXPECTED RESULTS The module should convert the vm name to lowercase before trying to create a default storage account. ##### ACTUAL RESULTS Creating storage account always fails as below. ``` TASK [azure_rm_virtualmachine] ************************************************* fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to create a unique storage account name for nameWithUpper. Try using a different VM name."} ```
main
azure rm virtualmachine module fails creating a virtualmachine when the name of vm contains upper case issue type bug report component name azure rm virtualmachine ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file configured module search path default w o overrides configuration n a os environment n a summary creating a new azure virtualmachine with upper cased letter fails without setting a specific storage account because azurermvirtualmachine create default storage account try to create a storage account with upper case as described in the storage account name can only use numbers and lower case letters steps to reproduce here is a sample task yaml azure rm virtualmachine name namewithupper resource group testing vm size standard public ip allocation method dynamic admin username adminusername admin password adminp open ports os type windows image publisher microsoftwindowsserver offer windowsserver sku windows server technical preview version latest expected results the module should convert the vm name to lowercase before trying to create a default storage account actual results creating storage account always fails as below task fatal failed changed false failed true msg failed to create a unique storage account name for namewithupper try using a different vm name
1
2,884
10,319,589,571
IssuesEvent
2019-08-30 17:57:26
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
I would like to join the Backdrop Contrib community!
Maintainer application
I have a port of Multiple Selects (D7) ready, and I'm a (co)maintainer of a couple of Drupal projects which I can port, and maintain in the (near) future. -tnx!- Harold
True
I would like to join the Backdrop Contrib community! - I have a port of Multiple Selects (D7) ready, and I'm a (co)maintainer of a couple of Drupal projects which I can port, and maintain in the (near) future. -tnx!- Harold
main
i would like to join the backdrop contrib community i have a port of multiple selects ready and i m a co maintainer of a couple of drupal projects which i can port and maintain in the near future tnx harold
1
2,154
7,481,259,120
IssuesEvent
2018-04-04 20:04:12
lansuite/lansuite
https://api.github.com/repos/lansuite/lansuite
reopened
TLD .bayern domain not accepted
bug pending-maintainer-response
<!--- Provide a general summary of the issue in the Title above --> <!-- Formatting tips: GitHub supports Markdown: https://guides.github.com/features/mastering-markdown/ Multi-line code blocks either with three back ticks, or four space indent. ```php <?php $foo = "bar"; ... ``` --> ## Expected Behavior Register with domain.bayern without an error <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior Top level domain .bayern not accepted <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include configuration, logs, etc. to reproduce, if relevant --> 1. Register on a fresh installation on lansuite with a .bayern domain ## Your Environment It is an default behaviour and not only in my environment.
True
TLD .bayern domain not accepted - <!--- Provide a general summary of the issue in the Title above --> <!-- Formatting tips: GitHub supports Markdown: https://guides.github.com/features/mastering-markdown/ Multi-line code blocks either with three back ticks, or four space indent. ```php <?php $foo = "bar"; ... ``` --> ## Expected Behavior Register with domain.bayern without an error <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior Top level domain .bayern not accepted <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ## Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include configuration, logs, etc. to reproduce, if relevant --> 1. Register on a fresh installation on lansuite with a .bayern domain ## Your Environment It is an default behaviour and not only in my environment.
main
tld bayern domain not accepted formatting tips github supports markdown multi line code blocks either with three back ticks or four space indent php php foo bar expected behavior register with domain bayern without an error current behavior top level domain bayern not accepted steps to reproduce for bugs register on a fresh installation on lansuite with a bayern domain your environment it is an default behaviour and not only in my environment
1
922
4,622,717,663
IssuesEvent
2016-09-27 08:36:17
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
nxos_config isn't idempotent (in some cases)
affects_2.2 bug_report networking P2 waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME nxos_config ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel 547cea556f) last updated 2016/09/20 12:12:18 (GMT +100) lib/ansible/modules/core: (devel 12a7027c49) last updated 2016/09/20 15:11:43 (GMT +100) lib/ansible/modules/extras: (devel db7a3f48e1) last updated 2016/09/20 11:53:00 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY ##### STEPS TO REPRODUCE ``` - name: setup nxos_config: commands: - no description - no shutdown parents: - interface Ethernet2/5 match: none provider: "{{ cli }}" - name: configure device with config nxos_config: src: basic/config.j2 provider: "{{ cli }}" match: none register: result - assert: that: - "result.changed == true" # https://github.com/ansible/ansible-modules-core/issues/4807 - "result.updates is not defined" - name: check device with config nxos_config: src: basic/config.j2 provider: "{{ cli }}" match: none register: result - assert: that: # Idempotent test # https://github.com/ansible/ansible-modules-core/issues/4807 - "result.changed == false" - "result.updates is not defined" ``` ``` cat templates/basic/config.j2 interface Ethernet2/5 description this is a test shutdown ``` ##### EXPECTED RESULTS ##### ACTUAL RESULTS ``` TASK [test_nxos_config : configure device with config] ************************* task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:14 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py <nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" && echo ansible-tmp-1474483552.07-278464852480214="` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" ) && sleep 0' <nxos01> PUT /tmp/tmpmSyMj7 TO /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py <nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/ /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py && sleep 0' <nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/" > /dev/null 2>&1 && sleep 0' changed: [nxos01] => { "changed": true, "invocation": { "module_args": { "after": null, "auth_pass": null, "authorize": false, "backup": false, "before": null, "config": null, "defaults": false, "force": false, "host": "nxos01", "lines": null, "match": "none", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "nxos01", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "transport": "cli", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER" }, "replace": "line", "save": false, "src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n", "ssh_keyfile": null, "timeout": 10, "transport": "cli", "use_ssl": false, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true } }, "warnings": [] } TASK [test_nxos_config : assert] *********************************************** task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:21 ok: [nxos01] => { "changed": false, "invocation": { "module_args": { "that": [ "result.changed == true", "result.updates is not defined" ] }, "module_name": "assert" }, "msg": "all assertions passed" } TASK [test_nxos_config : check device with config] ***************************** task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:27 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py <nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" && echo ansible-tmp-1474483562.39-244603509576702="` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" ) && sleep 0' <nxos01> PUT /tmp/tmpTQ_P7m TO /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py <nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/ /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py && sleep 0' <nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/" > /dev/null 2>&1 && sleep 0' changed: [nxos01] => { "changed": true, "invocation": { "module_args": { "after": null, "auth_pass": null, "authorize": false, "backup": false, "before": null, "config": null, "defaults": false, "force": false, "host": "nxos01", "lines": null, "match": "none", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "nxos01", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "transport": "cli", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER" }, "replace": "line", "save": false, "src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n", "ssh_keyfile": null, "timeout": 10, "transport": "cli", "use_ssl": false, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true } }, "warnings": [] } TASK [test_nxos_config : assert] *********************************************** task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:34 fatal: [nxos01]: FAILED! => { "assertion": "result.changed == false", "changed": false, "evaluated_to": false, "failed": true, "invocation": { "module_args": { "that": [ "result.changed == false", "result.updates is not defined" ] }, "module_name": "assert" } } to retry, use: --limit @/home/johnb/git/ansible-inc/test-network-modules/nxos.retry PLAY RECAP ********************************************************************* nxos01 : ok=46 changed=12 unreachable=0 failed=1 ```
True
nxos_config isn't idempotent (in some cases) - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME nxos_config ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel 547cea556f) last updated 2016/09/20 12:12:18 (GMT +100) lib/ansible/modules/core: (devel 12a7027c49) last updated 2016/09/20 15:11:43 (GMT +100) lib/ansible/modules/extras: (devel db7a3f48e1) last updated 2016/09/20 11:53:00 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY ##### STEPS TO REPRODUCE ``` - name: setup nxos_config: commands: - no description - no shutdown parents: - interface Ethernet2/5 match: none provider: "{{ cli }}" - name: configure device with config nxos_config: src: basic/config.j2 provider: "{{ cli }}" match: none register: result - assert: that: - "result.changed == true" # https://github.com/ansible/ansible-modules-core/issues/4807 - "result.updates is not defined" - name: check device with config nxos_config: src: basic/config.j2 provider: "{{ cli }}" match: none register: result - assert: that: # Idempotent test # https://github.com/ansible/ansible-modules-core/issues/4807 - "result.changed == false" - "result.updates is not defined" ``` ``` cat templates/basic/config.j2 interface Ethernet2/5 description this is a test shutdown ``` ##### EXPECTED RESULTS ##### ACTUAL RESULTS ``` TASK [test_nxos_config : configure device with config] ************************* task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:14 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py <nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" && echo ansible-tmp-1474483552.07-278464852480214="` echo $HOME/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214 `" ) && sleep 0' <nxos01> PUT /tmp/tmpmSyMj7 TO /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py <nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/ /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py && sleep 0' <nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483552.07-278464852480214/" > /dev/null 2>&1 && sleep 0' changed: [nxos01] => { "changed": true, "invocation": { "module_args": { "after": null, "auth_pass": null, "authorize": false, "backup": false, "before": null, "config": null, "defaults": false, "force": false, "host": "nxos01", "lines": null, "match": "none", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "nxos01", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "transport": "cli", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER" }, "replace": "line", "save": false, "src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n", "ssh_keyfile": null, "timeout": 10, "transport": "cli", "use_ssl": false, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true } }, "warnings": [] } TASK [test_nxos_config : assert] *********************************************** task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:21 ok: [nxos01] => { "changed": false, "invocation": { "module_args": { "that": [ "result.changed == true", "result.updates is not defined" ] }, "module_name": "assert" }, "msg": "all assertions passed" } TASK [test_nxos_config : check device with config] ***************************** task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:27 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/nxos/nxos_config.py <nxos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <nxos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" && echo ansible-tmp-1474483562.39-244603509576702="` echo $HOME/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702 `" ) && sleep 0' <nxos01> PUT /tmp/tmpTQ_P7m TO /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py <nxos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/ /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py && sleep 0' <nxos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/nxos_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474483562.39-244603509576702/" > /dev/null 2>&1 && sleep 0' changed: [nxos01] => { "changed": true, "invocation": { "module_args": { "after": null, "auth_pass": null, "authorize": false, "backup": false, "before": null, "config": null, "defaults": false, "force": false, "host": "nxos01", "lines": null, "match": "none", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "nxos01", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "transport": "cli", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER" }, "replace": "line", "save": false, "src": "interface Ethernet2/5\n description this is a test\n shutdown\n\n", "ssh_keyfile": null, "timeout": 10, "transport": "cli", "use_ssl": false, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true } }, "warnings": [] } TASK [test_nxos_config : assert] *********************************************** task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_nxos_config/tests/cli/src_match_none.yaml:34 fatal: [nxos01]: FAILED! => { "assertion": "result.changed == false", "changed": false, "evaluated_to": false, "failed": true, "invocation": { "module_args": { "that": [ "result.changed == false", "result.updates is not defined" ] }, "module_name": "assert" } } to retry, use: --limit @/home/johnb/git/ansible-inc/test-network-modules/nxos.retry PLAY RECAP ********************************************************************* nxos01 : ok=46 changed=12 unreachable=0 failed=1 ```
main
nxos config isn t idempotent in some cases issue type bug report component name nxos config ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary steps to reproduce name setup nxos config commands no description no shutdown parents interface match none provider cli name configure device with config nxos config src basic config provider cli match none register result assert that result changed true result updates is not defined name check device with config nxos config src basic config provider cli match none register result assert that idempotent test result changed false result updates is not defined cat templates basic config interface description this is a test shutdown expected results actual results task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml using module file home johnb git ansible inc ansible lib ansible modules core network nxos nxos config py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home johnb ansible tmp ansible tmp nxos config py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp nxos config py sleep exec bin sh c python home johnb ansible tmp ansible tmp nxos config py rm rf home johnb ansible tmp ansible tmp dev null sleep changed changed true invocation module args after null auth pass null authorize false backup false before null config null defaults false force false host lines null match none parents null password value specified in no log parameter port null provider host password value specified in no log parameter transport cli username value specified in no log parameter replace line save false src interface n description this is a test n shutdown n n ssh keyfile null timeout transport cli use ssl false username value specified in no log parameter validate certs true warnings task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml ok changed false invocation module args that result changed true result updates is not defined module name assert msg all assertions passed task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml using module file home johnb git ansible inc ansible lib ansible modules core network nxos nxos config py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmptq to home johnb ansible tmp ansible tmp nxos config py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp nxos config py sleep exec bin sh c python home johnb ansible tmp ansible tmp nxos config py rm rf home johnb ansible tmp ansible tmp dev null sleep changed changed true invocation module args after null auth pass null authorize false backup false before null config null defaults false force false host lines null match none parents null password value specified in no log parameter port null provider host password value specified in no log parameter transport cli username value specified in no log parameter replace line save false src interface n description this is a test n shutdown n n ssh keyfile null timeout transport cli use ssl false username value specified in no log parameter validate certs true warnings task task path home johnb git ansible inc test network modules roles test nxos config tests cli src match none yaml fatal failed assertion result changed false changed false evaluated to false failed true invocation module args that result changed false result updates is not defined module name assert to retry use limit home johnb git ansible inc test network modules nxos retry play recap ok changed unreachable failed
1
4,167
6,963,655,800
IssuesEvent
2017-12-08 18:17:41
mlibrary/cozy-sun-bear
https://api.github.com/repos/mlibrary/cozy-sun-bear
opened
Build in support for IE11 and IE10
browser compatibility EPUB
CSB must support older versions of IE to IE10. Full functionality is not a requirement. - [ ] Reading and paging must work - [ ] Table of contents must work - [ ] Preferences should work, but not mandatory
True
Build in support for IE11 and IE10 - CSB must support older versions of IE to IE10. Full functionality is not a requirement. - [ ] Reading and paging must work - [ ] Table of contents must work - [ ] Preferences should work, but not mandatory
non_main
build in support for and csb must support older versions of ie to full functionality is not a requirement reading and paging must work table of contents must work preferences should work but not mandatory
0
294,706
22,161,017,938
IssuesEvent
2022-06-04 14:07:15
nigelmann/res-http
https://api.github.com/repos/nigelmann/res-http
opened
Why is a reverse proxy useful to improve the security of the infrastructure ?
documentation
Hello m8 ! There is just one part of the documentation we need to discuss. https://github.com/HEIGVD-Course-API/API-2021-HTTP-Infra#step-4-ajax-requests-with-jquery Maybe tomorrow you have a little time for that ? I would be gr8ful In the mean time if you have some spare time you can check the documentation. I have added the missing parts. Cheers m8 !
1.0
Why is a reverse proxy useful to improve the security of the infrastructure ? - Hello m8 ! There is just one part of the documentation we need to discuss. https://github.com/HEIGVD-Course-API/API-2021-HTTP-Infra#step-4-ajax-requests-with-jquery Maybe tomorrow you have a little time for that ? I would be gr8ful In the mean time if you have some spare time you can check the documentation. I have added the missing parts. Cheers m8 !
non_main
why is a reverse proxy useful to improve the security of the infrastructure hello there is just one part of the documentation we need to discuss maybe tomorrow you have a little time for that i would be in the mean time if you have some spare time you can check the documentation i have added the missing parts cheers
0
235,298
19,322,232,906
IssuesEvent
2021-12-14 07:29:07
pingcap/tidb
https://api.github.com/repos/pingcap/tidb
closed
TiDB CI hang for more then 10 min
type/bug component/test component/tikv severity/major
## Bug Report ``` [2021-11-23T14:16:43.094Z] FAIL github.com/pingcap/tidb/session 600.096s ``` Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) in ci https://ci.pingcap.net/blue/organizations/jenkins/tidb_ghpr_check_2/detail/tidb_ghpr_check_2/47625/pipeline/64 <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) master <!-- Paste the output of SELECT tidb_version() -->
1.0
TiDB CI hang for more then 10 min - ## Bug Report ``` [2021-11-23T14:16:43.094Z] FAIL github.com/pingcap/tidb/session 600.096s ``` Please answer these questions before submitting your issue. Thanks! ### 1. Minimal reproduce step (Required) in ci https://ci.pingcap.net/blue/organizations/jenkins/tidb_ghpr_check_2/detail/tidb_ghpr_check_2/47625/pipeline/64 <!-- a step by step guide for reproducing the bug. --> ### 2. What did you expect to see? (Required) ### 3. What did you see instead (Required) ### 4. What is your TiDB version? (Required) master <!-- Paste the output of SELECT tidb_version() -->
non_main
tidb ci hang for more then min bug report fail github com pingcap tidb session please answer these questions before submitting your issue thanks minimal reproduce step required in ci what did you expect to see required what did you see instead required what is your tidb version required master
0
5,642
28,369,700,374
IssuesEvent
2023-04-12 16:03:48
deislabs/spiderlightning
https://api.github.com/repos/deislabs/spiderlightning
opened
installation fails on v0.4.1 in wsl (ubuntu 20.04) with "`GLIBC_2.32' not found (required by slight)"
πŸ› bug 🚧 maintainer issue
**Description of the bug** ╭── ~ ╰────▢ slight slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.32' not found (required by slight) slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.34' not found (required by slight) slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.33' not found (required by slight) **To Reproduce** the "unix" installation instructions. "unix?" ╭── ~ ╰────▢ /bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/deislabs/spiderlightning/main/install.sh)" >>> LATEST RELEASE: v0.4.1... >>> DONLOADING FROM: https://github.com/deislabs/spiderlightning/releases/download/v0.4.1/slight-linux-x86_64.tar.gz... >>> DOWNLOADED BINARY TAR. >>> EXTRACTED BINARY TAR. >>> INSTALLED BINARY. >>> CLEANED UP. **Additional context** ╭── ~ ╰────▢ uname -a Linux 5.15.90.1-microsoft-standard-WSL2 #1 SMP Fri Jan 27 02:56:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Li
True
installation fails on v0.4.1 in wsl (ubuntu 20.04) with "`GLIBC_2.32' not found (required by slight)" - **Description of the bug** ╭── ~ ╰────▢ slight slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.32' not found (required by slight) slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.34' not found (required by slight) slight: /lib/x86_64-linux-gnu/libc.so.6: version `GLIBC_2.33' not found (required by slight) **To Reproduce** the "unix" installation instructions. "unix?" ╭── ~ ╰────▢ /bin/bash -c "$(curl -fsSL https://raw.githubusercontent.com/deislabs/spiderlightning/main/install.sh)" >>> LATEST RELEASE: v0.4.1... >>> DONLOADING FROM: https://github.com/deislabs/spiderlightning/releases/download/v0.4.1/slight-linux-x86_64.tar.gz... >>> DOWNLOADED BINARY TAR. >>> EXTRACTED BINARY TAR. >>> INSTALLED BINARY. >>> CLEANED UP. **Additional context** ╭── ~ ╰────▢ uname -a Linux 5.15.90.1-microsoft-standard-WSL2 #1 SMP Fri Jan 27 02:56:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Li
main
installation fails on in wsl ubuntu with glibc not found required by slight description of the bug ╭── ╰────▢ slight slight lib linux gnu libc so version glibc not found required by slight slight lib linux gnu libc so version glibc not found required by slight slight lib linux gnu libc so version glibc not found required by slight to reproduce the unix installation instructions unix ╭── ╰────▢ bin bash c curl fssl latest release donloading from downloaded binary tar extracted binary tar installed binary cleaned up additional context ╭── ╰────▢ uname a linux microsoft standard smp fri jan utc gnu li
1
35,745
7,988,291,372
IssuesEvent
2018-07-19 10:31:40
Microsoft/devkit-sdk
https://api.github.com/repos/Microsoft/devkit-sdk
closed
[NOTE] Deprecate existing installation package
Code Ready Feature P1
As we are moving to the new [IoT Workbench](https://marketplace.visualstudio.com/items?itemName=vsciot-vscode.vscode-iot-workbench), the DevKit will be full supported by this new VS Code extension, which will be more convenient for developing, less tooltrain acquisition paints and more straightforward on creating a IoT Solution. We will deprecate current installation package, which includes: - Installation scripts - Task scripts for all examples - All binaries We will keep delivering new SDK library for IoT DevKit, add new features and fix bugs. This is not a 'deprecation' of the IoT DevKit, but a 'evolution' which will lead to a new IoT E2E dev experience. > 'Deprecation' not mean 'Disappearance', all historic packages are still keeping for downloading.
1.0
[NOTE] Deprecate existing installation package - As we are moving to the new [IoT Workbench](https://marketplace.visualstudio.com/items?itemName=vsciot-vscode.vscode-iot-workbench), the DevKit will be full supported by this new VS Code extension, which will be more convenient for developing, less tooltrain acquisition paints and more straightforward on creating a IoT Solution. We will deprecate current installation package, which includes: - Installation scripts - Task scripts for all examples - All binaries We will keep delivering new SDK library for IoT DevKit, add new features and fix bugs. This is not a 'deprecation' of the IoT DevKit, but a 'evolution' which will lead to a new IoT E2E dev experience. > 'Deprecation' not mean 'Disappearance', all historic packages are still keeping for downloading.
non_main
deprecate existing installation package as we are moving to the new the devkit will be full supported by this new vs code extension which will be more convenient for developing less tooltrain acquisition paints and more straightforward on creating a iot solution we will deprecate current installation package which includes installation scripts task scripts for all examples all binaries we will keep delivering new sdk library for iot devkit add new features and fix bugs this is not a deprecation of the iot devkit but a evolution which will lead to a new iot dev experience deprecation not mean disappearance all historic packages are still keeping for downloading
0