id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
1936405649
Need an Easy Way to Add Implied Automatic Modules to the Module Path (Parity with Maven) tl;dr: I just want to add a few implied automatic modules to the module path without managing transitive dependencies. To demonstrate, let me construct a very lightweight example module with three easily-met requirements: The module includes a module-info.java file. It uses Immutables to create @Value.Immutable interfaces (and generate their implementations). It needs Google Guava as a dependency. To repro, we can create a minimalistic Java file (note that the generated code will differ based on whether Guava is present or not): package org.example; import org.immutables.value.Value; @Value.Immutable public interface Empty {} Here's the corresponding module-info.java: module org.example { requires static org.immutables.value; requires com.google.errorprone.annotations; requires jsr305; } (jsr305 is an implied automatic module generated by com.google.code.findbugs:jsr305.) Let's try to get his code to build with Maven first, and then with Gradle. Maven I only have a passing knowledge of Maven, but it didn't take very long for me to generate a working pom.xml file: <?xml version="1.0" encoding="UTF-8"?> <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd"> <modelVersion>4.0.0</modelVersion> <groupId>org.example</groupId> <artifactId>maven-module</artifactId> <version>1.0-SNAPSHOT</version> <properties> <maven.compiler.source>17</maven.compiler.source> <maven.compiler.target>17</maven.compiler.target> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> </properties> <dependencies> <dependency> <groupId>org.immutables</groupId> <artifactId>value</artifactId> <version>2.10.0</version> <scope>provided</scope> </dependency> <dependency> <groupId>com.google.guava</groupId> <artifactId>guava</artifactId> <version>32.1.2-jre</version> </dependency> </dependencies> </project> Now granted, one could criticize Maven for including an implied automatic module on the module path, but this build logic just works. Gradle This won't build out-of-the-box; Gradle will complain that it can't find module jsr305. Now let's 1) temporarily get rid of the module-info.java file, and 2) just add the plugin to the build.gradle.kts: plugins { `java-library` id("org.gradlex.extra-java-module-info") version "1.4.2" } repositories { mavenCentral() } dependencies { compileOnly("org.immutables:value-annotations:2.10.0") annotationProcessor("org.immutables:value:2.10.0") implementation("com.google.guava:guava:32.1.2-jre") } This actually will not build either: Execution failed for task ':lib:compileJava'. > Could not resolve all files for configuration ':lib:compileClasspath'. > Failed to transform failureaccess-1.0.1.jar (com.google.guava:failureaccess:1.0.1) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}. > Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.guava/failureaccess/1.0.1/1dcf1de382a0bf95a3d8b0849546c88bac1292c9/failureaccess-1.0.1.jar. > Not a module and no mapping defined: failureaccess-1.0.1.jar > Failed to transform jsr305-3.0.2.jar (com.google.code.findbugs:jsr305:3.0.2) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}. > Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.code.findbugs/jsr305/3.0.2/25ea2e8b0c338a877313bd4672d3fe056ea78f0d/jsr305-3.0.2.jar. > Not a module and no mapping defined: jsr305-3.0.2.jar > Failed to transform j2objc-annotations-2.8.jar (com.google.j2objc:j2objc-annotations:2.8) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}. > Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.j2objc/j2objc-annotations/2.8/c85270e307e7b822f1086b93689124b89768e273/j2objc-annotations-2.8.jar. > Not a module and no mapping defined: j2objc-annotations-2.8.jar I was expecting a solution where I could add this plugin and this short snippet, and the build would just work like it did with Maven. extraJavaModuleInfo { automaticModule("com.google.code.findbugs:jsr305", "jsr305") } What I actually got is a solution where I also have to do extra work to manage all my transitive dependencies—which will become a real pain for larger projects. Each dep added could potentially introduce new build errors. Now perhaps there are some valid reasons for this design decision w.r.t. managing transitive dependencies. But if I have to do all this extra work just to get Java modules to work with Gradle, why not just use Maven—where things just work out-of-the-box. For many real-world projects, we can realistically expect that you're going to have to take some dependencies that don't use named modules. (E.g., for Undertow and XNIO, you will have to consume an implied automatic module as well.) This is unrelated to this plugin. This plugin is about patching existing Jars to add a module-info.class so that all Jars are real modules and you can fully use the Java Module System as originally intended. The behavior in Maven you describe is to put everything on the --module-path, no matter if the Jars are compatible or not. This may fail for certain setups (invalid names, split packages). In your example it works, but there are many constellations where it does not. And "it works" still means that you do not fully use the module system as there are Automatic Modules involved. The Automatic Modules mechanism in Java only exist as a kind of intermediate compatibility. A lot of the advantages of the Module System are lost once you involve automatic modules. So if you fully want to use the Module System, you should only use dependencies that are already real modules (Guava is not https://github.com/google/guava/issues/2970). But if you really want the behavior of Maven you describe in Gradle, this is a Gradle core issue on the topic: https://github.com/gradle/gradle/issues/12630#issuecomment-1125105629 Feel free to share your argumentation there. With current Gradle versions, you could re-configure tasks in your build to get the Maven behavior: tasks.withType<JavaCompile>().configureEach { doFirst { options.compilerArgs.add("--module-path") options.compilerArgs.add(classpath.asPath) classpath = files() } } tasks.withType<JavaExec>().configureEach { doFirst { jvmArguments.add("--module-path") jvmArguments.add(classpath.asPath) classpath = files() } } It's ultimately your call since this is your project, but I would ask that you reconsider. I'd ask this: what is more likely to attract users to this plugin? One day, they woke up and had this burning desire to convert all JARs to real Java modules. They ran into a (well-)known issue where Gradle cannot not find a module when it's a filename-based automodule. My money is on (2), especially when even javax.inject:javax.inject:1 is a filename-based automodule. It's also worth noting that the benefits of Java modules are not all-or-nothing. E.g., you can still benefit from strong encapsulation for your module's code, even if some of your dependencies have filename-based automodules. This could very well be a case where the perfect (real modules for all JARs) is the enemy of the good (strong encapsulation for your code), especially if people decide to either not use Java modules or switch to Maven. (In my case, I just switched to Maven.) (Also, I had some issues getting your workaround to work.) @mikewacker, like @jjohannes mentioned, this is the way how Gradle works and treat non-modular JARs. This plugin in particular allows you to add the missing metadata either by defining the proper module-info.class or by adding the Automatic-Module-Name entry to the JAR Manifest. While the former is recommended to benefit from the module system the most, the latter will allow you to ("quickly") solve your problem and make Gradle place the JAR on the module path. The example is below. extraJavaModuleInfo { automaticModule("javax.inject:javax.inject", "javax.inject") } The main problem that I noted there is that you have to do this for all your transitive dependencies, not just the modules that have a requires in module-info.java. (See the error message from the original post.) Maybe I misunderstood the request a litte bit. My point is that – from a technical perspective – this plugin cannot be changed to provide exactly the same behavior that Maven has to Gradle: Gradle modifies JARs (when using this plugin) Maven does not modify JARs That being said, I could imagine that we add an option to add the Automatic-Module-Name entry automatically to all Jars that are not Modules. This option would use the name of the Jar file as Module Name (same as what Java does implicitly). It can fail at build time for invalid names and inform the user to explicitly define a name for the problematic Jar. I personally do not see much value in such an option. But maybe it is interesting as one step in migrating an existing project to Modules. Or just for experimentation. I can imagine something like this: extraJavaModuleInfo { allPlainJarsAreAutomaticModules = true } (Happy about bettter suggestions for how to call this option.) If that is what you are looking for @mikewacker, I can reopen this issue (and adjust title and description accordingly). It's moreso that things just work in Maven. From that perspective, it doesn't really matter if that's accomplished by adding plain JARs to the module path, or by converting all plain JARs to (explicit) automatic modules. What Maven does is that it includes plain JARs on the module path, but it generates a warning: [WARNING] ********************************************************************************************************************************************************************************************************************* [WARNING] * Required filename-based automodules detected: [jsr305-3.0.2.jar, undertow-core-2.3.9.Final.jar, xnio-api-3.8.8.Final.jar, javax.inject-1.jar]. Please don't publish this project to a public artifact repository! * [WARNING] ********************************************************************************************************************************************************************************************************************* (Note: This warning only lists plain JARs that the module directly requires in module-info.java. It doesn't list all the plain JARs that you transitively depend on.) As a simple example, let's use one of the most common Java dependencies: Guava. Here is the error that you get if have a very simple build file with only the com.google.com.guava:guava:32.1.3-jre dep and an empty extraJavaModuleInfo{} section. (Let's assume that your module never requires a plain JAR in module-info.java.) Execution failed for task ':lib:compileJava'. > Could not resolve all files for configuration ':lib:compileClasspath'. > Failed to transform failureaccess-1.0.1.jar (com.google.guava:failureaccess:1.0.1) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}. > Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.guava/failureaccess/1.0.1/1dcf1de382a0bf95a3d8b0849546c88bac1292c9/failureaccess-1.0.1.jar. > Not a module and no mapping defined: failureaccess-1.0.1.jar > Failed to transform listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar (com.google.guava:listenablefuture:9999.0-empty-to-avoid-conflict-with-guava) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}. > Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.guava/listenablefuture/9999.0-empty-to-avoid-conflict-with-guava/b421526c5f297295adef1c886e5246c39d4ac629/listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar. > Not a module and no mapping defined: listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar > Failed to transform jsr305-3.0.2.jar (com.google.code.findbugs:jsr305:3.0.2) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}. > Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.code.findbugs/jsr305/3.0.2/25ea2e8b0c338a877313bd4672d3fe056ea78f0d/jsr305-3.0.2.jar. > Not a module and no mapping defined: jsr305-3.0.2.jar > Failed to transform j2objc-annotations-2.8.jar (com.google.j2objc:j2objc-annotations:2.8) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}. > Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.j2objc/j2objc-annotations/2.8/c85270e307e7b822f1086b93689124b89768e273/j2objc-annotations-2.8.jar. > Not a module and no mapping defined: j2objc-annotations-2.8.jar At this point, I'd just take the option to turn all plain JARs into automatic modules, rather than untangle this error. Obviously, one of the big benefits of Java modules is strong encapsulation, i.e., limiting the packages that your module exports. If I want the benefits of strong encapsulation, but I also have an unavoidable dependency on a plain JAR (e.g., Undertow, javax.inject), this is where that feature becomes useful. I just want to add those plain JARs to my module path so that the requires statement compiles; I don't want to manage all my transitive dependencies which are plain JARs. This is in 1.6. Turn it on with this: extraJavaModuleInfo { deriveAutomaticModuleNamesFromFileNames = true } @mikewacker let me know if you run into any issues using this. Thanks! Seems to work with Guava and Immutables, but I ran into an issue with Dagger. I'll open a separate issue for that.
gharchive/issue
2023-10-10T22:52:10
2025-04-01T04:34:25.600349
{ "authors": [ "iherasymenko", "jjohannes", "mikewacker" ], "repo": "gradlex-org/extra-java-module-info", "url": "https://github.com/gradlex-org/extra-java-module-info/issues/74", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2652416210
SceneVariableSet: Notify scene objects that use time macros when time changes Thought this was a clever solution (Making SceneVariableSet notify scene objects that use __from and __to macro when time range changes). But it makes it so that SceneQueryRunner executes double queries when a query is using these macros (as SceneQueryRunner is already subscribing to time range). Need a way to dedupe those updates. Possible solutions dedupe the run queries somehow (when they are called with the same time range) Opt-in to being notified of time range "__from" and "__to" variables (or opt-out) in VariableDependencyConfig Tested locally and it seems to work, and the code looks good. Maybe making it opt-in is safest for the moment? @kaydelaney after thinking more about it I did not like this solution much, becomes a bit complex to have it be opt-in from SceneVariableSet (would need to have some kind of special handling for this variable in the propagation chain). Also feels a bit messy that when time range change we propagate variable change updates for for __from and __to (causing double updates when a scene object depends on both). Opted for a simpler feature in VariableDependencyConfig that can handle the opt-in & and the time range subscription.
gharchive/pull-request
2024-11-12T14:46:25
2025-04-01T04:34:25.816980
{ "authors": [ "kaydelaney", "torkelo" ], "repo": "grafana/scenes", "url": "https://github.com/grafana/scenes/pull/966", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2168005336
Incorrect default value placeholders on host edit form Defect description Create ~/.ssh/config file with the following contents: Host localhost HostName 127.0.0.1 User override-user Open goto and create a host with similar parameter, but leave Login field untouched. Notice that the placeholder of Login field contains your actual login instead of override-user as it is defined in ~/.ssh/config. Also blocks IMPROVEMENT-47 as goto will not display proper ssh key name in the status line.
gharchive/issue
2024-03-04T23:49:18
2025-04-01T04:34:25.827060
{ "authors": [ "grafviktor" ], "repo": "grafviktor/goto", "url": "https://github.com/grafviktor/goto/issues/60", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1738070375
docs: fix broken links Fixes #425 I fixed all links that matched core.telegram.org/bots[^/] and were broken. I also checked links matching core.telegram.org/(?!bots), and none of them were broken. Fan tas tic
gharchive/pull-request
2023-06-02T11:59:09
2025-04-01T04:34:25.856887
{ "authors": [ "KnorpelSenf", "roj1512" ], "repo": "grammyjs/grammY", "url": "https://github.com/grammyjs/grammY/pull/427", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
723700799
Foundry 0.7.4 Multilevel toksen duplication with Furnace When I tried to make a macro run with Multilevel tokens in 0.7.4 I realised that there are two of everything related to Multilevel tokens, as you can see from the pictures, and every option (cloning, teleport, triggers, etc.) is on by default. When I set a macro for example in one of the posssible locations, it overwrites it with the combination of the two locations. For example I write Macro1 in the first Macro region and Macro2 in the second, I update reopen Multilevel tab and both regions are overwritten as Macro1,Macro2. It seems to be caused by The Furnace module, but these two worked together with 0.6.6. Not sure if it needs to be fixed in your side or Furnaces side, but I will post this to both. ### ### ### __ See https://github.com/kakaroto/fvtt-module-furnace/issues/62 Closing as I believe this is now fixed in Furnace 2.3.1
gharchive/issue
2020-10-17T08:01:37
2025-04-01T04:34:25.919969
{ "authors": [ "Webwra", "grandseiken" ], "repo": "grandseiken/foundryvtt-multilevel-tokens", "url": "https://github.com/grandseiken/foundryvtt-multilevel-tokens/issues/27", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1696660513
InitialState is null in App.razor OnInitializedAsync Method Hi, Thanks for the great example! I'm very close to getting this working but running across an issue where the InitialState is null in the App.razor when I start the app. I am using the BlazorServer App. I do initialize the InitialApplicationState class in the _Host.cshtml and the access token is populated when the app starts. However, when it gets to the OnInitializedAsync method in the App.razor, InitialState is null because it is not instantiated (object reference not set to an instance of an object error). Am I missing an instantiation somewhere? Thanks! Oops, forgot to add the param-InitialState="initialState" on the component. Works great now! Thanks @COBowler300. I am glad you got it working.
gharchive/issue
2023-05-04T20:30:50
2025-04-01T04:34:25.931038
{ "authors": [ "COBowler300", "grantcolley" ], "repo": "grantcolley/blazor-auth0", "url": "https://github.com/grantcolley/blazor-auth0/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
526916507
Crawler Error Hi everyone, I'd love to use this tool to help me with my search for any type of housing, however, I'm getting a crawler error. File "/Users/danijel/anaconda3/bin/wg-gesucht-crawler-cli", line 10, in <module> sys.exit(cli()) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 697, in main rv = self.invoke(ctx) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/cli.py", line 109, in cli wg_gesucht.search() File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/crawler.py", line 465, in search ad_list = self.fetch_ads(filters_to_check) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/crawler.py", line 289, in fetch_ads url_list.extend(self.process_filter_results(search_results)) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/crawler.py", line 240, in process_filter_results post_date_link = result.find("td", {"class": "ang_spalte_datum"}).find("a") So apparently they changed their source code, as I cant find the class ang_spalte_datum there any more. I just don't understand the logic well enough to find the new class they replaced it with. Any help is appreciated, I'm really betting on this bot to find a place to live Hey @sechsneun Answered this in the other issue, https://github.com/grantwilliams/wg-gesucht-crawler-cli/issues/10#issuecomment-568722522 Let me know if that helps?
gharchive/issue
2019-11-22T00:25:46
2025-04-01T04:34:25.933793
{ "authors": [ "grantwilliams", "sechsneun" ], "repo": "grantwilliams/wg-gesucht-crawler-cli", "url": "https://github.com/grantwilliams/wg-gesucht-crawler-cli/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1157504271
No graph4nlp_demo folder created when git clone I followed the instructions in the repo: created environment (python 3.8) git clone -b v0.5.5 https://github.com/graph4ai/graph4nlp.git While this created a graph4nlp folder, it does not have a graph4nlp_demo folder. I had to separately clone to the graph4nlp_demo repo graph4nlp and graph4nlp_demo are two separate repositories. You may need to git clone them separately to get both of them. Thanks.
gharchive/issue
2022-03-02T17:42:38
2025-04-01T04:34:25.935455
{ "authors": [ "SaizhuoWang", "vinven7" ], "repo": "graph4ai/graph4nlp_demo", "url": "https://github.com/graph4ai/graph4nlp_demo/issues/9", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
248602050
CarbonLink cache-query return 0 datapoints in some cases graphite-web 1.0.2 (0.9.15 not affected) This issue occurs only in two simultaneous circumstances: STORAGE_DIR in local_settings.py is defined via symbolic link There is an element in metric name of fewer than 4 characters in length Example $ grep 'STORAGE_DIR =' /opt/graphite/webapp/graphite/local_settings.py STORAGE_DIR = '/var/lib/graphite/whisper' $ ls -l /var/lib/graphite/whisper lrwxrwxrwx 1 root root 20 Jun 26 11:59 /var/lib/graphite/whisper -> /mnt/ssd/graphite/db $ curl '127.0.0.1/render/?from=-2min&format=json&target=Env.HTTP.NumConnections' | jq . [ { "target": "Env.HTTP.NumConnections", "datapoints": [ [ 48, 1502180500 ], [ 50, 1502180510 ], [ 50, 1502180520 ], [ null, 1502180530 ], [ null, 1502180540 ], [ null, 1502180550 ], [ null, 1502180560 ], [ null, 1502180570 ], [ null, 1502180580 ], [ null, 1502180590 ], [ null, 1502180600 ], [ null, 1502180610 ] ] } ] $ tail /var/log/graphite/cache.log | grep cache-query 2017-08-08,12:15:08.930 :: CarbonLink cache-query request for HTTP.NumConnections returned 0 datapoints The difference: Env.HTTP.NumConnections HTTP.NumConnections I found the issue in the code. An absolute path to whisper file is truncated to the length of the full metric name. This is not correct because absolute_path ends up with a file extension (.wsp), but relative_fs_path is not. >>> absolute_path='/var/lib/graphite/whisper/Env/HTTP/NumConnections.wsp' >>> real_fs_path='/mnt/ssd/graphite/db/Env/HTTP/NumConnections.wsp' >>> metric_path='Env.HTTP.NumConnections' >>> >>> relative_fs_path = metric_path.replace('.', os.sep) >>> relative_fs_path 'Env/HTTP/NumConnections' >>> absolute_path[:-len(relative_fs_path)] '/var/lib/graphite/whisper/Env/' >>> base_fs_path = os.path.dirname(absolute_path[:-len(relative_fs_path)]) >>> base_fs_path '/var/lib/graphite/whisper/Env' >>> real_base_fs_path = os.path.realpath(base_fs_path) >>> real_base_fs_path '/mnt/ssd/graphite/db/Env' >>> relative_real_fs_path = real_fs_path[len(real_base_fs_path):].lstrip('/') >>> relative_real_fs_path 'HTTP/NumConnections.wsp' >>> fs_to_metric(relative_real_fs_path) 'HTTP.NumConnections' In addition, it is not at all clear why the metric name is newly calculated, because it is originally contained in the variable metric_path. Or am I missing something? All working as expect with this patch for me: --- __init__.py 2017-08-08 09:00:16.000000000 +0300 +++ __init__-new.py 2017-08-08 09:44:21.000000000 +0300 @@ -6,15 +6,6 @@ EXPAND_BRACES_RE = re.compile(r'.*(\{.*?[^\\]?\})') def get_real_metric_path(absolute_path, metric_path): - # Support symbolic links (real_metric_path ensures proper cache queries) - real_fs_path = os.path.realpath(absolute_path) - if absolute_path != real_fs_path: - relative_fs_path = metric_path.replace('.', os.sep) - base_fs_path = os.path.dirname(absolute_path[:-len(relative_fs_path)]) - real_base_fs_path = os.path.realpath(base_fs_path) - relative_real_fs_path = real_fs_path[len(real_base_fs_path):].lstrip('/') - return fs_to_metric(relative_real_fs_path) - return metric_path Ahem... I doubt that fix is that simple. I'm not aware of what get_real_metric_path() doing though. @iksaif @DanCech ? I'm not clear on the exact purpose, but it seems that the problem would be solved by using: relative_fs_path = metric_path.replace('.', os.sep) + '.wsp' @DanCech yes, this should work too. But is not clear what get_real_metric_path() purpose? In my case with a symbolic link, all work fine without this function. Based on the code, this function extracts the metric path by truncating the real path and the resulting value is exactly the same as the metric_path. The propose is to support similar links. Between 0.9 and 1.x, the functionality was expanded to support symlinks beyond the filename in #1738. https://github.com/graphite-project/graphite-web/pull/1738 Ah, cool. Thanks, @cbowman0 ! Maybe you could make a fix then? I'm hitting this issue as well. Have made a PR to fix it.
gharchive/issue
2017-08-08T05:08:17
2025-04-01T04:34:25.963143
{ "authors": [ "DanCech", "cbowman0", "deniszh", "leochen4891", "zasca" ], "repo": "graphite-project/graphite-web", "url": "https://github.com/graphite-project/graphite-web/issues/2012", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1377688345
Fixing typo in documentation (apache) This PR will fix a typo error on apache configuration file 💚 All backports created successfully Status Branch Result ✅ 1.1.x Questions ? Please refer to the Backport tool documentation
gharchive/pull-request
2022-09-19T09:48:04
2025-04-01T04:34:25.966191
{ "authors": [ "anthony-quiros", "deniszh" ], "repo": "graphite-project/graphite-web", "url": "https://github.com/graphite-project/graphite-web/pull/2777", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
126435031
Can not open file for read - although the file exists. So all I did was: $ gcc incbin/incbin.c -o incbin $ ./incbin -o scripts.cpp src/scripts.rc and got an error: Ingwie@Ingwies-Macbook-Pro.local ~/W/IceTea $ file src/scripts.rc src/scripts.rc: ASCII c program text Ingwie@Ingwies-Macbook-Pro.local ~/W/IceTea $ ./out/incbin -o scripts.cpp ./src/scripts.rc failed to open `./src/scripts.rc' for reading What exactly could be the issue for this? In fact, this is the whole script: https://github.com/IngwiePhoenix/IceTea/blob/master/build.sh (will be commenting out the line that calls incbin soon, its actually not even required on UNIX :). But its analogous to the windows variant.) Fixed in 76a0059
gharchive/issue
2016-01-13T14:52:26
2025-04-01T04:34:25.968150
{ "authors": [ "IngwiePhoenix", "graphitemaster" ], "repo": "graphitemaster/incbin", "url": "https://github.com/graphitemaster/incbin/issues/19", "license": "unlicense", "license_type": "permissive", "license_source": "bigquery" }
2559669321
Drop eventuals in favor of Tokio watch + timers @Theodus suggested that Eventuals are really hard to debug and we should probably ship a standard solution like the Gateway currently does by using tokio watch. Hi, can I have this? This is a larger refactor that will update multiple parts of the code. I suggest you starting with only one: escrow_accounts or allocations and we can split the PR in multiple. Actually, to start you can update the dispute manager which is the smallest Eventual in the code. https://github.com/graphprotocol/indexer-rs/blob/main/common/src/attestations/dispute_manager.rs sure, I'll send PR as soon as possible
gharchive/issue
2024-10-01T16:06:29
2025-04-01T04:34:25.970815
{ "authors": [ "gusinacio", "taslimmuhammed" ], "repo": "graphprotocol/indexer-rs", "url": "https://github.com/graphprotocol/indexer-rs/issues/333", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
787163452
Graphql-Server + Graphene + Flask issues I'm trying to setup flask with graphene and graphql-server. As this is just a test which I am running, I'm using the latest beta releases: graphene: 3.0.0b7 graphql-server: 3.0.0b4 Note: I had to use the beta versions of both packages, otherwise I run into dependency issues with graphql-core Now I found some example which use graphql_server.flask.GraphQLView for the flask views. However, I couldnt use the graphene schema directly, I had to use schema.graphql_schema instead. Is this expected? Are there some side effects of this usage? The following code is a minimal working example: from flask import Flask from graphene import ObjectType, String, Schema from graphql_server.flask import GraphQLView class Query(ObjectType): hello = String(name=String(default_value='stranger')) goodbye = String() def resolve_hello(root, info, name): return 'Hello {}'.format(name) def resolve_goodbye(root, info): return 'See ya!' schema = Schema(query=Query) app = Flask(__name__) app.add_url_rule('/graphql', view_func=GraphQLView.as_view( 'graphql', schema=schema.graphql_schema, # TODO: Check what the consequences are of using the graphql_schema graphiql=True, graphiql_version='1.3.2' )) app.add_url_rule('/graphql/batch', view_func=GraphQLView.as_view( 'graphql_batch', schema=schema.graphql_schema, # TODO: Check what the consequences are of using the graphql_schema batch=True )) if __name__ == '__main__': app.run(host='0.0.0.0') Nevermind: Just found a note in the docs which exactly mentiones this... Nevermind: Just found a note in the docs which exactly mentiones this... Hi @jrast ich steh auf dem Schlauch! I have the same issue but can work it out how the docs can help me. Can you show me what you did to solve this issue?
gharchive/issue
2021-01-15T20:09:00
2025-04-01T04:34:26.007787
{ "authors": [ "JonathanBecks", "jrast" ], "repo": "graphql-python/graphql-server", "url": "https://github.com/graphql-python/graphql-server/issues/79", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1105449619
Please Add Mutation Example I saw the examples for juniper but only for query and I don't see any mutation example @alresarena2021 it's not much difference between declaring query and mutations. The only difference is that a mutation should be declared in a Mutation root object. See an example here: https://graphql-rust.github.io/juniper/master/quickstart.html#schema-example
gharchive/issue
2022-01-17T06:58:33
2025-04-01T04:34:26.009545
{ "authors": [ "alresarena2021", "tyranron" ], "repo": "graphql-rust/juniper", "url": "https://github.com/graphql-rust/juniper/issues/1018", "license": "bsd-2-clause", "license_type": "permissive", "license_source": "bigquery" }
2744815627
[graphiql] Autocomplete doesn't work Is there an existing issue for this? [X] I have searched the existing issues Current Behavior A few days ago, GraphiQL was working correctly, but today, without making any significant changes, I've noticed the autocomplete has stopped working. GraphiQL doesn't build the queries as it did a few days ago. I'm using graphiql with Spring boot 3.4.0 and spring-boot-starter-graphql. Expected Behavior Autocomlete working! Steps To Reproduce With a spring boot project. Run the application (without graphql errors) Open graphiql and try to make a query Environment GraphiQL Version: I don't know OS: Ubuntu 24.04.1 Browser: Google Chrome, Opera Bundler:I don't know react Version:I don't know graphql Version:I don't know Anything else? Fixed! Thanks!
gharchive/issue
2024-12-17T12:31:31
2025-04-01T04:34:26.014113
{ "authors": [ "xdelvalle" ], "repo": "graphql/graphiql", "url": "https://github.com/graphql/graphiql/issues/3839", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1325718206
reduce CO2 usage by running less unneccesary GH actions workflows because it's good for the earth 🌍 and the patience of contributors 😆 i wish you could use paths-ignore further down, to skip individual tasks note: i decided to upgrade actions/setup-node usage as well for similar reasons, it appears to be faster but brings some bugs, so i will finish this up tonight a lower priority fix - how to configure workflows to only re-run when only that workflow file changed? probably by using these same configs this makes me want to decouple cypress tests into seperate workflows, as they and vscode-graphql have seperate tests that should only need to be run together if graphql-language-service changes which is rare but usually a good thing haha another possibility, I wonder if some of these extremely edge-case node and graphql-js version matrix tests could just be run on merge to main? that way, a release breaks if it should, but it would be rare. maybe a few times a year at most we have an issue with either throughout the entire monorepo so you know what's fun? mocking various fetch polyfill libraries across node versions, where TextDecoder may or may not be present and more excitement
gharchive/pull-request
2022-08-02T11:34:54
2025-04-01T04:34:26.017067
{ "authors": [ "acao" ], "repo": "graphql/graphiql", "url": "https://github.com/graphql/graphiql/pull/2626", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1713666322
🛑 GrauNeko.com is down In a7d98c7, GrauNeko.com (https://graueneko.com) was down: HTTP code: 0 Response time: 0 ms Resolved: GrauNeko.com is back up in 86cf026.
gharchive/issue
2023-05-17T11:14:51
2025-04-01T04:34:26.034510
{ "authors": [ "graueneko" ], "repo": "graueneko/status-page", "url": "https://github.com/graueneko/status-page/issues/155", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1177755579
6666 - Migrate portal & console & cypress to node 16.10 Issue gravitee-io/issues#6666 Description Migrate portal console cypress to node 16.10 Additional context 📚  View the storybook of this branch here 🚀 CI was able to deploy the build of this PR, so you can now try it directly here Notes: The deployed app is linked to the management API of APIM master. (Same login and password as APIM master) Could you rebase the branch to fix the build ? thx Yes
gharchive/pull-request
2022-03-23T08:05:55
2025-04-01T04:34:26.118583
{ "authors": [ "ThibaudAV", "gcusnieux" ], "repo": "gravitee-io/gravitee-api-management", "url": "https://github.com/gravitee-io/gravitee-api-management/pull/1486", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2595192263
feat: ps - rework executionPhase & policy to use FlowPhase Issue https://gravitee.atlassian.net/browse/APIM-7242 Description We need to keep PS compatible with all APIM 4.3 to 4.5 The goal it's to allow APIM 4.6 to use new way to define policy compatibility into phase. So we add new FlowPhase instead of ExecutionPhase. add new ApiProtocolType like ApiType but "more precise" add policy flowPhaseCompatibility to define FlowPhase compatibility for each ApiProtocolType the Policy studio's internal code has been changed accordingly. But the external interface remains unchanged. Additional context 🧪  Gravitee.io Automatic Prerelease @gravitee/ui-schematics npm install @gravitee/ui-schematics@13.4.0-apim-7242-flowphase-f02a23a yarn add @gravitee/ui-schematics@13.4.0-apim-7242-flowphase-f02a23a 🧪  Gravitee.io Automatic Prerelease @gravitee/ui-policy-studio-angular npm install @gravitee/ui-policy-studio-angular@13.4.0-apim-7242-flowphase-f02a23a yarn add @gravitee/ui-policy-studio-angular@13.4.0-apim-7242-flowphase-f02a23a 🧪  Gravitee.io Automatic Prerelease @gravitee/ui-particles-angular npm install @gravitee/ui-particles-angular@13.4.0-apim-7242-flowphase-f02a23a yarn add @gravitee/ui-particles-angular@13.4.0-apim-7242-flowphase-f02a23a 📚  View the storybook of this branch here :tada: This PR is included in version 13.5.0 :tada: The release is available on: npm package (@latest dist-tag) npm package (@latest dist-tag) npm package (@latest dist-tag) npm package (@latest dist-tag) GitHub release Your semantic-release bot :package::rocket:
gharchive/pull-request
2024-10-17T16:18:40
2025-04-01T04:34:26.128423
{ "authors": [ "ThibaudAV", "graviteeio" ], "repo": "gravitee-io/gravitee-ui-particles", "url": "https://github.com/gravitee-io/gravitee-ui-particles/pull/439", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2409838050
feat(dgeni): generate color palettes and render in design-land PR Checklist Please check if your PR fulfills the following requirements: [x] The commit message follows our guidelines: https://github.com/graycoreio/daffodil/blob/develop/CONTRIBUTING.md#commit [x] Tests for the changes have been added (for bug fixes / features) [x] Docs have been added / updated (for bug fixes / features) PR Type What kind of change does this PR introduce? [ ] Bugfix [x] Feature [ ] Code style update (formatting, local variables) [ ] Refactoring (no functional changes, no api changes) [ ] Build related changes [ ] CI related changes [ ] Documentation content changes [ ] Other... Please describe: What is the current behavior? Fixes: #2905 What is the new behavior? moves lots of docs stuff to @daffodil/documentation to allow design land to access it too Does this PR introduce a breaking change? [ ] Yes [ ] No Other information @xelaint the styling in the design land color component definitely needs some work. I think what I have sufficiently demonstrates how to use the data available though so hopefully you ca take it from here @griest024 is this PR dependent on #2909 ? Yes @griest024 can we remove $error and $daff-white from the autogenerated list? I realized they need to be reworked and should not be part of this color palettes documentation. @griest024 can we remove $error and $daff-white from the autogenerated list? I realized they need to be reworked and should not be part of this color palettes documentation. will do, should any single colors ever be part of the generation or only the palettes that contain multiple shades? I just went ahead and only gened shaded palettes since the implementation is much cleaner @griest024 does this need to be reworked based on the changes Damien made in the last few weeks? @xelaint yes but afaik those changes aren't all merged yet so I will wait until they are to rebase this.
gharchive/pull-request
2024-07-15T23:56:46
2025-04-01T04:34:26.216871
{ "authors": [ "griest024", "xelaint" ], "repo": "graycoreio/daffodil", "url": "https://github.com/graycoreio/daffodil/pull/2912", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
712637057
Feature: Add custom endpoints Magento 2 modules can have their own api endpoints. This PR makes it possible to use those. For example: `$magento = new Magento(); $response = $magento->api('/suttonsilver-pricelists')->get('pricelist/12');` Thanks!
gharchive/pull-request
2020-10-01T08:51:32
2025-04-01T04:34:26.218781
{ "authors": [ "VincentBean", "ahinkle" ], "repo": "grayloon/magento-laravel-api", "url": "https://github.com/grayloon/magento-laravel-api/pull/34", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1095932568
eliminate flicker when zooming When zooming in/out there is sometimes a brief moment when tiles have no tile data, so the map is repainted with the background colour. The visual effect is a brief moment when the whole map is blank, which is a disorienting effect. Fixed with 06bdd9998cd0dbc9ba888fcd50be84284bb7ed1d and eecec083abce467a28ea79cba9188d551c6c98d9
gharchive/issue
2022-01-07T02:55:02
2025-04-01T04:34:26.306416
{ "authors": [ "greensopinion" ], "repo": "greensopinion/flutter-vector-map-tiles", "url": "https://github.com/greensopinion/flutter-vector-map-tiles/issues/22", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1731789
undefined method `route_instance_path' on adding comments When trying to add admin notes, using active_admin 0.3.2 and rails 3.0.10, the following Exception is thrown on the redirect after posting the comment: NoMethodError in Admin::CommentsController#show: undefined method 'route_instance_path' for nil:NilClass Stack trace show the responsible code to be: activeadmin (0.3.2) lib/active_admin/comments.rb:47 This seems to be caused by the changes made in commit 7c2b355bd01ae0bb76b1af1c240d9ea13b617f43, reverting the commit makes it work again. Please tell if you need the full framework trace. Hey, I'm experiencing the same error but in a different flavour. When posting a comment in the comment panel inside the resource "show" i get undefined method 'route_instance_path' for nil:NilClass in activeadmin (0.3.2) lib/active_admin/comments.rb:47:in _callback_before_79'` I'm not using active_reload. I'm using activeadmin 0.3.2 on Rails 3.1.1 I'm having the same problem described by @cosenmarco. My environment is this: https://gist.github.com/1437310 I'll re-open the bug as it seems to affect others, but try adding some useful debug info as to why route_instance_path is not defined. same problem. NoMethodError in Admin::CommentsController#show undefined method `route_instance_path' for nil:NilClass when adding a comment. Thanks
gharchive/issue
2011-09-24T22:18:48
2025-04-01T04:34:26.311218
{ "authors": [ "cosenmarco", "fbuenemann", "mauriziodemagnis", "nafkot" ], "repo": "gregbell/active_admin", "url": "https://github.com/gregbell/active_admin/issues/528", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1466081822
Should I use this library ? Hi All I'm considering using this library but is see that the since April no code was added . Any thoughts? Depends on your requirement. If you have a lot of complex operations then better not use this. I dont see much support here._ 'd recommend https://github.com/table-library/react-table-library if you need a good tool and can accept to go with unpopular lib, which utilizes modern concepts such as composition over configuration, otherwise https://github.com/TanStack/table is a way to go - bit outdated but well maintained and reliable. Do not recommend https://github.com/olifolkerd/tabulator despite author @olifolkerd claims that it is react-ready, it lags and hangs when used in react, bugs are ignored. Also going to leave Material React Table as a possible alternative here. The official MUI X DataGrid and AG Grid suggested above are great suggestions too, but they might require paid licenses for some features. Also going to leave Material React Table as a possible alternative here. The official MUI X DataGrid and AG Grid suggested above are great suggestions too, but they might require paid licenses for some features. I was having some dependency issues with this library. After testing many of the suggested ones I found that sometimes they are not so flexible in terms of custom styles. I used only a few part of this library, so I decided to implement a simple, compatible, drop-in pure MUI+React replacement https://gist.github.com/jsmolina/00dcdcda7897a7c6b23d8657e2d25716 It's just a gist, I do not plan to create a whole library from it.
gharchive/issue
2022-11-28T09:24:21
2025-04-01T04:34:26.316376
{ "authors": [ "Alex-Github-Account", "KevinVandy", "Shreeabcd", "jsmolina", "yaniv-upstream" ], "repo": "gregnb/mui-datatables", "url": "https://github.com/gregnb/mui-datatables/issues/1967", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
322127359
Page is greater than the total available Hi, I'm having some problems using this library. When I'm on the second page, or ahead and I make a search that should return less pages than the one I'm on, I get the following error: Provided options.page of 1is greater than the total available page length of0`` And here is my code: `export default class test extends React.Component { constructor(props) { super(props); this.changePage = this.changePage.bind(this); this.state = { columns: ["foo", "bar", "ho", "hey", "ho", "let's", "Go", ""], data: this.tratarDados(ultimosCoisosStore.getUltimosCoisos()), options: { responsive: 'scroll', selectableRows: false, filter: false, print: false, download: false, rowsPerPageOptions: [5,10,15], page: 0 } } this.updateData = this.updateData.bind(this); this.tratarDados = this.tratarDados.bind(this); this.returnIcones = this.returnIcones.bind(this); this.refreshList = this.refreshList.bind(this); coisoDispatcher.handleAction({ type: 'UPDATE_ULTIMOS_PEDIDOS' }); } componentDidMount() { ultimosCoisosStore.addChangeListener('ULTIMOS_COISOS_UPDATED', this.updateData); coiso.addChangeListener('REFRESH_LIST_ULTIMOS_COISOS', this.refreshList); } componentWillUnmount() { coiso.removeChangeListener('REFRESH_LIST_ULTIMOS_COISOS', this.refreshList); ultimosCoisosStore.addChangeListener('ULTIMOS_COISOS_UPDATED', this.updateData); } tratarDados(data) { let temp = []; for (var x in data) { let data_temp = Object.values(data[x]); data_temp[1] = parseFloat(data_temp[1]); //data_temp[2] = moment(data_temp[2]).format("DD/MM/YYYY HH:mm:ss"); let icones = this.returnIcones(data_temp[6], x, data_temp[7]); data_temp.pop(); data_temp.push(icones); temp.push(Object.values(data_temp)); } return temp; } returnIcones(status, id, show_delete) { if (status === 3 || status === 4) { return ( <BotoesOk key={id} nf_id={id} status={status} ok={true} show_delete={show_delete}/> ) } else { return ( <BotoesOk key={id} nf_id={id} status={status} ok={false} show_delete={show_delete}/> ) } } refreshList() { //this.setState({options.page: 0}); pdvDispatcher.handleAction({ type: 'UPDATE_ULTIMOS_COISOS' }); } updateData() { let dados = this.tratarDados(ultimosCoisosStore.getUltimosCoisos()); this.setState({data: dados}); } render() { return ( <Rodal onClose={this.props.hideModal} className="ultimos-coisos-modal" visible={true} width={1200} > <MUIDataTable key={Math.random()} title={"Últimos Coisos"} data={this.state.data} columns={this.state.columns} options={this.state.options} /> </Rodal> ) } }` Could someone help me unsderstand what is happening? It looks like there's a bug. I will look into this today This issue should be resolved. Upgrade to version 2.0.0-beta-3 Hi Greg, First of all, thx for this library, saves me tons of work. I'm a newbee to javascript so there could be a very simple explanation to this,, but i am experiencing this very same issue on version 2.0.0-beta-37. My code: ''''js import * as React from 'react'; import MUIDataTable from "mui-datatables"; import { createMuiTheme, MuiThemeProvider } from '@material-ui/core/styles'; export class ClientList extends React.Component { state = { page: 0, count: 100, rowsPerPage: 10, data: [] }; getMuiTheme = () => createMuiTheme({ overrides: { MuiMenuItem: { root: { fontSize: '1.25rem' } } , MuiChip: { label: { fontSize: '1.25rem' } } , MuiInputLabel: { root: { fontSize: '1.5rem' } } , MuiTooltip: { tooltip: { fontSize: '1.25rem' } } , MuiInputBase: { input: { fontSize: '1.25rem' } } , MuiSelect: { selectMenu: { fontSize: '1.25rem' } } , MuiTypography: { h6: { fontSize: '2rem' } , caption: { fontSize: '1.25rem' } } , MUIDataTableHeadCell: { root: { fontSize: '1.25rem' , fontWeight: 'bold' } } , MUIDataTableBodyCell: { root: { fontSize: '1.25rem' //,backgroundColor: "#FF0000" } } } }) componentDidMount() { this.getClientListWrapper(0,10); } onChangeRowsPerPage = rowsPerPage => { this.setState({ rowsPerPage }); }; // get data //getData = () => { // this.getClientList().then(data => { // this.setState({ data }); // }); //} //get list of clients getClientList = () => { return new Promise((resolve, reject) => { fetch('api/Client/Index') .then(function (response) { return resolve(response.json()); }) //.then(function (myJson) { console.log(JSON.stringify(myJson)); }) ; }); } //route to client overview page showClientOverview = (id) => { this.props.history.push("/client/edit/" + id); }; getClientListWrapper = (page, rowsPerPage) => { this.getClientList().then(data => { this.setState({ page: page, rowsPerPage: rowsPerPage, data }); }); }; render() { const columns = [{ name: "Id", options: { display: false } }, "Clientnummer", "Naam", "Airline"]; const { data, page } = this.state; const data1 = data.map(item => { return [item.clientId, item.clientNumber, item.fullName, item.companyName] }); const options = { filter: true ,filterType: 'dropdown' ,responsive: 'stacked' ,serverSide: false ,count: data1.length ,page: page ,rowsPerPage: this.state.rowsPerPage ,rowsPerPageOptions: [10, 15, 50, 100] ,onChangeRowsPerpage: this.onChangeRowsPerPage ,onRowClick: (rowData, rowMeta) => { this.showClientOverview(rowData[0]) } ,onTableChange: (action, tableState) => { //console.log(action, tableState); // a developer could react to change on an action basis or // examine the state as a whole and do whatever they want switch (action) { case 'changePage': this.getClientListWrapper(tableState.page, tableState.rowsPerPage); break; } } //TODO: load labels from label table ,textLabels: { body: { noMatch: "Clienten laden...", toolTip: "Sorteren", } , pagination: { next: "Volgende Pagina", previous: "Vorige Pagina", rowsPerPage: "Aantal objecten per pagina:", displayRows: "van", } , toolbar: { search: "Zoek", downloadCsv: "Download CSV", print: "Print", viewColumns: "Filter Kolommen", filterTable: "Filter Tabel", } , filter: { all: "Alles", title: "Filters", reset: "Verwijder Filters", } , viewColumns: { title: "Toon Kolommen", titleAria: "Toon/Verberg Kolommen", } , selectedRows: { text: "geselecteerde rij(en)", delete: "Verwijder", deleteAria: "Verwijder Geselecteerde Rijen", } } }; return ( <MuiThemeProvider theme={this.getMuiTheme()}> <MUIDataTable title={"Team Vervoort - Clienten"} data={data1} columns={columns} options={options} /> </MuiThemeProvider> ); } } '''' I have 21 clients in my database. When navigating to last page (2) there is one last client on that page. If i change rowsPerPage to 15 then, the page goes blank. Refreshing it defaults to first page, rowsPerPage = 10. However: changing rowsPerPage to 50 or 100 works perfectly. Any ideas?
gharchive/issue
2018-05-11T00:12:32
2025-04-01T04:34:26.324427
{ "authors": [ "gregnb", "rafinha90", "skalma" ], "repo": "gregnb/mui-datatables", "url": "https://github.com/gregnb/mui-datatables/issues/48", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2116835737
Consider implementing From where T: Into<#inner_type>? Currently derive(From) produces https://github.com/greyblake/nutype/blob/c8e3f72346c4294fce2982c5f1ea71224552bbe3/nutype_macros/src/common/gen/traits.rs#L125-L134 Is there anything holding back the production of something like the following? quote! { impl<T: Into<#inner_type>> ::core::convert::From<T> for #type_name { #[inline] fn from(raw_value: T) -> Self { Self::new(raw_value.into()) } } } @schneiderfelipe Hi, thanks for bringing this. I guess I had a thought about it, the reason I've decided not to implement it is clarity / explicitness. But you may make me reconsider this. What's you real life use case, where you think it would make things a bit easier? Well, it would make newtypes behave more like the inner type. For instance, something like the following would work, but currently doesn't: #[newtype(derive(From))] struct MyVec(Vec<isize8>); // This works let my_vec: Vec<isize8> = [0, 1, 2].into(); // This does not work, requires a separate From impl let my_vec: MyVec = [0, 1, 2].into(); @schneiderfelipe Sold! :) @danma3x Would you be interested in addressing this one? Sure! @schneiderfelipe Sold! :) @danma3x Would you be interested in addressing this one? There's a problem with my proposed change: it produces conflicting implementations of the trait From, e.g., error[E0119]: conflicting implementations of trait `From<_>` for type `derives::test_trait_from_string::__nutype_Name__::Name` --> test_suite/tests/string.rs:357:9 | 357 | #[nutype(derive(From))] | ^^^^^^^^^^^^^^^^^^^^^^^ | | | first implementation here | conflicting implementation for `derives::test_trait_from_string::__nutype_Name__::Name` | = note: this error originates in the attribute macro `nutype` (in Nightly builds, run with -Z macro-backtrace for more info) In particular, for the case of integers, things seem much worse: error[E0277]: the trait bound `u32: From<i32>` is not satisfied --> test_suite/tests/integer.rs:532:22 | 532 | let amount = Amount::from(350); | ^^^^^^ the trait `From<i32>` is not implemented for `u32` | = help: the following other types implement trait `From<T>`: <u32 as From<bool>> <u32 as From<char>> <u32 as From<u8>> <u32 as From<u16>> <u32 as From<NonZeroU32>> <u32 as From<Ipv4Addr>> = note: required for `i32` to implement `Into<u32>` note: required for `traits::test_trait_from::__nutype_Amount__::Amount` to implement `From<i32>` --> test_suite/tests/integer.rs:529:9 | 529 | #[nutype(derive(From))] | ^^^^^^^^^^^^^^^^^^^^^^^ unsatisfied trait bound introduced here 530 | pub struct Amount(u32); | ^^^^^^ = note: this error originates in the attribute macro `nutype` (in Nightly builds, run with -Z macro-backtrace for more info) So the compiler seems to figure out that 350 should be a u32 just fine if there is only a single From implementation, but "falls back" to i32 and relies on a separate From<i32> implementation otherwise, which should not exist obviously. I'm afraid my proposal can't be done 😞 @greyblake what do you think? The changes I made are in the master branch of my fork https://github.com/schneiderfelipe/nutype/commit/cca70474cba07f55de89f00e88fff2d89126a17e Considering nutypes with validation may be fallible, and From shouldn't support fallible conversions, shouldn't the trait implementation be TryFrom, not From? Considering nutypes with validation may be fallible, and From shouldn't support fallible conversions, shouldn't the trait implementation be TryFrom, not From? Conversions are fallible when there is validation. When there is no validation set, conversions are infallible. If there is validation, nutype won't allow you to derive From. Though you can still derive TryFrom if there is no validation. Ok I gave a try for a generic From implementation, but it's not gonna work the way we want. I you want to reproduce the example, I pushed my changes in generic-from-impl-demo, see also https://github.com/greyblake/nutype/commit/d73ddeb214f47d754fb2a0b80156682437301564 commit. Consider the following example: #[nutype(derive(Into, From))] pub struct Amount(i32); This won't compile: error[E0119]: conflicting implementations of trait `From<Amount>` for type `Amount` The expands into the following: impl ::core::convert::From<Amount> for i32 { #[inline] fn from(value: Amount) -> Self { value.into_inner() } } impl<T> ::core::convert::From<T> for Amount where i32: ::core::convert::From<T>, { #[inline] fn from(raw_value: T) -> Self { let inner_value = i32::from(raw_value); Self::new(inner_value) } } Note, that Into<i32> for Amount is actually defined as From<Amount> for i32, to keep the symmetry. Into<i32> for Amount will got implemented automatically through blanket implementation. This is common practice / idiom of the Rust language. But in this case it becomes also a source of a problem. It's not very obvious to see, but Rust finds 2 conflicting implementation for for From<Amount> for Amount. The std has reasonable implementation of impl<T> From<T> for T, meaning converting from type T to T should just return self. Other path Rust could take is: Amount -> i32 ->Amount, because our generic implementation of From enable this. This could be addressed by implementing Into trait as solely Into<i32> from Amount, meaning we lose From<Amount> for i32 implementation. Between these 2 available options, there is no clear winner, a trade off needs to be make. In this regard, my personal preference would be the current status quo. With that I am closing the issue. Thanks for you attention, if you followed me! @greyblake nicely explained! I'm afraid this won't ever be possible, except maybe if specialisation gets stabilised (rust-lang/rust#31844). Thank you for taking the time though, really appreciated!
gharchive/issue
2024-02-04T02:16:00
2025-04-01T04:34:26.347146
{ "authors": [ "asasine", "greyblake", "schneiderfelipe" ], "repo": "greyblake/nutype", "url": "https://github.com/greyblake/nutype/issues/124", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
429605409
obv implementation Implemented obv. It is another volume based indicator. Please let me know what do you think of this indicator. Hey, thanks for the new contribution :) Please let me know what do you think of this indicator. I haven't heard about this indicator before. The fact that the indicator accumulates volume differences and may return very huge numbers look a little bit weird for me. But I would like to add it to the library :) I haven't heard about this indicator before. The fact that the indicator accumulates volume differences and may return very huge numbers look a little bit weird for me. Quoting from investopedia Despite being plotted on a price chart and measured numerically, the actual individual quantitative value of OBV is not relevant. The indicator itself is cumulative, while the time interval remains fixed by a dedicated starting point, meaning the real number value of OBV arbitrarily depends on the start date. Instead, traders and analysts look to the nature of OBV movements over time; the slope of the OBV line carries all of the weight of analysis. Its a very good indicator for volume divergences. @shreyasdeotare Thanks! OBV is available in the newly released version 0.1.4: https://crates.io/crates/ta/0.1.4
gharchive/pull-request
2019-04-05T06:42:30
2025-04-01T04:34:26.351553
{ "authors": [ "greyblake", "shreyasdeotare" ], "repo": "greyblake/ta-rs", "url": "https://github.com/greyblake/ta-rs/pull/13", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1519823949
Cannot Restore WAX Account Description I am unable to restore my WAX paper backup certificate using "Import from key certificate". Everything seems to work fine and then I get a spinning wheel of death at the "Updating accounts..." display. I'm not doing anything especially strange, just following the prompts, and then it gets stuck at that spinner. See details the Steps to Reproduce section. Platform Desktop (MacOS) Steps To Reproduce Happens using desktop version 1.3.8 on macOS Ventura 13.1 (22C65). Created an EOS and WAX account on mobile. Backed up to paper certificate so I could migrate to desktop and remove them from mobile. Could not import either from the launch screen after fresh install. Kept getting stuck after all verified with a spinning wheel of death at "Updating accounts..." After multiple attempts, finally got the EOS account to import by adding the EOS blockchain and going to "Import from key certificate" instead. When it worked correctly, it gave me a timer countdown and eventually worked just fine. Then I tried to do the same for WAX, added the blockchain, tried to use "Import from key certificate" and tried Option #1 and Option #2 (I really want Option #2) but I always get the spinning wheel of death at "Updating account..." and it doesn't go away after waiting a very long time. I am on a slow (tethered hotspot) at the moment, but nothing too crazy, I'm able to browse websites, etc. Opened up the developer console and I only see some UI related warning that seem irrelevant, and in the network tab I see repeated calls to get_table_rows but nothing that looks like an error. When I close the "Updating account..." window then everything goes away, but no account added, and no progress. So I can't restore this account (must have tried over 20 times by now and same results each time). Relevant log output I don't see any errors in the developer tools. Anything else? I was able to hit this endpoint just fine: https://wax.greymass.com/ { "server_version":"12902721", "chain_id":"1064487b3cd1a897ce03ae5b6a865651747e2e152090f99c1d19d44e01aea5a4", "head_block_num":222827194, "last_irreversible_block_num":222826865, "last_irreversible_block_id":"0d48117112baad3c44cf03bcf887e4c227255a1c76a6ec86dc5594812ff2d4f1", "head_block_id":"0d4812ba91df59048a9da66bfbee81cfe9e2b9210934c140530275500606fe6f", "head_block_time":"2023-01-05T01:02:08.000", "head_block_producer":"sentnlagents", "virtual_block_cpu_limit":762817, "virtual_block_net_limit":1048576000, "block_cpu_limit":200000, "block_net_limit":1048576, "server_version_string":"v2.0.12wax02", "fork_db_head_block_num":222827194, "fork_db_head_block_id":"0d4812ba91df59048a9da66bfbee81cfe9e2b9210934c140530275500606fe6f", "server_full_version_string":"v2.0.12wax02-12902721dbd9c8a8d2734ede4f8978c1ce8c17b8" } Thanks, I'll respond privately. Thanks for the email 👍 I think I've isolated what's going on. The account does indeed lack the required RAM to update the permissions, which is why it's getting stuck. We have added an error handler in our dev build so that it'll at least display this error message. Since this is a pretty common scenario - we have also updated one of Anchor's services to also provide this RAM in instances where this happens on accounts trying to recover. I believe if you try again (no need to update Anchor), you should be able to recover the account now. 1.3.9 is also now released which actually displays an error instead of just getting stuck at this point in the process: https://github.com/greymass/anchor/releases/tag/v1.3.9
gharchive/issue
2023-01-05T01:05:49
2025-04-01T04:34:26.361618
{ "authors": [ "aaroncox", "raycardillo" ], "repo": "greymass/anchor", "url": "https://github.com/greymass/anchor/issues/1324", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
378818353
How to get templateLayoutFactory I need to create markers with custom layout. In issue #27 contextTypes were used to receive ymaps, however after the definiton of the same contextTypes structure I found out that ymaps in the context was empty. The second way I tried was to access ymaps with HOC. And again, incoming ymaps from that HOC hasn't templateLayoutFactory. Maybe I should somehow tell withYMaps to load that module, if it is a module and If not, then how can I access that thing? Sorry, it was realy some kind of a dumb question after all. Prefer to leave this issue here in case someone else will occur the same problem. The main issue is that templateLayoutFactory is realy a module itself and it is not loaded by default. It you want it, then just add it it modules, that should be loaded or load full package, or if you need it on-demand, then just wrap your component in withYMaps, set second param true (component will wait until API is loaded) and as third param is an array, to which you should just add: templateLayoutFactrory to the modules list, that should be loaded. There is also an example of how to do a HOC for templateLayotFactory in migration guide: https://react-yandex-maps.now.sh/migration-guide#onapiavaliable-is-removed-from-the-library It is kinda confusing, I know. We will move a bunch of those to FAQ at some point @gribnoysup, thanks
gharchive/issue
2018-11-08T16:38:41
2025-04-01T04:34:26.367759
{ "authors": [ "ShiiRochi", "gribnoysup" ], "repo": "gribnoysup/react-yandex-maps", "url": "https://github.com/gribnoysup/react-yandex-maps/issues/119", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2476320341
🛑 B-K Lighting, Inc. is down In d31d8be, B-K Lighting, Inc. (https://bklighting.com) was down: HTTP code: 0 Response time: 0 ms Resolved: B-K Lighting, Inc. is back up in 332532f after 26 minutes.
gharchive/issue
2024-08-20T18:52:57
2025-04-01T04:34:26.391389
{ "authors": [ "gripfastistech" ], "repo": "gripfastistech/status", "url": "https://github.com/gripfastistech/status/issues/90", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1428621971
Comfort Activity Switch Name Looks like HK behavior is different with new iOS. Need to fix something to make switch name show activity name, and not "[ZONE] Comfort Activity". I am still experiencing this issue on iOS 17.0.2, macOS 14.0 (23A344) Carrier Infinity: System Control Model # SYSTXCCWIC01-B System Control Firmware Version CESR131626-04.47 I have had some luck with identifying labels but the hold function seems to change their behavior and it cannot be reverted within HomeKit. Using the Carrier Infinity app does work as expected. The switches are significantly improved on 1.6.13 or later. Give it a try, and feel free to reopen (or make a new issue) if you're still experiencing issues.
gharchive/issue
2022-10-30T05:54:04
2025-04-01T04:34:26.397416
{ "authors": [ "MatthewNrmOK", "grivkees" ], "repo": "grivkees/homebridge-carrier-infinity", "url": "https://github.com/grivkees/homebridge-carrier-infinity/issues/385", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
980673270
RangeInput track color prop docs Signed-off-by: GurkiranSingh gurkiransinghk@gmail.com Doc are added for the RangeInput track color prop. https://github.com/grommet/grommet/pull/5275 Hah, @jcfilben -- Sorry, I had this open all day and made a bunch of comments overlapping with yours! Looks like we are on a similar train of thought. Apologies for the confusion.
gharchive/pull-request
2021-08-26T21:51:46
2025-04-01T04:34:26.399752
{ "authors": [ "g4rry420", "halocline" ], "repo": "grommet/grommet-site", "url": "https://github.com/grommet/grommet-site/pull/304", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1313712081
Networking: Bad Gateway should lead to explicit pipeline failure Current behavior: If an endpoint isn't available, nginx responds with its default 502 response. Expected behavior: Rudra makes it explicity, that tests are failing, because the host cannot be resolved. Similar issues with timeouts, etc. should also be taken into account. Modified behavior to print out warning when all responses are 502s.
gharchive/issue
2022-07-21T19:34:13
2025-04-01T04:34:26.417905
{ "authors": [ "grossamos" ], "repo": "grossamos/rudra", "url": "https://github.com/grossamos/rudra/issues/2", "license": "BSD-2-Clause", "license_type": "permissive", "license_source": "github-api" }
36931173
by default ... "Delete workspace before build starts" should be enabled I can't think of a good reason why "Delete workspace before build starts" should not be enabled. It will prevent previous build artifacts/report on the master of prior builds from being incorporated by the current artifacts/report running on the slave. Not doing so can lead to false broken builds. e.g its possible a prior build had generates a broken test report that exists on the master build. It's rare but development may require previous report to generate differences against the current report. They can disable this field ... rather than forcing all most developments who create new builds to enable this parameter. Not having this as a default causes a lot of confusion as any error logs based on workspace files will lead people and developers down the wrong path wondering what is wrong with their build setups. I agree this should be a default value. DotCi now by default cleans workspace before build per https://github.com/groupon/DotCi/blame/master/src/main/java/com/groupon/jenkins/buildtype/dockercompose/BuildConfiguration.java#L165 shellCommands.add("chmod -R u+w . ; find . ! -path \"./deploykey_rsa.pub\" ! -path \"./deploykey_rsa\" -delete"); Interesting side note: Docker user beware of volume mounting as root. It's recommend that docker-compose.override.yml be setup to create the userid of the slave agent. Because testing can result in creation of files and if image is running as root then those files will now be owned by root. However, the next build will lack permission to to perform the necessary find delete in the subsequent build, as the slave agent userid.
gharchive/issue
2014-07-01T21:00:57
2025-04-01T04:34:26.453377
{ "authors": [ "danmconrad", "tmack8001", "vvitayau" ], "repo": "groupon/DotCi", "url": "https://github.com/groupon/DotCi/issues/46", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
212287833
objc-tests: InteropTestsRemoteWithCronet test4MBResponsesAreAccepted failure https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3190/consoleText /jenkins/workspace/gRPC_pull_requests_macos/workspace_objc_macos_dbg_native/src/objective-c/tests/InteropTests.m:164: error: -[InteropTestsRemoteWithCronet test4MBResponsesAreAccepted] : ((error) == nil) failed: "Error Domain=io.grpc Code=3 "Received message larger than max (4194316 vs. 4194304)" UserInfo={io.grpc.TrailersKey={ }}" - Finished with unexpected error: Error Domain=io.grpc Code=3 "Received message larger than max (4194316 vs. 4194304)" UserInfo={io.grpc.TrailersKey={ Test Case '-[InteropTestsRemoteWithCronet test4MBResponsesAreAccepted]' failed (2.314 seconds). This is keeping the macos master build red. Is the fix complex? seen again in https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3314/testReport/junit/(root)/objc_macos_opt_native/objc_tests/ pr: https://github.com/grpc/grpc/pull/10076 against v1.2.x seen again in https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3333/testReport/junit/(root)/objc_macos_dbg_native/objc_tests/ pr https://github.com/grpc/grpc/pull/9986 against v1.2.x Seen again in https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3346/testReport/junit/(root)/objc_macos_dbg_native/objc_tests/ pr #10087 against v1.2.x
gharchive/issue
2017-03-06T23:54:20
2025-04-01T04:34:26.595009
{ "authors": [ "apolcyn", "dgquintas", "y-zeng" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/issues/10003", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
263996006
Artifact build on Windows failed https://sponge.corp.google.com/target?id=da133ab3-ab38-456d-b057-008de7e91964&target=github/grpc&searchFor=&show=ALL&sortBy=STATUS same as https://github.com/grpc/grpc/issues/12872 which has been fixed already.
gharchive/issue
2017-10-09T19:05:37
2025-04-01T04:34:26.596474
{ "authors": [ "adelez", "jtattermusch" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/issues/12911", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
280645112
New Failure: csharp.Grpc.Core.Tests.ContextPropagationTest Test: csharp.Grpc.Core.Tests.ContextPropagationTest Poll Strategy: None URL: https://kokoro2.corp.google.com/job/grpc/job/ubuntu/job/master/job/grpc_basictests_multilang/691 https://sponge.corp.google.com/invocation?id=ca86c237-d2ae-4cb0-89b3-1dd41dd267a0&searchFor= this should also be fixed by https://github.com/grpc/grpc/pull/13675.
gharchive/issue
2017-12-08T23:16:26
2025-04-01T04:34:26.598688
{ "authors": [ "dgquintas", "jtattermusch" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/issues/13690", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
377469077
grpc-java qps worker: qps_driver often reports "Worker 1 could not be properly quit because Received RST_STREAM with error code 8" The entire scenario passes, but qps_driver reports problem terminating java qps_woker. It happens only for grpc-java. The status reported by java worker seems to be GRPC_STATUS_RESOURCE_EXHAUSTED = 8 2018-11-05 13:28:44,247 START: qps_worker_java_0 2018-11-05 13:28:44,251 START: qps_worker_java_1 2018-11-05 13:28:44,254 START: qps_json_driver.java_protobuf_async_streaming_ping_pong_secure 2018-11-05 13:28:44,258 WAITING: 1 queued, 1 jobs running, 0 complete, 0 failed (load 1.00) next: qps_json_driver.quit @ 1.00 cpu 2018-11-05 13:29:32,423 ++ dirname tools/run_tests/performance/run_qps_driver.sh + cd tools/run_tests/performance/../../.. + cmake/build/qps_json_driver '--scenarios_json={"scenarios": [{"name": "java_protobuf_async_streaming_ping_pong_secure", "warmup_seconds": 15, "benchmark_seconds": 30, "num_servers": 1, "server_config": {"async_server_threads": 1, "channel_args": [{"str_value": "latency", "name": "grpc.optimization_target"}], "security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "threads_per_cq": 0, "server_type": "ASYNC_SERVER"}, "client_config": {"security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "channel_args": [{"str_value": "latency", "name": "grpc.optimization_target"}], "async_client_threads": 1, "outstanding_rpcs_per_channel": 1, "rpc_type": "STREAMING", "payload_config": {"simple_params": {"resp_size": 0, "req_size": 0}}, "client_channels": 1, "threads_per_cq": 0, "load_params": {"closed_loop": {}}, "client_type": "ASYNC_CLIENT", "histogram_params": {"max_possible": 60000000000.0, "resolution": 0.01}}, "num_clients": 1}]}' --scenario_result_file=scenario_result.json RUNNING SCENARIO: java_protobuf_async_streaming_ping_pong_secure I1105 13:28:44.266914471 14417 driver.cc:276] Starting server on localhost:10400 (worker #0) D1105 13:28:44.267247177 14417 ev_posix.cc:169] Using polling engine: epollex D1105 13:28:44.267283569 14417 dns_resolver.cc:338] Using native dns resolver D1105 13:28:44.267631541 14417 dns_resolver.cc:279] Start resolving. I1105 13:28:44.268252223 14417 subchannel.cc:874] Connect failed: {"created":"@1541424524.268217537","description":"Failed to connect to remote host: OS Error","errno":111,"file":"/tmpfs/src/github/grpc/src/core/lib/iomgr/tcp_client_posix.cc","file_line":205,"os_error":"Connection refused","syscall":"connect","target_address":"ipv4:127.0.0.1:10400"} I1105 13:28:44.268271388 14417 subchannel.cc:756] Subchannel 0x7f3b680036f0: Retry in 999 milliseconds D1105 13:28:44.268297072 14417 dns_resolver.cc:259] In cooldown from last resolution (from 1 ms ago). Will resolve again in 999 ms D1105 13:28:45.266394537 14429 dns_resolver.cc:279] Start resolving. I1105 13:28:45.267204871 14429 subchannel.cc:715] Failed to connect to channel, retrying I1105 13:28:45.267701347 14417 subchannel.cc:837] New connected subchannel at 0x55bda5385810 for subchannel 0x7f3b680036f0 I1105 13:28:46.097366545 14417 driver.cc:335] Starting client on localhost:10410 (worker #1) D1105 13:28:46.097465420 14417 driver.cc:357] Client 0 gets 1 channels D1105 13:28:46.097501750 14417 dns_resolver.cc:279] Start resolving. I1105 13:28:46.098013094 14417 subchannel.cc:837] New connected subchannel at 0x7f3b68001bf0 for subchannel 0x7f3b68005d40 I1105 13:28:46.837003397 14417 driver.cc:378] Initiating I1105 13:28:46.855068021 14417 driver.cc:399] Warming up I1105 13:29:01.855263396 14417 driver.cc:405] Starting I1105 13:29:01.871796201 14417 driver.cc:432] Running I1105 13:29:31.855285081 14417 driver.cc:446] Finishing clients I1105 13:29:31.883236617 14417 driver.cc:460] Received final status from client 0 I1105 13:29:31.883348844 14417 driver.cc:492] Finishing servers I1105 13:29:31.888333061 14417 driver.cc:506] Received final status from server 0 I1105 13:29:31.889280898 14417 report.cc:82] QPS: 3526.7 I1105 13:29:31.890097805 14417 report.cc:122] QPS: 3526.7 (inf/server core) I1105 13:29:31.890113801 14417 report.cc:127] Latencies (50/90/95/99/99.9%-ile): 246.0/400.5/444.7/520.8/652.1 us I1105 13:29:31.890120604 14417 report.cc:137] Server system time: 0.00% I1105 13:29:31.890126595 14417 report.cc:139] Server user time: 71.74% I1105 13:29:31.890131975 14417 report.cc:141] Client system time: 0.00% I1105 13:29:31.890143351 14417 report.cc:143] Client user time: 70.10% I1105 13:29:31.890148851 14417 report.cc:148] Server CPU usage: 0.00% I1105 13:29:31.890154007 14417 report.cc:153] Client Polls per Request: 0.00 I1105 13:29:31.890158888 14417 report.cc:155] Server Polls per Request: 0.00 I1105 13:29:31.890164413 14417 report.cc:160] Server Queries/CPU-sec: 4911.88 I1105 13:29:31.890169730 14417 report.cc:162] Client Queries/CPU-sec: 5031.00 + '[' performance_test.performance_experiment_singlevm '!=' '' ']' + tools/run_tests/performance/bq_upload_result.py --bq_result_table=performance_test.performance_experiment_singlevm Warning: Table performance_experiment_singlevm already exists Successfully uploaded scenario_result.json to BigQuery. 2018-11-05 13:29:32,423 PASSED: qps_json_driver.java_protobuf_async_streaming_ping_pong_secure [time=48.2sec, retries=0:0] 2018-11-05 13:29:32,423 START: qps_json_driver.quit 2018-11-05 13:29:32,426 WAITING: ETA 48.2 sec; 0 queued, 1 jobs running, 1 complete, 0 failed (load 1.00) 2018-11-05 13:29:32,452 E1105 13:29:32.441533137 14537 driver.cc:550] Worker 0 could not be properly quit because Received RST_STREAM with error code 8 E1105 13:29:32.450617305 14537 driver.cc:550] Worker 1 could not be properly quit because Received RST_STREAM with error code 8 2018-11-05 13:29:32,452 FAILED: qps_json_driver.quit [ret=1, pid=14536, time=0.0sec] E.g. here: https://source.cloud.google.com/results/invocations/b3638c0b-37c1-4ddf-998f-561d98823f41/targets;collapsed=/grpc%2Fcore%2Fmaster%2Flinux%2Fgrpc_e2e_performance_singlevm/log The worker log: 2018-11-05 13:29:35,801 ++ dirname tools/run_tests/performance/run_worker_java.sh + cd tools/run_tests/performance/../../.. + cd ../grpc-java + benchmarks/build/install/grpc-benchmarks/bin/benchmark_worker --driver_port=10410 Nov 05, 2018 1:28:46 PM io.grpc.benchmarks.driver.LoadClient <init> INFO: Client Config server_targets: "localhost:37451" client_type: ASYNC_CLIENT security_params { use_test_ca: true server_host_override: "foo.test.google.fr" } outstanding_rpcs_per_channel: 1 client_channels: 1 async_client_threads: 1 rpc_type: STREAMING load_params { closed_loop { } } payload_config { simple_params { } } histogram_params { resolution: 0.01 max_possible: 6.0E10 } channel_args { name: "grpc.optimization_target" str_value: "latency" } [GC (Allocation Failure) [PSYoungGen: 524800K->18141K(611840K)] 524800K->18165K(2010112K), 0.0229622 secs] [Times: user=0.14 sys=0.03, real=0.02 secs] Nov 05, 2018 1:29:32 PM io.grpc.benchmarks.driver.LoadWorker$WorkerServiceImpl quitWorker INFO: Received quitWorker request. Nov 05, 2018 1:29:32 PM io.grpc.benchmarks.driver.LoadWorker main INFO: DriverServer has terminated. Heap PSYoungGen total 611840K, used 102588K [0x00000000d5580000, 0x0000000100000000, 0x0000000100000000) eden space 524800K, 16% used [0x00000000d5580000,0x00000000da7f7d10,0x00000000f5600000) from space 87040K, 20% used [0x00000000f5600000,0x00000000f67b7498,0x00000000fab00000) to space 87040K, 0% used [0x00000000fab00000,0x00000000fab00000,0x0000000100000000) ParOldGen total 1398272K, used 24K [0x0000000080000000, 0x00000000d5580000, 0x00000000d5580000) object space 1398272K, 0% used [0x0000000080000000,0x0000000080006000,0x00000000d5580000) Metaspace used 20736K, capacity 20896K, committed 21248K, reserved 1069056K class space used 2387K, capacity 2441K, committed 2560K, reserved 1048576K 2018-11-05 13:29:35,801 PASSED: qps_worker_java_1 [time=51.5sec, retries=0:0] @carl-mastrangelo could the error GRPC_STATUS_RESOURCE_EXHAUSTED be reported by grpc-java because of [GC (Allocation Failure) [PSYoungGen: 524800K->18141K(611840K)] 524800K->18165K(2010112K), 0.0229622 secs] [Times: user=0.14 sys=0.03, real=0.02 secs]? Would increasing java heap size help? Btw, this is happening even for single channel ping pong scenarios like "java_generic_async_streaming_ping_pong_secure" or "java_protobuf_async_unary_ping_pong_secure", which seems odd, because with these scenarios, there's max 1 RPC in flight at any given time I'm able to reproduce locally by running tools/run_tests/run_performance_tests.py -l java -r 'java_protobuf_unary_ping_pong_insecure' and it happens in 100% of cases it seems. Based on some local experiments I have a suspicion that the quitWorker logic might be flawed: https://github.com/grpc/grpc-java/blob/65bd38476f2007a65f249bb852c99069ab9f2c87/benchmarks/src/main/java/io/grpc/benchmarks/driver/LoadWorker.java#L254 The driverServer.shutdownNow(); is invoked right after the quitWorker RPC is finished and this seems to be causing the C++ qps_driver to receiver error code 8 (GRPC_STATUS_RESOURCE_EXHAUSTED) from that quitWorker invocation. The GC (Allocation Failure) is a red herring. That just means Young was filled up, and it's doing a GC. Looking through the Java code, we rarely use ResourceExhausted, so I don't think this would be generated locally. shutdown now terminates new incoming RPCs, but should let all existing RPCs finish gracefully. Shutdown commands are serialized along with writes on the socket. Unfortunately I could only repro it once. I am reassigning back to you since I can't diagnose it.
gharchive/issue
2018-11-05T16:08:39
2025-04-01T04:34:26.607328
{ "authors": [ "carl-mastrangelo", "jtattermusch" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/issues/17101", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
456761153
PHP: insecure environment read function 'getenv' used" What version of gRPC and what language are you using? PHP 7.2.19 PECL GRPC 1.21.3 What operating system (Linux, Windows,...) and version? Alpine Linux 3.9.4 What runtime / compiler are you using (e.g. python version or version of gcc) PHP What did you do? Update to 1.21.3 (not sure the what the version was before). Nothing helpful to describe here I'm afraid. What did you expect to see? Expected no warnings in log since default logging level is "ERROR" (and GRPC_VERBOSITY is set to "ERROR"). What did you see instead? [17-Jun-2019 08:02:15] WARNING: [pool application] child 171 said into stderr: "D0617 08:02:15.289927347 171 env_linux.cc:71] Warning: insecure environment read function 'getenv' used" Anything else we should know about your project / environment? Search for this issue and found https://github.com/grpc/grpc/issues/8104 but this is outdated. The issue showed itself after updating to 1.21.3, but I'm not sure which version we came from. I found references of other people recently having this issue https://github.com/grpc/grpc/issues/18833#issuecomment-499518864. Maybe @villers has more insight into what is happening. Also looked into some commits (I'm be no means an expert, just trying to be helpful). The following commit seems to have recently touched the logging levels: 90fbdc92f522af9f98297e08c0ed174361977d46 Does caused by this commit https://github.com/grpc/grpc/pull/18539/files ? I am seeing this as well. Any plans for fixing this? I think this line blow is the cause https://github.com/grpc/grpc/pull/18539/commits/3107cda311853ce57f3a67847ce15dd14a5c9fd5#diff-f4dd247c3672ef404b2cec6be6573a39R208 @kellegous could this be related to https://github.com/grpc/grpc/pull/18539 in your opinion? (as the author). @Swahjak It's possible but I'm not sure how. That change uses getenv directly and doesn't use gpr_getenv which is what issues that warning. It could be possible that somewhere in grpc, the libc getenv is being replaced with gpr_getenv but I haven't looked to see if that's the case. file /src/core/lib/gpr/env_linux.cc line 71: gpr_log(GPR_DEBUG, "Warning: insecure environment read function '%s' used", line 72: insecure_func_used); When SET ENV GRPC_VERBOSITY=ERROR Already print 'Warning: insecure environment read function 'getenv' used' why? I attempted to get a backtrace to the offending call to gpr_getenv this morning and, as best I can tell, this is what is emitting the warning. Breakpoint 1, gpr_getenv ( name=name@entry=0x7ffff7781608 <g_env_str_grpc_verbosity> "GRPC_VERBOSITY") at /tmp/pear/temp/grpc/src/core/lib/gpr/env_linux.cc:69 69 /tmp/pear/temp/grpc/src/core/lib/gpr/env_linux.cc: No such file or directory. (gdb) bt #0 gpr_getenv (name=name@entry=0x7ffff7781608 <g_env_str_grpc_verbosity> "GRPC_VERBOSITY") at /tmp/pear/temp/grpc/src/core/lib/gpr/env_linux.cc:69 #1 0x00007ffff74f2598 in grpc_core::GlobalConfigEnv::GetValue ( this=this@entry=0x7ffff7781620 <g_env_grpc_verbosity>) at /tmp/pear/temp/grpc/src/core/lib/gprpp/global_config_env.cc:75 #2 0x00007ffff74f28b4 in grpc_core::GlobalConfigEnvString::Get ( this=this@entry=0x7ffff7781620 <g_env_grpc_verbosity>) at /tmp/pear/temp/grpc/src/core/lib/gprpp/global_config_env.cc:126 #3 0x00007ffff74ef573 in gpr_global_config_get_grpc_verbosity () at /tmp/pear/temp/grpc/src/core/lib/gpr/log.cc:78 #4 gpr_log_verbosity_init () at /tmp/pear/temp/grpc/src/core/lib/gpr/log.cc:78 #5 0x00007ffff74f2d69 in do_basic_init () at /tmp/pear/temp/grpc/src/core/lib/surface/init.cc:69 #6 0x00007ffff7fbf01a in ?? () from /lib/ld-musl-x86_64.so.1 #7 0x00005555563ab6a0 in core_globals () #8 0x00007ffff7fbef99 in pthread_mutexattr_settype () from /lib/ld-musl-x86_64.so.1 #9 0x00007ffff7784ee8 in g_init_mu () from /usr/local/lib/php/extensions/no-debug-non-zts-20180731/grpc.so file: src/php/ext/grpc/php_grpc.c PHP_RINIT_FUNCTION(grpc) { if (!GRPC_G(initialized)) { apply_ini_settings(TSRMLS_C); grpc_init(); register_fork_handlers(); grpc_php_init_completion_queue(TSRMLS_C); GRPC_G(initialized) = 1; } return SUCCESS; } RINIT function before grpc_init() apply_ini_settings() OR MINIT Function where is it used getenv() ? The warning should be removed after #19691 is merged. We tested this on master and verified the warning is gone. Please watch out for the next upcoming 1.23 release. Fixed in 1.23.0RC1. 1.23.0 will be released in 2 weeks.
gharchive/issue
2019-06-17T06:19:58
2025-04-01T04:34:26.619755
{ "authors": [ "Swahjak", "axot", "kellegous", "lvzhihao", "stanley-cheung", "taka-oyama" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/issues/19366", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
467114561
memory_usage_test fail under mac D0711 10:54:11.677362000 140735725650816 ev_posix.cc:174] Using polling engine: poll D0711 10:54:11.678205000 140735725650816 dns_resolver_ares.cc:485] Using ares dns resolver D0711 10:54:11.692659000 140735725650816 test_config.cc:384] test slowdown factor: sanitizer=1, fixture=1, poller=1, total=1 D0711 10:54:11.692816000 140735725650816 test_config.cc:384] test slowdown factor: sanitizer=1, fixture=1, poller=1, total=1 D0711 10:54:11.693971000 140735725650816 ev_posix.cc:174] Using polling engine: poll D0711 10:54:11.693988000 140735725650816 ev_posix.cc:174] Using polling engine: poll D0711 10:54:11.694074000 140735725650816 dns_resolver_ares.cc:485] Using ares dns resolver D0711 10:54:11.694077000 140735725650816 dns_resolver_ares.cc:485] Using ares dns resolver I0711 10:54:11.694176000 140735725650816 server.cc:180] creating server on: [::]:27749 I0711 10:54:11.696370000 140735725650816 subchannel.cc:1076] New connected subchannel at 0x7ff5afe06160 for subchannel 0x7ff5afe04760 memory_usage_client(23863,0x7fff96ef4380) malloc: *** error for object 0x7ff5afe02150: pointer being freed was not allocated *** set a breakpoint in malloc_error_break to debug Fixed memory_usage_test failure under Linux by #19811. Not sure if it fixed the situation under Mac since there is another test failed #19819 .
gharchive/issue
2019-07-11T21:25:57
2025-04-01T04:34:26.625232
{ "authors": [ "bigfacebear", "yang-g" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/issues/19614", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
964126327
C++ callback service OnDone called twice and app crahsed What version of gRPC and what language are you using? 1.39.0, C++ What operating system (Linux, Windows,...) and version? Windows, OS Version: 10.0.19043 N/A Build 19043 What runtime / compiler are you using (e.g. python version or version of gcc) Microsoft (R) C/C++ Optimizing Compiler Version 19.16.27045 for x64 What did you do? Can't disclose my source, will try to come up w/ a standalone UT, but this is the flow (verified OnDone called twice) Create bidi stream "classic" impl for OnReadDone behavior void GrpcSession::OnReadDone(bool ok) { if (ok) { try { switch (_curReq.message_case()) { // BIZ logic, dispatch requests yada yada } StartRead(&_curReq); } catch (const exception& ex) { FinishWrapper({ INTERNAL, ex.what() }); } } else { FinishWrapper(Status::OK); } } void GrpcSession::OnWriteDone(bool ok) { if (ok) { Response* resp; { lock_guard<mutex> _{ _respMtx }; ASSERT(!_respQueue.empty(), "Queue must be not empty"); _respQueue.pop(); if (!_respQueue.empty()) { resp = _respQueue.front().get(); } else { resp = nullptr; } } if (resp) { StartWrite(resp); } else { _respQueueEmpty.notify_all(); } } else { FinishWrapper({ INTERNAL, "OnWriteDone failed" }); } } void GrpcSession::FinishWrapper(const Status& st) { if (_finishRequested.test_and_set()) // this is atomic_flag { return; } if (!st.ok()) { GS_ERROR("Completing session due to err=%s", st.error_message().c_str()); } Finish(st); } Run client which "bombs" the service w/ messages Close the client app abruptly OnReadDone will invoke Finish (since ok is false) OnDone is called Outstanding write will fail, but I won't issue another call to Finish since FinishWarpper is idempotent Library is initiating another OnDone crash (since its seems ctx_ was destroyed at: void CallOnDone() override { reactor_.load(std::memory_order_relaxed)->OnDone(); grpc_call* call = call_.call(); auto call_requester = std::move(call_requester_); if (ctx_->context_allocator() != nullptr) { ctx_->context_allocator()->Release(ctx_); } this->~ServerCallbackReaderWriterImpl(); // explicitly call destructor ::grpc::g_core_codegen_interface->grpc_call_unref(call); call_requester(); } What did you expect to see? What did you see instead? Anything else we should know about your project / environment? Your code rocks :) I LOVE the new callback approach, and would like to incorporate it deeper in our projects My bad, seems calling Finish while there is an Write in the air is a big no no
gharchive/issue
2021-08-09T15:41:14
2025-04-01T04:34:26.631946
{ "authors": [ "mosdav" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/issues/26947", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
269015630
Bump version to 1.7.1 (mostly to deliver https://github.com/grpc/grpc/pull/13169). [trickle] No significant performance differences [microbenchmarks] No significant performance differences Superseded by https://github.com/grpc/grpc/pull/13168
gharchive/pull-request
2017-10-27T07:43:09
2025-04-01T04:34:26.634028
{ "authors": [ "grpc-testing", "jtattermusch" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/13170", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
372594814
In gRPC-C++ podspec, copy the certificate to bundle resources In GRPC.podspec, the root certificates file from etc/roots.pem is copied to the bundle resources, so that it can be loaded at runtime for SSL to function properly. However, roots.pem is not preserved in gRPC-C++.podspec, so a project depending on that pod won't be able to establish SSL connections. Having the project maintain its own copy of the certificates file and keep it in sync with gRPC repo is undesirable. **************************************************************** libgrpc.so VM SIZE FILE SIZE ++++++++++++++ ++++++++++++++ [ = ] 0 0 [ = ] **************************************************************** libgrpc++.so VM SIZE FILE SIZE ++++++++++++++ ++++++++++++++ [ = ] 0 0 [ = ] [trickle] No significant performance differences Objective-C binary sizes *****************STATIC****************** New size Old size 1,997,294 Total (=) 1,997,294 No significant differences in binary sizes ***************FRAMEWORKS**************** New size Old size 11,021,771 Total (>) 11,021,759 No significant differences in binary sizes Corrupt JSON data (indicates timeout or crash): bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.new: 10 bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.old: 10 [microbenchmarks] No significant performance differences **************************************************************** libgrpc.so VM SIZE FILE SIZE ++++++++++++++ ++++++++++++++ [ = ] 0 0 [ = ] **************************************************************** libgrpc++.so VM SIZE FILE SIZE ++++++++++++++ ++++++++++++++ [ = ] 0 0 [ = ] [trickle] No significant performance differences Objective-C binary sizes *****************STATIC****************** New size Old size 2,015,303 Total (=) 2,015,303 No significant differences in binary sizes ***************FRAMEWORKS**************** New size Old size 11,115,426 Total (=) 11,115,426 No significant differences in binary sizes Corrupt JSON data (indicates timeout or crash): bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.new: 10 bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.old: 10 [microbenchmarks] No significant performance differences Basic Tests MacOS [opt] - #16201 Basic Tests Multi-language Linux - #15308
gharchive/pull-request
2018-10-22T16:29:55
2025-04-01T04:34:26.638176
{ "authors": [ "grpc-testing", "muxi", "var-const" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/16962", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
539373246
Backport #21487 to v1.26.x. Reference: https://github.com/grpc/grpc/pull/21487 This fixes broken Ruby artifact builds on MacOS. CC @veblush @apolcyn Does this mean a patch only for Ruby? Since v1.26.0 hasn't been tagged yet, I don't think this requires a patch at all. Failures: https://github.com/grpc/grpc/issues/20385 Going ahead and merging as this is purely a Ruby change and all relevant Ruby tests have already passed.
gharchive/pull-request
2019-12-17T23:52:38
2025-04-01T04:34:26.641006
{ "authors": [ "gnossen", "srini100" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/21502", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
597537124
Removing obsolete C++ tutorial content Contributes to https://github.com/grpc/grpc.io/issues/180. cc @ejona86 @jtattermusch @srini100 One of the disadvantages of hosting doc pages as part of a repo is that, AFAIK, it isn't possible to create a redirect. So, this PR removes the page content but directs readers to the page's new home. If any of you feel that we should just drop the file, I'm ok with that too. Let me know what you prefer.
gharchive/pull-request
2020-04-09T20:34:28
2025-04-01T04:34:26.642557
{ "authors": [ "chalin" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/22633", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
605663648
dont run resource_quota_server test case under epoll1 See internal b/151212019 for context. based on @karthikravis, @yashykt and @yang-g opinions it's the best to just stop running the test case on epoll1 (which is the only poller when the test case is flaky) the way to exclude a poller didn't exist in generate_tests.bzl, so I added it. also unmarking the test as flaky (as it won't be run at all). At some point we might get rid of epoll1 entirely. This fix is pretty high impact as it should reduce the number of distinct items we're seeing in the flakiness dashboard basically in half:
gharchive/pull-request
2020-04-23T16:11:26
2025-04-01T04:34:26.644881
{ "authors": [ "jtattermusch" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/22751", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
826979740
Fix link to test file @drfloob @bradfol Sorry I missed this! If you'd like to sign the CLA and reopen the PR, I'll be happy to merge it.
gharchive/pull-request
2021-03-10T02:39:47
2025-04-01T04:34:26.645849
{ "authors": [ "bradfol", "drfloob" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/25666", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
868449400
xds_end2end_test test infra: Eds Args refactoring and enhancing WaitForBackend This change is  known issue #26128
gharchive/pull-request
2021-04-27T04:42:00
2025-04-01T04:34:26.647160
{ "authors": [ "donnadionne" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/26093", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1226988091
[promises] Convert lame client @markdroth I think remaining failures are unrelated.
gharchive/pull-request
2022-05-05T17:42:08
2025-04-01T04:34:26.648274
{ "authors": [ "ctiller" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/29587", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1554134795
GCP Observability Logging: Base64 Encode metadata, message and status-details Tested on Stanley's CI Stop this feed On Mon, Jan 23, 2023, 6:52 PM Stanley Cheung @.***> wrote: @.**** approved this pull request. — Reply to this email directly, view it on GitHub https://github.com/grpc/grpc/pull/32184#pullrequestreview-1266722899, or unsubscribe https://github.com/notifications/unsubscribe-auth/AVMQI7VRERDNJKH73HCBZ2DWT47V5ANCNFSM6AAAAAAUEQNCGI . You are receiving this because you are subscribed to this thread.Message ID: @.***>
gharchive/pull-request
2023-01-24T01:31:11
2025-04-01T04:34:26.651642
{ "authors": [ "Tette69", "yashykt" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/32184", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1976796755
[Python Misc] Revert change to print backtrace in server Fix: https://github.com/grpc/grpc/issues/34853 In order to make debugging easier, we have begun printing backtraces in servers. However, this change has the unintended consequence of printing errors to stderr by default, which may not be expected by some users. This PR reverts the change. We recommend that users set up a logging sink if they want to see errors. We will add this to our documentation later. Hi there, any news on when this is going to be released? It's blocking Pulumi on Python 3.12 for everybody :-( https://github.com/pulumi/pulumi/issues/14258 🎉 @XuanWang-Amos, @gnossen, do you know when this will be released? @XuanWang-Amos, when will this be released? Thank you!
gharchive/pull-request
2023-11-03T19:19:35
2025-04-01T04:34:26.654265
{ "authors": [ "XuanWang-Amos", "justinvp", "zyv" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/34877", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
132490048
Add grpc_version.py to PYTHON-MANIFEST.in Should fix #5152 Can one of the admins verify this patch? Can one of the admins verify this patch? this is ok to test This makes most of the distribtests pass https://grpc-testing.appspot.com/view/Artifacts/job/gRPC_distribtest/62/architecture=x64,language=python,platform=linux/console
gharchive/pull-request
2016-02-09T18:07:45
2025-04-01T04:34:26.656289
{ "authors": [ "grpc-jenkins", "jtattermusch", "soltanmm" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/5155", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
142476855
Update clang-format to 3.8 Pros: we standardize on just one version of clang the new version sorts includes I mostly prefer the new decisions Cons: I only mostly prefer the new decisions Churn Let's bite the bullet on this. LGTM.
gharchive/pull-request
2016-03-21T21:29:42
2025-04-01T04:34:26.657986
{ "authors": [ "ctiller", "vjpai" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/5895", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
150154180
Temporarily reduce throughput test depth to 1 RPC per channel due to flake. Need to root-cause the flake. LGTM. Undo LGTM. Looks like the test is still timing out: TIMEOUT: qps_json_driver.cpp_single_channel_throughput_secure [pid=9036] Yes, just realized that. It's become more urgent to root-cause this, and maybe it wasn't caused by this particular change. BTW, note that this failed on the perf smoke test, which is a single-machine test. So, it's not about single-machine vs multi-machine. Superseded by#6270
gharchive/pull-request
2016-04-21T18:27:42
2025-04-01T04:34:26.660241
{ "authors": [ "jtattermusch", "vjpai" ], "repo": "grpc/grpc", "url": "https://github.com/grpc/grpc/pull/6250", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
126461814
Fatal error: travis-ci build Cannot read property 'contents' of undefined this is similar to #330. here is travis-ci build-log and my repository. I am able to build locally , but i can't succeed travis-ci build. Can any one look at this? I am encountering the same issue with my travis-ci build for an Angular + Firebase Application... Gruntfile.js defines the task imagemin: { dist: { files: [{ expand: true, cwd: '<%= yeoman.app %>/images', src: '{,*/}*.{png,jpg,jpeg,gif}', dest: '<%= yeoman.dist %>/images' }] } }, grunt build completes successfully locally... Running "imagemin:dist" (imagemin) task Minified 9 images (saved 77.32 kB) Done, without errors. Execution Time (2016-01-13 17:04:49 UTC) imagemin:dist 1.6s 100% Total 1.6s grunt build errors out via travis-ci... Warning: Running "imagemin:dist" (imagemin) task Fatal error: Cannot read property 'contents' of undefined Execution Time (2016-01-13 17:00:32 UTC) loading tasks 9ms 2% imagemin:dist 441ms 98% Total 450ms Use --force to continue. Aborted due to warnings. I have tried to determine if there is a syntax issue in how I am declaring src in the task... however, that does not appear to be the case... as I can successfully complete the build locally without error using multiple combinations, including the example provided by Grunt for grunt-contrib-imagemin Ref: https://github.com/gruntjs/grunt-contrib-imagemin Any thoughts or suggestions? As a follow-up... I have successfully moved passed this issue by swapping out "grunt-contrib-imagemin": "^1.0.0" for "grunt-image": "^1.2.1" ... and subsequently updating my Gruntfile.js by replacing imagemin tasks with image tasks... Local builds and Travis builds are now succeeding with the same results. Gruntfile.js image: { dist: { files: [{ expand: true, cwd: '<%= yeoman.app %>/images', src: '{,*/}*.{png,jpg,jpeg,gif}', dest: '<%= yeoman.dist %>/images' }] } }, Sample build output below: Running "image:dist" (image) task ✔ app/images/npmjs.png -> before=3.75 kB after=158 B reduced=3.59 kB(95.9%) ✔ app/images/bootstrap.png -> before=41.5 kB after=21.93 kB reduced=19.57 kB(47.2%) ✔ app/images/bowerjs.png -> before=17.71 kB after=6.65 kB reduced=11.06 kB(62.5%) ✔ app/images/nodejs.png -> before=8.45 kB after=2.93 kB reduced=5.51 kB(65.3%) ✔ app/images/starterlog.png -> before=34.53 kB after=10.58 kB reduced=23.95 kB(69.3%) ✔ app/images/gruntjs.png -> before=84.38 kB after=27.98 kB reduced=56.4 kB(66.8%) ✔ app/images/firebase.png -> before=32.03 kB after=12.19 kB reduced=19.83 kB(61.9%) ✔ app/images/fontawesome.png -> before=12.01 kB after=5.63 kB reduced=6.38 kB(53.1%) ✔ app/images/angular.png -> before=41.5 kB after=18.63 kB reduced=22.87 kB(55.1%) Done, without errors. Execution Time (2016-01-13 17:47:44 UTC) image:dist 2.6s 99% Total 2.6s it works :+1: !!!! after removing "^" when specifying a version of vinyl-fs in package.json file , it builds successfully with this build-log. Thanks @juliusosokinas , @lorijoan , @ernestorocha , @FellowHobbyist and everybody else. I am closing this issue now!
gharchive/issue
2016-01-13T16:44:56
2025-04-01T04:34:26.678513
{ "authors": [ "FellowHobbyist", "anbestephen" ], "repo": "gruntjs/grunt-contrib-imagemin", "url": "https://github.com/gruntjs/grunt-contrib-imagemin/issues/345", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
561854760
Failed to upload state: AccessDenied (As root user) Hi Terragrunt team, I'm trying to use Terragrunt for a from scratch project I am doing and found out that I can't upload the state to the S3 bucket created with terragrunt initusing the root account. I think it could be related to #770 and #978. I'm using latest version: $ terragrunt -version terragrunt version v0.21.11 $ terraform -version Terraform v0.12.19 + provider.aws v2.48.0 Output of terragrunt apply: module.iam_amontalban.aws_iam_user.this[0]: Creating... Failed to save state: failed to upload state: AccessDenied: Access Denied status code: 403, request id: 992F1BEF89128A5B, host id: XvdXpf1q2kx7NIuD26MU+OaaIbtEjGMdZgI6IGqXZe0sxducyoj7d/bIe0PwixdiPatZNC+VbJI= Error: Failed to persist state to backend. Policy in the S3 bucket created with Terragrunt: $ aws-vault exec acme_root -- aws s3api get-bucket-policy --bucket acme-bucket { "Policy": "{\"Version\":\"2012-10-17\",\"Statement\":[{\"Sid\":\"RootAccess\",\"Effect\":\"Allow\",\"Principal\":{\"AWS\":\"arn:aws:iam::123456789012:root\"},\"Action\":\"s3:*\",\"Resource\":\"arn:aws:s3:::acme-bucket\"}]}" } Current running user: $ aws-vault exec acme_root -- aws sts get-caller-identity { "UserId": "123456789012", "Account": "123456789012", "Arn": "arn:aws:iam::123456789012:root" } Not even an aws s3 cp works: $ aws-vault exec acme_root -- aws s3 cp variables.tf s3://acme-bucket --sse=aws:kms upload failed: ./variables.tf to s3://acme-bucket/variables.tf An error occurred (AccessDenied) when calling the PutObject operation: Access Denied Let me know if you need any other information. Thanks! Investigating this a little bit looks like the policy generated in #978 is wrong, it is: { "Version": "2012-10-17", "Statement": [ { "Sid": "RootAccess", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::123456789012:root" }, "Action": "s3:*", "Resource": "arn:aws:s3:::acme-bucket" } ] } When it should be like: { "Version": "2012-10-17", "Statement": [ { "Sid": "RootAccess", "Effect": "Allow", "Principal": { "AWS": "arn:aws:iam::123456789012:root" }, "Action": "s3:*", "Resource": [ "arn:aws:s3:::acme-bucket", "arn:aws:s3:::acme-bucket/*" ] } ] } After applying the above policy to the S3 bucket I was able to store the state. Thanks for reporting! I think you may be right. Would you be up for a quick PR to fix this? Hi @brikis98 I've added the fix for this issue, can you check it. Thanks. Should be fixed in https://github.com/gruntwork-io/terragrunt/releases/tag/v0.21.13. Please check (binaries should show up in a few min).
gharchive/issue
2020-02-07T20:27:00
2025-04-01T04:34:26.685801
{ "authors": [ "amontalban", "brikis98", "nicomfer" ], "repo": "gruntwork-io/terragrunt", "url": "https://github.com/gruntwork-io/terragrunt/issues/1038", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1500130125
Migrate terraform to terragrunt for high number of modules I have microservice non-prod env. where each team has its own env. which means a lot of envs, in terms of terrafrom right now I have two github repos: The first one has root tf modules. The Second one has reusable tf modules. The first one contains 10 root modules each one of them calls just a single module that resides in the second repo, this module in the second repo calls 13 modules, so one of the modules in the second repo represents a proxy repo. to recap, the first repo structure: tf-env1.tf tf-env2.tf tf-env3.tf .... tf-env10.tf The second one has a structure: api-gateway.tf vpc.tf rds.tf ec2.tf s3.tf lambda.tf lb.tf SPECIAL_Module.tf the call chain goes as: in repo1 : tf-envX.tf | in repo2 : SPECIAL_Module.tf | in repo2 : api-gateway.tf,vpc.tf,rds.tf,s3.tf .... etc As part of migrating to terragrunt in order to give the developers an ability to control their env. via managing their env. variables and free up the SRE team to just writing new terraform modules. I'm thinking of getting rid of SPECIAL_Module.tf, since it has a lot of local controls on how terraform works, but the issue is if I remove SPECIAL_Module.tf terragrunt doesn't support more than one source on each hcl file. Another option is to create an hcl for each module from the second repo, but this means I would manage # of modules * # of envs files., so it came to my mind to group the modules in the second repo into 3 or 4 groups. One solution I thought about is to move SPECIAL_Module.tf to the first module and include it in each env., but this means a huge single terraform file would reside inside terragrunt repo. which I'm not sure if it's an abuse for terragrunt. So is there a way to migrate from terraform to terragrunt with a high # of modules and a high # of envs? hi, ideas on how to adopt Terragrunt can be collected from examples: https://github.com/gruntwork-io/terragrunt-infrastructure-live-example https://terragrunt.gruntwork.io/docs/features/keep-your-terraform-code-dry/ Code from SPECIAL can be referenced in dependencies https://terragrunt.gruntwork.io/docs/reference/config-blocks-and-attributes/#dependencies The first link is very simple example use case, in the second one the # of modules are 3 with 3 env. which is a different case.
gharchive/issue
2022-12-16T12:27:47
2025-04-01T04:34:26.697384
{ "authors": [ "aladdin55", "denis256" ], "repo": "gruntwork-io/terragrunt", "url": "https://github.com/gruntwork-io/terragrunt/issues/2397", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
918019161
Adapt count parser to tf 0.15.5 The plan output for no changes changed in tf 0.15.5, so we need to update the regex. Thanks for review! Build failed, but upon investigation they are transient errors unrelated to this change set, so will go ahead and merge this in!
gharchive/pull-request
2021-06-10T22:49:36
2025-04-01T04:34:26.698675
{ "authors": [ "yorinasub17" ], "repo": "gruntwork-io/terratest", "url": "https://github.com/gruntwork-io/terratest/pull/925", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1149724270
🛑 bretterhofer.at is down In 7e8ce6b, bretterhofer.at (https://bretterhofer.at) was down: HTTP code: 0 Response time: 0 ms Resolved: bretterhofer.at is back up in 9cbb890.
gharchive/issue
2022-02-24T20:37:14
2025-04-01T04:34:26.713080
{ "authors": [ "grzchr15" ], "repo": "grzchr15/uptime", "url": "https://github.com/grzchr15/uptime/issues/4314", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2061779015
Active Identifier Unhandled Error thrown in read handler I'm receiving the below error fairly frequently within the logs, and after awhile it the xbox plugin stops working until a restart. The below error seems to correlate to whenever I open home app, or run a random home shortcut and polling of the xbox occurs even when I'm not interacting with it. Below is my config. Which is simple in nature. "devices": [ { "name": "Living Room Xbox Series X", "host": "10.0.1.132", "xboxLiveId": "REMOVEDID", "webApiControl": true, "getInputsFromDevice": true, "filterGames": true, "filterApps": true, "filterSystemApps": true, "filterDlc": true, "inputsDisplayOrder": 1, "sensorPower": false, "sensorInput": false, "sensorScreenSaver": false, "webApiPowerOnOff": true, "webApiRcControl": false, "webApiVolumeControl": false, "webApiToken": "REMOVEDTOKEN", "webApiClientId": "REMOVEDCLIENTID", "enableDebugMode": false, "disableLogInfo": false, "disableLogDeviceInfo": false, "infoButtonCommand": "nexus", "volumeControl": -1, "enableRestFul": false, "restFulDebug": false, "enableMqtt": false, "mqttDebug": false, "mqttAuth": false } ], "_bridge": { "username": "0E:D7:D1:31:C2:57", "port": 42538 }, "platform": "XboxTv" } Log snippet: 01/01/2024, 14:42:43] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, set Power: ON [01/01/2024, 14:42:53] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Connected. [01/01/2024, 14:42:53] [homebridge-xbox-tv] -------- Living Room Xbox Series X --------' [01/01/2024, 14:42:53] [homebridge-xbox-tv] Manufacturer: Microsoft [01/01/2024, 14:42:53] [homebridge-xbox-tv] Model: Xbox Series X [01/01/2024, 14:42:53] [homebridge-xbox-tv] Serialnr: REMOVED [01/01/2024, 14:42:53] [homebridge-xbox-tv] Firmware: 10.0.25398 [01/01/2024, 14:42:53] [homebridge-xbox-tv] Locale: en-US [01/01/2024, 14:42:53] [homebridge-xbox-tv] ---------------------------------- [01/01/2024, 14:42:59] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, set Power: OFF [01/01/2024, 14:43:16] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Disconnected. [01/01/2024, 14:44:16] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 14:44:17] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Current Media: STOP [01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Target Media: STOP [01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Mute: ON [01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Volume: 0 [01/01/2024, 14:44:49] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:44:49] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:45:22] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:45:27] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:45:43] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:52:24] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 14:52:24] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:52:30] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:52:32] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 14:52:32] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:52:34] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 14:52:35] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 14:53:31] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 14:53:31] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 16:01:52] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 16:01:52] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 16:12:38] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 16:12:39] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 16:13:04] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF [01/01/2024, 16:13:04] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 16:13:11] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [01/01/2024, 16:13:12] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. I am also getting this error, though I haven't managed to capture it when debug is enabled yet: [01/01/2024, 12:11:13] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId') Have now been able to capture the error with debug logging, hopefully the attached will help you @grzegorz914 homebridge.log.txt this should be fixed with latest update 2.12.3 Thanks for your fantastic effort! Thanks @grzegorz914 I have updated to 2.12.4 and will let you know if I see the error again. Still seeing it in my logs on 2.12.4 unfortunately: [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Power: OFF [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Current Media: STOP [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Target Media: STOP [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Mute: ON [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Volume: 0 [02/01/2024, 17:21:30] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Power: OFF [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Current Media: STOP [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Target Media: STOP [02/01/2024, 17:21:30] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Mute: ON [02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Volume: 0 [02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Power: OFF [02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Current Media: STOP [02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Target Media: STOP [02/01/2024, 17:21:34] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info. [02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Mute: ON [02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Volume: 0 Still seeing the error here too after updating log.txt So the steps to reproduce this are: Restart Homebridge/child bridge while console is powered off (plugin successfully polls console and logs info) Power on console (plugin still successfully polls console and logs info) Power off the console (plugin polls console successfully for power/media state but errors getting input information). Polling will continue to fail until either the console is powered back on or Homebridge/child bridge is restarted. Please go to Homebridge/xboxTv/ folder and remove inputs, inputsNames and inputsTatgetVisibility files then restart plugin and let me know. Updated to 2.12.5 and removed those files and that seems to have worked, thanks for your efforts @grzegorz914! Updated and tested, everything is working great so far. Thanks @grzegorz914 ! Logs are staying clean, responsiveness of home actions and device state representation has improved back to similar duration prior to implementation of this plugin. Kudos, keep up the fantastic work. Plan on adding second xbox to this configuration and thoroughly testing so that I can roll out for the family. This is going to be game changer for them and my use case for automation of the AV systems.
gharchive/issue
2024-01-01T21:21:15
2025-04-01T04:34:26.748179
{ "authors": [ "chris4prez", "grzegorz914", "insimbi-stuart" ], "repo": "grzegorz914/homebridge-xbox-tv", "url": "https://github.com/grzegorz914/homebridge-xbox-tv/issues/189", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2069225292
2.12.18 - Send connect request error: ReferenceError: tokenExist is not defined Just updated to 2.12.18 before I was going to test and provide feedback on issues 191 - Enabling a second Xbox within the plugin failing But before I could start testing, after update I'm getting the below errors and everything is no longer working. I blew away my entire configuration for child bridge, and single xbox, deleted home hub, xbox device and recreated all (performed auth, etc.) yet still receiving same error below without on/off working any more. [1/7/2024, 12:48:36 PM] [homebridge-xbox-tv] Restarting child bridge... [1/7/2024, 12:48:36 PM] Got SIGTERM, shutting down child bridge process... [1/7/2024, 12:48:41 PM] [homebridge-xbox-tv] Child bridge process ended [1/7/2024, 12:48:41 PM] [homebridge-xbox-tv] Process Ended. Code: 143, Signal: null [1/7/2024, 12:48:48 PM] [homebridge-xbox-tv] Restarting Process... [1/7/2024, 12:48:49 PM] [homebridge-xbox-tv] Launched child bridge with PID 52327 [1/7/2024, 12:48:49 PM] Registering platform 'homebridge-xbox-tv.XboxTv' [1/7/2024, 12:48:49 PM] [homebridge-xbox-tv] Loaded homebridge-xbox-tv v2.12.18 child bridge successfully [1/7/2024, 12:48:49 PM] Loaded 0 cached accessories from cachedAccessories.0E67955C5891. [1/7/2024, 12:48:49 PM] Homebridge v1.7.0 (HAP v0.11.1) (homebridge-xbox-tv) is running on port 54592. [1/7/2024, 12:48:57 PM] Living Room Xbox Series X 0EDB is running on port 43213. [1/7/2024, 12:48:57 PM] Please add [Living Room Xbox Series X 0EDB] manually in Home app. Setup Code: 562-07-087 [1/7/2024, 12:49:00 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined [1/7/2024, 12:49:09 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined [1/7/2024, 12:49:20 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined [1/7/2024, 12:49:29 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined [1/7/2024, 12:49:39 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined [1/7/2024, 12:49:49 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined [1/7/2024, 12:49:59 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined [1/7/2024, 12:50:09 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined thanks issues resolved.
gharchive/issue
2024-01-07T17:56:53
2025-04-01T04:34:26.757876
{ "authors": [ "chris4prez" ], "repo": "grzegorz914/homebridge-xbox-tv", "url": "https://github.com/grzegorz914/homebridge-xbox-tv/issues/194", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1844300158
aria2的ARIANG_RPC_SECRET_AUTO默认参数建议设置为false ARIANG_RPC_SECRET_AUTO false 如果设置为true,当远程访问aria2下载页面时,会在页面里默认设置为token的值,也就是任何人只要打开aria2页面就可以下载了,有风险。 当ARIANG_RPC_SECRET_AUTO设置为false,任何人打开aria2下载页面,如果不设置token的值,是无法下载的。 我自行测试,设置ARIANG_RPC_SECRET_AUTO ture时,在任何地方打开我的aria2就显示为连接状态。设置ARIANG_RPC_SECRET_AUTO false时,只有自己设置了token的浏览器才能进行下载,否则连不上。 当aria2远程下载暴露在公网时,别人不知道你的token值,也无法操作。 新版已默认设置为false。
gharchive/issue
2023-08-10T02:26:31
2025-04-01T04:34:26.769015
{ "authors": [ "gshang2017", "youland" ], "repo": "gshang2017/docker", "url": "https://github.com/gshang2017/docker/issues/202", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
54047692
Filtering parameters similar to rails' filter_parameters It would be useful to filter out sensitive details from logs at a global level - similar to how rails does this could also be inherited from the rails configuration That's an application-level concern, not a logging one. Rails builds it into the framework and utilises those methods rather than their raw equivalent where appropriate.
gharchive/issue
2015-01-12T12:22:07
2025-04-01T04:34:26.770535
{ "authors": [ "gshutler", "stephenbinns" ], "repo": "gshutler/hatchet", "url": "https://github.com/gshutler/hatchet/issues/13", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1365167126
Updating readme to be more accurate. Signed-off-by: David Tippett dtip@amazon.com This commit updates some of the readme to better reflect what is being run at the Wikimedia Foundation and Snagajob. Just saw Flaxsearches PR so closing as duplicate XD
gharchive/pull-request
2022-09-07T20:47:13
2025-04-01T04:34:26.771836
{ "authors": [ "dtaivpp" ], "repo": "gsingers/opensearch-learning-to-rank-base", "url": "https://github.com/gsingers/opensearch-learning-to-rank-base/pull/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2152902701
Defer animation on list items until scrolled into view Similar to https://pub.dev/packages/flutter_staggered_animations I would like to be able to only animate a list's items once they appear. Regardless of scrolling, or items changing, they should only animate for the first time. Sorry for the delayed response on this. Interesting! I'll have to look at how they made that work. Worst case, I anticipate it would be pretty easy for the author of flutter_staggered_animations to create an animation type that would accept any flutter_animate instance. Keeping open and will look into it when time permits. Feel free to take a stab at it in the meantime. Okay, I've been giving this a bunch of thought, and have the start of a flexible API in mind that I'm fairly sure can be implemented: foo.animate(hiddenBehavior: myBehavior); HiddenBehavior(behavior, visibleFraction=0.5, resetOnHide=false, keepAlive=true) // behaviors: HiddenBehavior.PLAY // always play HiddenBehavior.SKIP // skip to end if initially hidden HiddenBehavior.PAUSE // remain paused at beginning if initially hidden HiddenBehavior.WAIT // pause until visible This would facilitate all of the behaviors I can think of: current behavior (play on creation) only play visible items, new items scroll in already at the end (or start) of their animation play each item only when it first scrolls in play every time it scrolls back in Feedback is very welcome. I'm not 100% sure I love the naming of HiddenBehavior, but I think it's a bit more semantically intuitive than VisibilityBehavior (which is perhaps the more expected choice).
gharchive/issue
2024-02-25T19:17:56
2025-04-01T04:34:26.775388
{ "authors": [ "gskinner", "harkairt" ], "repo": "gskinner/flutter_animate", "url": "https://github.com/gskinner/flutter_animate/issues/131", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1829374758
Update mlSearcher фичи по тайм-ауту одной карты, рестарту сервера для каждой карты новые карты из ветки Анны убрал фичу с выбором количества сохраняемых агентов. пусть все сохраняются, в следующий раз не будет возможности забыть выставить правильное значение фикс табличек: теперь там лежат кортежи <coverage, test_count, error_count, steps_count> Что-то линтер ругается.
gharchive/pull-request
2023-07-31T14:52:45
2025-04-01T04:34:26.783571
{ "authors": [ "emnigma", "gsvgit" ], "repo": "gsvgit/VSharp", "url": "https://github.com/gsvgit/VSharp/pull/69", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1610239306
support loading a collection which contains everything Allow loading a collection of precomputed output. The input is a path to a collection that is a dictionary with keys X, Y and samples. @emulate(collection="test.pkl") def test(a, b): pass Codecov Report Patch coverage: 52.94% and project coverage change: +23.20 :tada: Comparison is base (27ce372) 11.16% compared to head (a18e472) 34.36%. :mega: This organization is not using Codecov’s GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more Additional details and impacted files @@ Coverage Diff @@ ## master #14 +/- ## =========================================== + Coverage 11.16% 34.36% +23.20% =========================================== Files 9 9 Lines 412 419 +7 =========================================== + Hits 46 144 +98 + Misses 366 275 -91 Impacted Files Coverage Δ galilei/emulator.py 55.46% <52.94%> (+40.59%) :arrow_up: galilei/experimental.py 67.56% <0.00%> (+67.56%) :arrow_up: galilei/backends/sklearn.py 90.90% <0.00%> (+90.90%) :arrow_up: Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here. :umbrella: View full report at Codecov. :loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
gharchive/pull-request
2023-03-05T15:14:55
2025-04-01T04:34:26.811617
{ "authors": [ "codecov-commenter", "guanyilun" ], "repo": "guanyilun/galilei", "url": "https://github.com/guanyilun/galilei/pull/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
165726894
sm2 加解密问题 老师,你好 我用sm2做加解密测试的时候, 加密可以出来结果 解密的时候失败了,这是什么问题 API的调用正确吗,还是其他的什么问题 int test_sm2_enc() { EC_KEY *key1; EC_GROUP *group1; int nid,ret; key1 = EC_KEY_new(); if(!key1) return 0; int nid = NID_sm2p256v1; group1 = EC_GROUP_new_by_curve_name(nid); if(group1 == NULL) return 0; ret = EC_KEY_set_group(key1,group1); if(ret != 1) return 0; ret = EC_KEY_generate_key(key1); //EC_KEY_print_fp(stdout,key1,0); /*---- ----*/ printf("----加密测试----\n"); char tmp_buf[48] = {0}; memset(tmp_buf,'1',48); char out_buf[48+128] = {0}; int buf_size = 48 +128; if(!SM2_encrypt_with_recommended(out_buf,&buf_size, (const unsigned char *)tmp_buf,sizeof(tmp_buf),key1)) { printf("SM2 Encryp ERROR !!!\n"); return -1; } else { printf("SM2 Encrypt Result[%d] :\n",buf_size); //print_hex(out_buf,buf_size); printf("----测试成功----\n\n"); } printf("----解密测试----\n"); unsigned char msg[128] = {0}; size_t msglen = 0; if(!SM2_decrypt_with_recommended(msg,&msglen,out_buf,buf_size,key1)) { printf("解密失败!!!\n"); return -1; }else { printf("SM2 Decrypt Result[%d] :\n",msglen); //print_hex(msg,msglen); printf("----测试成功----\n\n"); } return 0; } 另外想问一个国密SSL的问题 现在代码树中定义了SSL_kECC 密钥交换算法,我想问这是什么算法 从文档上看和SSL_kRSA算法应该是差不多(用非对称密钥加密预主密钥),但是似乎openssl里面没有这种类似的算法 是不是要自己实现 ----加密测试---- SM2 Encrypt Result[145] : 04 1F D6 11 66 1E 53 60 A0 C2 98 E4 78 08 7D 50 E6 4C 98 CC 0C 3E 68 7B 5C F5 D9 3E B3 CE 38 CA E0 D2 99 02 EF 2A 68 18 3F 80 2D 76 93 2C DD F5 3C 87 54 62 F0 C2 87 51 47 EC B9 E5 DA 63 57 1E 78 15 AE 1B 36 83 F1 58 0B 1C 8A 58 3F F8 EE 30 0A 6D 3C 25 E1 BC AE E3 52 CA CA 6C F4 31 E6 FD 32 34 A7 C7 24 EE 6A 07 91 02 9B 73 1F DC 8E BD 90 91 26 02 CB 7C 35 F2 B9 94 83 8A 5F C1 B7 F8 A4 62 F6 32 E0 B1 F3 F4 55 C5 39 9F DA EA 80 48 54 ----测试成功---- ----解密测试---- 解密失败!!! 另外cryto/sm2/sm2test.c 编译之后运行也是出错的 sm2 sign p256 passed sm2 sign b257 passed error: sm2test.c 359 139772977178256:error:3406A06A:SM2 routines:SM2_decrypt:error:sm2_enc.c:520: sm2 enc p256 failed SM2 test failed 通过关注项目得到的信息是目前版本提供了完整的加解密示例 是不是我没找对示例代码的位置 sm2test.c可以看做是SM2加密签名的示例,涵盖了SM2内部接口和EVP接口的调用。 建议: 项目目前似乎没有.gitignore文件,个人感觉.gitignore在开发过程中比较重要,我从openssl项目中拷贝了.gitignore过来稍作改动可以参考一下 基于第一点,发现在最新一次更新中居然有109个文件的改动,其中大部分是.save文件,如果是必要的,最好单独提交一次commit,如果不是,.gitignore 中加入Makefile.sava或者*.save能避免不必要的更新 提议更新再频繁一点,比如一个功能做一次更新,一次更新夹杂太多的文件,容易淹没了真正需要关注的更改(比如最新一次更新),不利于其他开发者理解项目的进展情况,比如通过阅读代码我发现对于国密SSL的支持,本项目已经做了相当的改造,但是从commit 信息找到相关的信息就比较困难,可以将commit 细化,分类 (当然可能会增加工作量,我也只是建议) 最后,对于SM2加解密失败的问题,示例文件 sm2test.c 单独拿出来编译,也是解密会失败的,显然没有做过良好的测试,我做了简单的修改后可以正常用于测试了 提交了pull request 望采纳 https://github.com/guanzhi/GmSSL/pull/35 非常感谢您的建议! .gitgnore我们后续我们会加上去 更新频率由目前开发方式导致暂时无法加快,并且由于暑期的影响近期频率会降低,不过我们正在准备社区化,期望在有更多贡献者加入后会形成对开发者比较友好的更新方式 处于和上面相同的原因,没有提前通知我们的pull request难以直接合并,不过我们会在理解完代码之后将您的代码并入,并通过主页和AUTHORS文件等方式注明您的贡献。 老师,目前我在做国密SSL的改造,想问几个关于密钥交换算法的问题 1.openssl中目前的密钥交换算法SSL_kECDHe 和 SSL_kEECH 算法有什么区别 2.GmSSL新定义的SSL_kECC算法又是怎么回事 3.GM/T 0024文档中描述的ECDHE 和ECC 密钥交换和以上的算法的对应关系是什么 SSL_kECC预计实现国密SSL的ECC-SM3-SM4套件,SSL_kECDHe和SSL_kEECH分别对应ECDHE-XXX、ECDH-XXX等SSL标准套件。 我是想问,国密套件中的ECDHE算法,比如(ECDHE_SM4_SM3),ECDHE对应的是SSL_kECDHe吗,如果是SSL_kECDHe的话,那按照openssl的实现 等于国密套件中的ECDHE,ECC和RSA密钥交换都没有ServerKeyExchange消息,但是GM/T 0024又描述了这几种消息(6.4.4.3),我觉得比较迷惑,还望解答一下 GM标准中的ECDHE对应双证书SM2密钥交换,ECC对应SM2公钥加密,这在SSL协议中没有对应的算法,因此不是SSL_kECDHe,需要定义新的SSL_kXXXXX。 就是说 1、ECC_SM4_SM3 对应的是SM2公钥加密,这在SSL协议中没有,需要自己定义实现 2、ECDHE_SM4_SM3 对应的【双证书SM2密钥交换】,不是SSL_kECDHe,也不是SSL_kEECDH,同样需要自己定义并实现?那这个【双证书SM2密钥交换】应该怎么理解?我应该查哪个相关的资料 SM2公钥加密密钥交换可以拿RSA密钥交换来作为参考,容易理解, 但是这个双证书是个什么概念呢,有什么可参考的呢 我这么理解: 服务器取服务器证书中的公钥,通过ServerKeyExchange发送 客户端取客户端证书中的公钥,通过ClientKeyExchange发送 通过SM2密钥交换算法,协商出预主密钥,实现密钥交换 全程完全使用证书中的公钥来做密钥交换,不会生成临时密钥对?(SSL_kECDHe,SSL_kEECDH都生成了临时密钥对) 是应该这样实现吗 关老师,我学习openssl时间不长,问题有点缺乏逻辑,还望包含 1.服务器都各有2对证书密钥(加密证书,签名证书),叫做双证书? ECDHE算法,服务器,客户端都有两对SM2证书密钥 2.在密钥交换和消息认证中,无论是RSA,ECC,ECDHE都区分加密证书和签名证书,都需要双证书,不只是ECDHE_XXX套件需要双证书? 我开发过程中完全漏了双证书, 那等于在国密ssl的实现中,双证书才是重点,国密算法GmSSL项目都完全支持好了,快吐血了 服务器都各有2对证书密钥(加密证书,签名证书),这应该叫做单证书,单证书是指证书中的密钥仅支持加密或签名操作中的一种,我的理解。 在 2016年7月25日,18:59,XiaoTian <notifications@github.commailto:notifications@github.com> 写道: 服务器都各有2对证书密钥(加密证书,签名证书),叫做双证书 老师,我还想最后确认两个问题 1 无论是 RSA_XXX,ECC_XXX,ECDHE_XXX,在国密标准里服务器都需要使用双证书? 2 【双证书SM2密钥交换】,客户端需要有双证书吗,客户端的交换密钥是来自客户端的加密证书还是使用临时密钥? 老师 关于【国密SM2双证书的密钥交换】,我使用了服务器的加密密钥和客户端的临时密钥的方式实现(可以成功建立连接),这样做是合理的吗,还是客户端和服务器都必须要使用双证书 国密套件中的ECC_XXX 和 RSA_XXX套件实现也是要使用双证书吗(目前单证书已经实现了) 我要在下周结束SSL开发,还望老师帮忙确认一下这两个问题 http://www.cacrnet.org.cn/upload/fckeditor/File/2014/罗俊ssl vpn技术规范宣贯最终版0725挂网.pdf 其中有对12个密码套件使用的密钥和证书的说明,这个应该是比较权威的解释了。 太感谢了,正是我需要的信息!!!! 正式产品开发的话,建议去新华书店买一份纸质的规范,京东上应该也有。 网上的电子版不保证与实际发布的纸质版完全一致。 掌晓愚 在 2016年8月2日,17:27,xiyayadamaozai <notifications@github.commailto:notifications@github.com> 写道: 太感谢了,正是我需要的信息!!!! ― You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHubhttps://github.com/guanzhi/GmSSL/issues/33#issuecomment-236852005, or mute the threadhttps://github.com/notifications/unsubscribe-auth/AL-bho9aZ13rZYY0_qyN_xaeWsDpqKKdks5qbw1sgaJpZM4JNJLZ. 你好,老师,现在GmSSL编译的能编译支持arm64的么?你们给的编译后只支持x86_64的. 支持的,我测试过的,arm-linux 下编译运行都没问题 我是想把GmSSL编译成静态库用于iOS开发,是在mac环境下编译的.发现不支持arm64的. iOS模拟器和arm32支持吗? 你好,在mac环境下编译的我发现只支持x86_64,也就是说只支持iOS模拟器,对于iOS开发真机arm64是不支持的。 发自 网易邮箱大师 在2016年08月13日 05:32,Zhi Guan 写道: iOS模拟器和arm32支持吗? — You are receiving this because you commented. Reply to this email directly, view it on GitHub, or mute the thread. 你可以把编译错误信息帖上来看看 另外请新开一个issue 你好,这是iOS开发真机报错: ld: warning: ld: warning: ignoring file /Users/a123456/Desktop/SM2 SM4/libcrypto.a, file was built for archive which is not the architecture being linked (arm64): /Users/a123456/Desktop/SM2 SM4/libcrypto.aignoring file /Users/a123456/Desktop/SM2 SM4/libssl.a, file was built for archive which is not the architecture being linked (arm64): /Users/a123456/Desktop/SM2 SM4/libssl.a ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(CharacterKeyboard.o)) was built for newer iOS version (9.0) than being linked (7.0) ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(NumberKeyboardThird.o)) was built for newer iOS version (9.0) than being linked (7.0) ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(KeyboardTool.o)) was built for newer iOS version (9.0) than being linked (7.0) ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(Keyboard.o)) was built for newer iOS version (9.0) than being linked (7.0) ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(XYKeyboard.o)) was built for newer iOS version (9.0) than being linked (7.0) ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(NumberKeyboardFirst.o)) was built for newer iOS version (9.0) than being linked (7.0) ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(NumberKeyboardSecond.o)) was built for newer iOS version (9.0) than being linked (7.0) ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(UIView+LVExtension.o)) was built for newer iOS version (9.0) than being linked (7.0) Undefined symbols for architecture arm64: "_EVP_sm3", referenced from: _test_sm2_enc in sm2Plugin.o "_EC_POINT_point2hex", referenced from: _test_sm2_enc in sm2Plugin.o "_EC_KEY_new", referenced from: -[sm2Plugin sm2:] in sm2Plugin.o "_EC_POINT_point2oct", referenced from: _SM2_do_encrypt in sm2_enc.o "_KDF_get_x9_63", referenced from: _SM2_do_encrypt in sm2_enc.o "_CRYPTO_malloc", referenced from: _SM2_do_encrypt in sm2_enc.o "_BN_new", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_GROUP_get_order", referenced from: _SM2_do_encrypt in sm2_enc.o "_EVP_DigestInit_ex", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_POINT_new", referenced from: _SM2_do_encrypt in sm2_enc.o "_BN_CTX_new", referenced from: _SM2_do_encrypt in sm2_enc.o -[sm2Plugin sm2:] in sm2Plugin.o _test_sm2_enc in sm2Plugin.o "_EVP_MD_CTX_create", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_KEY_get0_group", referenced from: _SM2_do_encrypt in sm2_enc.o -[sm2Plugin sm2:] in sm2Plugin.o _test_sm2_enc in sm2Plugin.o "_EC_GROUP_get_cofactor", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_KEY_get0_public_key", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_GROUP_get_degree", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_POINT_hex2point", referenced from: -[sm2Plugin sm2:] in sm2Plugin.o "_BN_rand_range", referenced from: _SM2_do_encrypt in sm2_enc.o "_OpenSSLDie", referenced from: _SM2_do_encrypt in sm2_enc.o _test_sm2_enc in sm2Plugin.o "_BN_free", referenced from: _SM2_do_encrypt in sm2_enc.o "_EVP_DigestUpdate", referenced from: _SM2_do_encrypt in sm2_enc.o "_EVP_DigestFinal_ex", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_POINT_mul", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_POINT_free", referenced from: _SM2_CIPHERTEXT_VALUE_free in sm2_enc.o "_BN_CTX_free", referenced from: _SM2_do_encrypt in sm2_enc.o "_BN_num_bits", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_POINT_is_at_infinity", referenced from: _SM2_do_encrypt in sm2_enc.o "_EVP_MD_CTX_destroy", referenced from: _SM2_do_encrypt in sm2_enc.o "_EC_KEY_new_by_curve_name", referenced from: -[sm2Plugin sm2:] in sm2Plugin.o "_CRYPTO_free", referenced from: _SM2_CIPHERTEXT_VALUE_free in sm2_enc.o ld: symbol(s) not found for architecture arm64 clang: error: linker command failed with exit code 1 (use -v to see invocation)
gharchive/issue
2016-07-15T07:24:05
2025-04-01T04:34:26.846407
{ "authors": [ "LiTianjue", "conezxy", "guanzhi", "sinv", "tongyu123" ], "repo": "guanzhi/GmSSL", "url": "https://github.com/guanzhi/GmSSL/issues/33", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2116687355
🛑 Sefaz - Rio Grande do Sul is down In 42ce968, Sefaz - Rio Grande do Sul (https://onboardapi.guichepass.com.br/sefaz?code=2) was down: HTTP code: 500 Response time: 157 ms Resolved: Sefaz - Rio Grande do Sul is back up in 8cd33f9 after 1 hour, 13 minutes.
gharchive/issue
2024-02-03T21:41:26
2025-04-01T04:34:26.894957
{ "authors": [ "suporte-gpass" ], "repo": "guichevirtual/statuspage", "url": "https://github.com/guichevirtual/statuspage/issues/1814", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1651005455
css style border-radius not work I tried with to generate image with html and css style like below command: +++HTML ` <meta charset="UTF-8"> <body> <img src="${basics.image}" alt="HTML5 Icon" width="100" height="100" style="border-radius: 50%;"> </body> `+++ I expected it will generate an image with the circle shape, BUT it didn't. If you're sure your HTML itself is correct, please read up on the limitations of HTML altChunk in MS Word. It is possible that what you are trying to do is not supported by MS Word.
gharchive/issue
2023-04-02T16:13:38
2025-04-01T04:34:26.907188
{ "authors": [ "jjhbw", "minuth" ], "repo": "guigrpa/docx-templates", "url": "https://github.com/guigrpa/docx-templates/issues/304", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2465271471
[BUG] autoFocus not triggering keyboard on iOS devices Version of the library: 1.2.4 Expected Behavior When the OTPInput component is rendered with autoFocus, the first slot should be focused, and the keyboard should automatically appear on iOS devices. Current Behavior When using the OTPInput component with the autoFocus prop set to true, the first slot is correctly focused on iOS devices (both web and native), but the keyboard does not automatically appear. Users must tap on the focused slot to make the keyboard appear. Steps to Reproduce Create a new project using the input-otp library Implement the OTPInput component with the autoFocus prop set to true Run the project on an iOS device (either in a web browser or as a native app) Observe the behavior when the OTPInput component is rendered I've created a CodeSandbox to demonstrate this issue I also noticed this problem today on iOS devices How awkward. No idea why the keyboard wouldn't be triggering... Hey @MorganMeirFitussi I just found out that won't work with iOS. It won't work with a normal <input />, too. See this is the only workaround and it's not pretty https://gist.github.com/cathyxz/73739c1bdea7d7011abb236541dc9aaa?permalink_comment_id=4412959#gistcomment-4412959 Even tho there's a workaround available for that, it shouldn't be responsibility of input-otp's but more of a general knowledge on iOS inputs.
gharchive/issue
2024-08-14T08:54:38
2025-04-01T04:34:26.917327
{ "authors": [ "ETOPS7", "MorganMeirFitussi", "guilhermerodz" ], "repo": "guilhermerodz/input-otp", "url": "https://github.com/guilhermerodz/input-otp/issues/61", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
31942025
Use nodechecker.com to figure out if a module has passing tests We can sort them a bit lower if the tests are failing or don't exist Example: http://nodechecker.com/#info/gulp We may want to host our own since the interval at which they scan may not be fast enough and we can filter out only gulpplugin/gulpfriendly modules Closing this because the plugin page has been rewritten and integrated with our main website: https://github.com/gulpjs/gulpjs.github.io
gharchive/issue
2014-04-22T03:01:11
2025-04-01T04:34:26.936809
{ "authors": [ "contra", "phated" ], "repo": "gulpjs/plugins", "url": "https://github.com/gulpjs/plugins/issues/76", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1305724521
Local user? Hi, I was wondering if it is possible to use a local user to set the alarm, i.e not one linked to a hie-connect account with a simple pin code? As I just want HA to be able to set not but not able to unset the alarm for safety. Thanks James Hi, Unfortunately local users dont have ISAPI api access. So It can't be used in this integration. But I'll check a way to disable disarm functionality. Or, It can be forced to ask a password (with a complexity of your choice) when disarm is clicked at HA. Both needs some development. Let me know which will help you most. Then I can start development of it. I vote for keycode to use alarm control panel integration and alarm-panel which is included in hassio: https://www.home-assistant.io/dashboards/alarm-panel/ Hi Thanks. I would go for a key code which we can set to disarm would be most useful for the masses. Having looking at the user options, I can set a arm only account up on Hik-connect so this solves the need to disable the disarm function in first instances. I will trial over the weekend/next week with a separate Hik—connect account and update here for other peoples reference. Hi, So, I'll start implementing the keycode soon. Hello, Thanks you so much for having made the HACS integration of ax pro. I have switched my 100 devices to Home assistant last weekend because of that! Please don't make the keycode mandatory. The current integration is great. We can arm or disarm with geolacation for example. If you implement the keycode, please make it optional. Thank you. I used the following script to test `from hikaxpro import HikAxPro axpro = HikAxPro("IP_ADDRESSS", "HIK_CONNECT_USERNAME", "PASSWORD") res = axpro.zone_status() print(res) `
gharchive/issue
2022-07-15T08:12:21
2025-04-01T04:34:26.951106
{ "authors": [ "DejanBukovec", "gunkutzeybek", "nicolasv55", "obrien-james" ], "repo": "gunkutzeybek/hikaxpro_hacs", "url": "https://github.com/gunkutzeybek/hikaxpro_hacs/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
852473587
Please share your processed data set (via Google or Baidu Cloud Disk), thank you Please share your processed data set (via Google or Baidu Cloud Disk), thank you.For some reasons, it is more difficult to download data from the original site. I will only use the data for research Sure, I will release it soon. The datasets are more than 400 GB, so it will take a while to upload. Thank you very much. Data is released. I cannot access preprocessed data... Is Google or Baidu Cloud Disk available? thx thx thx... Hi Kingaza, I’m trying to resolve this problem. The dataset is too big (around 400 GB). I need find an appropriate place to host the dataset. Best, Pengfei From: kingaza @.> Sent: Wednesday, June 9, 2021 1:44 AM To: guopengf/FL-MRCM @.> Cc: guopengf @.>; State change @.> Subject: Re: [guopengf/FL-MRCM] Please share your processed data set (via Google or Baidu Cloud Disk), thank you (#3) I cannot access preprocessed data... Is Google or Baidu Cloud Disk available? thx thx thx... — You are receiving this because you modified the open/close state. Reply to this email directly, https://github.com/guopengf/FL-MRCM/issues/3#issuecomment-857394856 view it on GitHub, or https://github.com/notifications/unsubscribe-auth/AHUEVDZ76JIGWR4343DLZJLTR35SJANCNFSM42Q5DA3Q unsubscribe. https://github.com/notifications/beacon/AHUEVD3VXZPAHEYEI663JDTTR35SJA5CNFSM42Q5DA32YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOGMNM5KA.gif thx again~
gharchive/issue
2021-04-07T14:31:37
2025-04-01T04:34:26.978562
{ "authors": [ "guopengf", "kingaza", "zym1130232" ], "repo": "guopengf/FL-MRCM", "url": "https://github.com/guopengf/FL-MRCM/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1570102754
点击“提交”希望能有个图标表示提交成功,不然界面没有变化,一直感觉没有提交成功 希望大佬后续能加个图标,比如加个对号什么的,有个提示 想要這個功能 +1
gharchive/issue
2023-02-03T16:26:20
2025-04-01T04:34:26.979904
{ "authors": [ "Donge-wx", "trevim99" ], "repo": "guopenghui/obsidian-language-learner", "url": "https://github.com/guopenghui/obsidian-language-learner/issues/95", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
338501665
能不能尊重一下我的劳动成果? 应该署名参考我的github吧。。。这个分类和里面的内容和我的一模一样。 https://github.com/AweiLoveAndroid/CommonDevKnowledge/blob/master/interview/summary.md 不好意思,之前没看到你这个repo,从掘金搜了一下面试题集,整理了一下问题,写了下答案。 @AweiLoveAndroid 整合内容的目的只是为了让别人宣传你的名字吗? 我认为分享的意义是帮助他人,我知道你的初心不是不想帮助他人,但如果原文链接对你非常重要,我建议你以后不要公开发表,因为你不能保证所有人都不会犯错。 谢绝撕逼,世界和平。我只是搜集了各种面试题,并把自己博客的内容作为答案放在了里面,本来是自己备忘用的,因为是MIT协议,被鸿洋发到公众号了。我把所有的网页浏览记录都加上了。 另外有两点: 1 我只是搜集问题,问题的答案来源于我的博客。 2 不要要搞个大新闻,在知乎上发个阿里人抄袭你是什么意思,这是我入职阿里之前写的东西,不要动不动就把公司扯上。
gharchive/issue
2018-07-05T09:45:58
2025-04-01T04:34:26.984228
{ "authors": [ "AweiLoveAndroid", "guoxiaoxing", "tcqq" ], "repo": "guoxiaoxing/android-interview", "url": "https://github.com/guoxiaoxing/android-interview/issues/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
312579365
Try to handle the case where a element is missing an r attribute I've run into an Excel spreadsheet where some <c> elements do not have address references. It looks like it's some kind of short hand where it's implicitly the next cell, and Excel opens the spreadsheet fine. Here's a pretty-printed snippet from the spreadsheet: <sheetData> <row r="1" spans="1:1" ht="18" customHeight="1"> <c r="A1" s="1" t="s"> <v>0</v> </c> </row> <row r="2" spans="1:1" ht="12.75" customHeight="1"> <c r="A2" s="2" t="s"> <v>1</v> </c> </row> <row r="3" spans="1:1" ht="12.75" customHeight="1"> <c r="A3" s="2" t="s"> <v>2</v> </c> </row> <row r="5" spans="1:12" ht="12.75" customHeight="1"> <c r="A5" s="4" t="s"> <v>51</v> </c> <c s="4" t="s"> <v>52</v> </c> <c s="4" t="s"> <v>53</v> </c> <c s="4" t="s"> <v>54</v> </c> ... Exceljs breaks with: (node:3101) TypeError: Cannot read property 'match' of undefined at Object.decodeAddress (/Users/andreaslind/work/exceljs/lib/utils/col-cache.js:93:21) at value.cells.forEach.cellModel (/Users/andreaslind/work/exceljs/lib/doc/row.js:325:46) at Array.forEach (<anonymous>) at module.exports.set model [as model] (/Users/andreaslind/work/exceljs/lib/doc/row.js:319:17) at model.rows.forEach.rowModel (/Users/andreaslind/work/exceljs/lib/doc/worksheet.js:595:17) at Array.forEach (<anonymous>) at module.exports._parseRows (/Users/andreaslind/work/exceljs/lib/doc/worksheet.js:592:16) at module.exports.set model [as model] (/Users/andreaslind/work/exceljs/lib/doc/worksheet.js:606:10) at value.worksheets.forEach.worksheetModel (/Users/andreaslind/work/exceljs/lib/doc/workbook.js:197:23) at Array.forEach (<anonymous>) (The line numbers might be off, as I'm currently running an older version, unfortunately) I'm not sure the enclosed solution is per spec. Maybe someone else knows more? And yes, it's actually the same spreadsheet that caused https://github.com/guyonroche/exceljs/pull/536 -- I'll follow up if I'm able to find out how it was generated. I just learned that the Excel file in question is an export from Xero, which is a piece of accounting software.
gharchive/pull-request
2018-04-09T15:32:53
2025-04-01T04:34:27.026480
{ "authors": [ "papandreou" ], "repo": "guyonroche/exceljs", "url": "https://github.com/guyonroche/exceljs/pull/537", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
164661792
almost ready 基本写完了 实现了内置的校验函数 datetime 和 date可能会有bug unicode,dict,list这3个不用,dict,list在schema.py里面实现了,unicode去掉,不打算支持python2。 还差个枚举enum add_validater,remove_validater也不需要,去掉。
gharchive/pull-request
2016-07-09T10:45:25
2025-04-01T04:34:27.028571
{ "authors": [ "guyskk", "vibiu" ], "repo": "guyskk/validater", "url": "https://github.com/guyskk/validater/pull/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2533041182
会员视频无法下载 https://www.bilibili.com/bangumi/play/ep835859?spm_id_from=333.999.0.0 例如上面的,显示解析错误 登录的账号是 会员账号吗 😂 我不是会员账号没试过下载会员视频 肯定是大会员账号,不只是会员视频,就像是这种普通视频,只要地址是这种格式,就解析错误。 https://www.bilibili.com/bangumi/play/ep835674?spm_id_from=333.999.0.0&from_spmid=666.25.episode.0 非会员的这种格式已于v0.0.9修复, 会员的视频 我这边就无法测了有空的话麻烦帮忙测下😂 会员视频解析错误:Error: TypeError: Cannot read properties of undefined (reading 'duration') 肯定是大会员账号,不只是会员视频,就像是这种普通视频,只要地址是这种格式,就解析错误。 bilibili.com/bangumi/play/ep835674?spm_id_from=333.999.0.0&from_spmid=666.25.episode.0 如果用非会员的连接弹窗后 点会员的那一集下载呢 会报错吗 肯定是大会员账号,不仅仅是会员视频,就像是这种普通视频,只要地址是这种格式,就解析错误。 bilibili.com/bangumi/play/ep835674?spm_id_from=333.999.0.0&from_spmid=666.25.episode.0 如果用非会员的连接弹窗后点会员的那一套下载呢会报错吗 会卡死 试下用这个版本 看看 https://github.com/gxr404/BilibiliVideoDownloadFork/releases/tag/v0.0.10 我这边借了个号 试了下 没啥问题 试下用这个版本看看https://github.com/gxr404/BilibiliVideoDownloadFork/releases/tag/v0.0.10 我借了个号试了下没啥问题 是可以下载了,但是地址不对,显示的多P视频就不是视频本身 https://www.bilibili.com/bangumi/play/ep835914?spm_id_from=333.999.0.0 没懂啥意思 该会员视频链接 是个合集 含其他多p视频 所以填入 会员视频和 非会员视频 都会弹窗,如果只需下其中某一集 弹窗里选择即可 没懂啥意思会员视频链接是个合集含其他多p视频所以填入会员视频的和非会员视频都会弹窗,如果只需下其中一个集弹窗里选择即可 需要的视频就不在多p列表,你打开看一下,要下载是右侧列表里的,而不是上面的正片 你是指 要下载的 红框内的 而不是 蓝框内的? 这个有点特殊😂 你是指要下载的红框内的而不是蓝框内的? 这个有点特殊😂 对,应该是B站隐藏了下方视频的真实地址,上面总挂着3个正片 可以下载了,但是还有一点小瑕疵, https://www.bilibili.com/bangumi/play/ep835859?spm_id_from=333.999.0.0 这个视频有4K版本,但是上面的可选分辨率最高只有1080P高码率 大概看了下 如果使用 仅支持1080P高码率的视频链接 打开弹窗 选项只有1080P高码率 如果使用 支持4k的视频链接 打开弹窗 就有 选项4k OK,完美解决,感谢大神! 如果选择4k下载 视频仅支持1080高码率 则 会自动降级到1080高码率(降级功能一直都有), 只是现在v0.0.11版本显示有问题 没显示视频真实的分辨率, 在v0.0.12修复了
gharchive/issue
2024-09-18T08:14:58
2025-04-01T04:34:27.090052
{ "authors": [ "gxr404", "youzhui" ], "repo": "gxr404/BilibiliVideoDownloadFork", "url": "https://github.com/gxr404/BilibiliVideoDownloadFork/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2471768721
🛑 Undertale社区档案馆 - UTCARC is down In bbae30c, Undertale社区档案馆 - UTCARC (https://arc.utcwiki.com) was down: HTTP code: 0 Response time: 0 ms Resolved: Undertale社区档案馆 - UTCARC is back up in 90289c7 after 11 minutes.
gharchive/issue
2024-08-18T03:24:24
2025-04-01T04:34:27.104013
{ "authors": [ "gzombiejun" ], "repo": "gzombiejun/upptime", "url": "https://github.com/gzombiejun/upptime/issues/592", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1315688650
Require conda package [ ] Recipe to build conda package [ ] Step to build conda package [ ] Publish conda package to h2oai channel superseded by https://github.com/h2oai/authn-py/issues/72
gharchive/issue
2022-07-23T15:51:08
2025-04-01T04:34:27.116405
{ "authors": [ "Mathanraj-Sharma", "zoido" ], "repo": "h2oai/authn-py", "url": "https://github.com/h2oai/authn-py/issues/31", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
199415331
Error on gradle I got this error when I want to use this library and put it on gradle: Error:Could not GET 'https://repos.zeroturnaround.com/nexus/content/repositories/zt-public-releases/com/zeroturnaround/jrebel/android/jr-android-gradle/maven-metadata.xml'. Received status code 401 from server: Unauthorized Enable Gradle 'offline mode' and sync project What should I do to resolve it? Hi. That is not an issue of this library. It seems that the maven-metadata.xml file is protected with basic auth. I guess the file is used by JRabel plugin, so try disabling it. Thanks.
gharchive/issue
2017-01-08T10:50:49
2025-04-01T04:34:27.250139
{ "authors": [ "h6ah4i", "yasaman93" ], "repo": "h6ah4i/android-advancedrecyclerview", "url": "https://github.com/h6ah4i/android-advancedrecyclerview/issues/335", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
509225401
Define list of packages to download As part of https://github.com/habitat-sh/habitat/issues/6902 we want to provide some pre-defined lists of packages. Note: the files here use the '#' comment character and whitespace for clarity; and won't work with habitat until issue https://github.com/habitat-sh/habitat/issues/7040 is addressed. Background: Historically we provided a tarball 'LATEST.tgz' that contained all the packages in core for all targets. That amounted to about 15GB, and was both too much and too little, in that many of the packages weren't needed, and for many patterns (effortless) other origins were needed. This is an initial attempt to provide some starter lists of packages. The basic naming pattern is TASK_ARCH_CHANNEL. This due to a limitation of the input file format. The simple newline separated list of package idents doesn't allow for specification of channel or target architecture inline, so we're using a file naming convention to The current tasks are builder (setting up an on prem builder) core_deps (a reduced starter set from core with common build time deps) core_full (everything for a particular architecture) effortless (starter set for the effortless pattern) Each is broken out by the architecture and channel required; to complete some two downloads, once from stable and once from unstable will be required. For example, to get the complete effortless infrastructure for linux, hab pkg download --download-directory download_pkgs --channel=unstable --target x86_64-linux --file quickstart_lists/effortless_x86_64-windows_unstable hab pkg download --download-directory download_pkgs --channel=stable --target x86_64-linux --file quickstart_lists/effortless_x86_64-windows_stable Closes #6902 Signed-off-by: Mark Anderson mark@chef.io A few minor comments, overall looking good @markan
gharchive/pull-request
2019-10-18T17:45:51
2025-04-01T04:34:27.285655
{ "authors": [ "chefsalim", "markan" ], "repo": "habitat-sh/on-prem-builder", "url": "https://github.com/habitat-sh/on-prem-builder/pull/210", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1897276538
📓 Пользователь может авторизоваться Пользовательская история и сценарии Функция: Я, как зарегистрированный пользователь Хочу иметь возможность войти в систему Чтобы начать пользоваться функционалом платформы Сценарий: Успешная авторизация пользователя Допустим существует зарегистрированный пользователь с логином 'test@example.com' и паролем '12345' Когда пользователь осуществляет авторизацию с использованием логина 'test@example.com' и пароля '12345' Тогда пользователю выдается токен авторизации Сценарий: Неудачная авторизация пользователя Допустим существует зарегистрированный пользователь с логином 'test@example.com' и паролем '12345' Когда пользователь осуществляет авторизацию с использованием логина 'test@example.com' и пароля '54321' Тогда пользователь получает ошибку авторизации Дополнительная информация No response Функционал авторизации делаю выключенным, так как пока у нас нет развернутого Kratos на стендах
gharchive/issue
2023-09-14T20:40:37
2025-04-01T04:34:27.287347
{ "authors": [ "picolino" ], "repo": "habralab/garnet-team", "url": "https://github.com/habralab/garnet-team/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2414882311
cuteOneLineFlower-emma Generates a three-petal colored flower. Uses random seed value to move the position of the flower, stem, and leaves, as well as change the shape of the petals slightly. This is my first coding project and I'm a beginner coder :) [x] I have read the steps to getting a blot [x] I am submitting art that... [x] is algorithmically generated (will change each time the program is run) [x] is drawable on a blot (fits in the work area & doesn't overlap too much) [x] is original (not copied from somewhere else) [ ] Optional, if you used a tutorial or based your art on something else, please include the link here: [ ] Optional, if you remixed this from something else, mention it here: Seeing as this hasn't been updated in over a month, I'm going to close the PR for now. You're free to make a new one in the future with the same artwork if you'd like! Reach out to me @ alexren on the slack if you have any questions!
gharchive/pull-request
2024-07-18T00:11:18
2025-04-01T04:34:27.292981
{ "authors": [ "emma-x1", "qcoral" ], "repo": "hackclub/blot", "url": "https://github.com/hackclub/blot/pull/710", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
646777284
update scrappy user on profile change Updates github and website on profile change This will update the profile picture every time anything in the user's profile changes instead of only when the profile picture is updated. Cool, this looks great! Merging now, will test once merged and fix any bugs that arise. Thanks so much 🙌
gharchive/pull-request
2020-06-27T22:12:54
2025-04-01T04:34:27.297820
{ "authors": [ "MatthewStanciu", "saharshy29" ], "repo": "hackclub/scrappy", "url": "https://github.com/hackclub/scrappy/pull/32", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2486220562
[common] 해커톤 앱 제출 제목 팀 앱 제출 1차 팀 이름 common 팀 리포지토리 https://github.com/hackersground-kr/hg-common-hackergroud IntelliJ 닫기 버튼을 눌러 초기 화면으로 어떻게 가는지 모르겠어요. /invalid
gharchive/issue
2024-08-26T08:04:39
2025-04-01T04:34:27.306073
{ "authors": [ "clcok", "yeseong0412" ], "repo": "hackersground-kr/hackers-ground", "url": "https://github.com/hackersground-kr/hackers-ground/issues/951", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2011254203
Texture the player prefab The player prefabs needs to be UV-Mapped and get some textures Done, at least partially. What remains is improvements.
gharchive/issue
2023-11-26T23:07:54
2025-04-01T04:34:27.306857
{ "authors": [ "Fueredoriku", "toberge" ], "repo": "hackerspace-ntnu/Red-Planet-Rampage", "url": "https://github.com/hackerspace-ntnu/Red-Planet-Rampage/issues/336", "license": "BSD-4-Clause", "license_type": "permissive", "license_source": "github-api" }
266612187
BUY TICKETS INDIVIDUALLY No way to distinguish if ppl buy multiple tickets so changed policy to only allowing individual purchases. Is this enforced in the event brite? No, people can order multiple tickets. I could modify the Eventbrite to only allow one ticket
gharchive/pull-request
2017-10-18T19:22:13
2025-04-01T04:34:27.382213
{ "authors": [ "casey-chow", "dfan97" ], "repo": "hackprinceton/static", "url": "https://github.com/hackprinceton/static/pull/22", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
545080314
Error in opening the audio file As you have mentioned in the document, I replaced the path of the mp3 file. But it is showing the following error: Traceback (most recent call last): File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media\codecs\wave.py", line 59, in init self._wave = wave.open(file) File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\wave.py", line 510, in open return Wave_read(f) File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\wave.py", line 164, in init self.initfp(f) File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\wave.py", line 131, in initfp raise Error('file does not start with RIFF id') wave.Error: file does not start with RIFF id During handling of the above exception, another exception occurred: Traceback (most recent call last): File "voice_based_email_for_blind.py", line 26, in music = pyglet.media.load(ttsname, streaming = False) File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media_init_.py", line 143, in load raise first_exception File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media_init_.py", line 133, in load loaded_source = decoder.decode(file, filename, streaming) File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media\codecs\wave.py", line 111, in decode return StaticSource(WaveSource(filename, file)) File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media\codecs\wave.py", line 61, in init raise WAVEDecodeException(e) pyglet.media.codecs.wave.WAVEDecodeException: file does not start with RIFF id Please help me to solve this issue . @exist-world It seems like you have an invalid or unsupported audio file. Recall that only FLAC, AIFF, and RIFF WAV files are supported. If it's one of those, and it's still not working, please let me know. May I know your os? Try this. Read the file with librosa, then convert it back to a temporary .wav file. Then read it back with the 'wave' package. https://stackoverflow.com/questions/25672289/failed-to-open-file-file-wav-as-a-wav-due-to-file-does-not-start-with-riff-id/57349558#57349558 thanks for ur suggestion i will try and tell My os is windows 10 Now that audio file is working...but choosing option 1 gives me following error... Traceback (most recent call last): File "voice_pro.py", line 107, in mail.login('emailID','pswrd') #login part File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\smtplib.py", line 730, in login raise last_exception File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\smtplib.py", line 721, in login initial_response_ok=initial_response_ok) File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\smtplib.py", line 642, in auth raise SMTPAuthenticationError(code, resp) smtplib.SMTPAuthenticationError: (535, b'5.7.8 Username and Password not accepted. Learn more at\n5.7.8 https://support.google.com/mail/?p=BadCredentials i17sm4151845pfr.67 - gsmtp') And chosing option 2 gives me the following error Traceback (most recent call last): File "voice_pro.py", line 89, in if int(text) == 1: ValueError: invalid literal for int() with base 10: 'Tu' @exist-world bro, Here [mail.login('emailID','pswrd') #login part] You have to put your email id and password for login . After that if any error occurs then please tell me. yeah ..thank you !..Now 1 st option is working but in victim id i gave receiver mail id...it shows mail sent But receiver didnt get that mail And giving 2 nd option shows me the following error Traceback (most recent call last): File "voice_pro.py", line 89, in if int(text) == 1: ValueError: invalid literal for int() with base 10: 'Tu' Help me to solve this one..! @exist-world Its because of 'Tu' not '2'. I mean script recognise the 'Tu' not 2 and that's why it fetch the error. Just edit it and add [ if int(text) == 2 or text == "tu": ] in the second if statement not the first one. Now, I really don't know why u have not received the mail in inbox. Have you checked in spam folder? If not then use your primary email id in both as a sender and receiver email id. No i dont use any of that kind..its okay...but whenever i use option 2 its seems like its checking for option 1 only...its shows folln error You said : Tu Traceback (most recent call last): File "voice_pro.py", line 90, in if int(text)==1 or text == 'one': ValueError: invalid literal for int() with base 10: 'Tu' @exist-world Sorry for late response. Paste those with folln codes: if text == '1' or text == 'One' or text == 'one': elif text == '2' or text == 'tu' or text == 'two' or text == 'Tu' or text == 'to' or text == 'To' : Soon I will update the latest one. As I told you before it was almost 3 years old project. Okay ..pls upload the latest one as soon as possible...Thanks in advance i have one doubt...you gave mail id of the receiver as static in code...for getting dynamic input from the user for mail id how to use ?? Please help quickly to solve this out.. @afuafu21 It arises some minor problem. I have added this feature in the latest one. But the problem is I have not enough time to test it. So, Its getting a bit late to upload. And there are another problem, if someone's email id is alphanumeric then the problem occurs, till now it should be in alphabet only. Ohh Its okay...When u will upload the latest one.please upload soon or if u don't mind send it to my mail...Because this month 18 is the last date for my project submission and review....if u upload soon it is so helpful for me....because of your contribution only i did this much in that project Thanks a lot... On Thu, Mar 5, 2020, 10:03 AM Sayak Naskar notifications@github.com wrote: @afuafu21 https://github.com/afuafu21 It arises some minor problem. I have added this feature in the latest one. But the problem is I have not enough time to test it. So, Its getting a bit late to upload. And there are another problem, if someone's email id is alphanumeric then the problem occurs, till now it should be in alphabet only. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/hacky1997/voice-based-email-for-blind/issues/5?email_source=notifications&email_token=AOU64DHYZ75QVF2P7C4HO2TRF4TRDA5CNFSM4KCPYZU2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEN3VMVA#issuecomment-595023444, or unsubscribe https://github.com/notifications/unsubscribe-auth/AOU64DFXQNAT7TVKYMEEOY3RF4TRDANCNFSM4KCPYZUQ . @afuafu21 Please provide your email id. I have not tested it yet. Okay Thank you this is my mail id-mini20pro20@gmail.com On Fri, Mar 6, 2020, 4:30 PM Sayak Naskar notifications@github.com wrote: @afuafu21 https://github.com/afuafu21 Please provide your email id. I have not tested it yet. — You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub https://github.com/hacky1997/voice-based-email-for-blind/issues/5?email_source=notifications&email_token=AOU64DHTJUFKAGJRIG2UQIDRGDJWDA5CNFSM4KCPYZU2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEOA64MA#issuecomment-595717680, or unsubscribe https://github.com/notifications/unsubscribe-auth/AOU64DFMGXSGFUXFRD6ABPDRGDJWDANCNFSM4KCPYZUQ . Issue has been resolved.
gharchive/issue
2020-01-03T17:14:41
2025-04-01T04:34:27.407192
{ "authors": [ "afuafu21", "exist-world", "hacky1997" ], "repo": "hacky1997/voice-based-email-for-blind", "url": "https://github.com/hacky1997/voice-based-email-for-blind/issues/5", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
437048585
Hacl* library package How do I include the Hacl* library in the lib folder when extracting my code to OCaml? Currently I am stuck with this, which can extract successfully for F* codes only using the F* libraries. fstar.exe --z3cliopt 'timeout=600000' --use_hints --use_hint_hashes --odir out --codegen OCaml <modulename>.fst OCAMLPATH="../../fstar/bin" ocamlfind opt -package fstarlib -linkpkg -g out/<modulename>.ml -o test.exe After I perform the fstar --codegen OCaml --include lib on my file while including the Hacl* library, I used the following Makefile to get the .cmxa hacl library. FILES= \ FStar_All.ml \ FStar_BitVector.ml \ FStar_Calc.ml \ FStar_Exn.ml \ FStar_Heap.ml \ FStar_List_Tot.ml \ FStar_List_Tot_Base.ml \ FStar_List_Tot_Properties.ml \ FStar_Math_Lemmas.ml \ FStar_Math_Lib.ml \ FStar_Mul.ml \ FStar_Pervasives.ml \ FStar_Pervasives_Native.ml \ FStar_PredicateExtensionality.ml \ FStar_Preorder.ml \ FStar_PropositionalExtensionality.ml \ FStar_Seq.ml \ FStar_Seq_Base.ml \ FStar_Seq_Properties.ml \ FStar_ST.ml \ FStar_StrongExcludedMiddle.ml \ FStar_TSet.ml \ FStar_UInt.ml \ FStar_UInt8.ml \ FStar_UInt16.ml \ FStar_UInt32.ml \ FStar_UInt64.ml OBJS=$(FILES:.ml=.o) CMX=$(FILES:.ml=.cmx) all: hacllib.cmxa hacllib.cmxa: OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -thread -w -58 -c $(FILES) OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -a $(CMX) -o hacllib.cmxa clean: rm -f *.cmi *.cmo *.cmx *.cmxa *.o *.a *~ However, there was this error: Error: The files fStar_BitVector.cmi and fStar_Seq_Base.cmi make inconsistent assumptions over interface FStar_Pervasives_Native make: *** [Makefile:36: hacllib.cmxa] Error 2 Can anyone enlighten me on whether I am on the right track and how to fix this? You appear to have inconsistent state between fstar and hacl. Can you clean both repos and recompile from scratch? On Mon, Apr 29, 2019, 10:56 jamesbarne notifications@github.com wrote: After I perform the fstar --codegen OCaml --include lib on my file while including the Hacl* library, I used the following Makefile to get the .cmxa hacl library. FILES= FStar_All.ml FStar_BitVector.ml FStar_Calc.ml FStar_Exn.ml FStar_Heap.ml FStar_List_Tot.ml FStar_List_Tot_Base.ml FStar_List_Tot_Properties.ml FStar_Math_Lemmas.ml FStar_Math_Lib.ml FStar_Mul.ml FStar_Pervasives.ml FStar_Pervasives_Native.ml FStar_PredicateExtensionality.ml FStar_Preorder.ml FStar_PropositionalExtensionality.ml FStar_Seq.ml FStar_Seq_Base.ml FStar_Seq_Properties.ml FStar_ST.ml FStar_StrongExcludedMiddle.ml FStar_TSet.ml FStar_UInt.ml FStar_UInt8.ml FStar_UInt16.ml FStar_UInt32.ml FStar_UInt64.ml OBJS=$(FILES:.ml=.o) CMX=$(FILES:.ml=.cmx) all: hacllib.cmxa hacllib.cmxa: OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -thread -w -58 -c $(FILES) OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -a $(CMX) -o hacllib.cmxa clean: rm -f *.cmi *.cmo *.cmx *.cmxa *.o *.a *~ However, there was this error: Error: The files fStar_BitVector.cmi and fStar_Seq_Base.cmi make inconsistent assumptions over interface FStar_Pervasives_Native make: *** [Makefile:36: hacllib.cmxa] Error 2 Can anyone enlighten me on whether I am on the right track and how to fix this? — You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/project-everest/hacl-star/issues/154#issuecomment-487505105, or mute the thread https://github.com/notifications/unsubscribe-auth/ABFUVS4C32ULZXFOZK3FGH3PS2Z3LANCNFSM4HIKOBBA . I recomplied the fstar and hacl from scratch but am still facing the same problem. When attempting to recompile kremlin too, the build failed with a similar error regarding the FStar_Pervasives_Native module. Rebuilding kremlin Running: build_kremlin make: Entering directory '/cygdrive/d/lbeverest/everest/kremlin' ocamlbuild -I src -I lib -I parser -I kremlib -use-menhir -use-ocamlfind -classic-display -menhir "menhir --infer --explain" Kremlin.native Tests.native ln -sf Kremlin.native krml make -C kremlib make[1]: Entering directory '/cygdrive/d/lbeverest/everest/kremlin/kremlib' D:/lbeverest/everest/FStar/bin/fstar.exe --record_hints --use_hints --use_two_phase_tc true --odir .extract --cache_checked_modules --cmi --already_cached 'FStar -FStar.Kremlin.Endianness LowStar' FStar.Kremlin.Endianness.fst && \ touch FStar.Kremlin.Endianness.fst.checked D:\lbeverest\everest\FStar\ulib\FStar.Pervasives.fst(0,0-0,0): (Warning 241) Unable to load D:\lbeverest\everest\FStar\bin\..\ulib\FStar.Pervasives.fst.checked since checked file D:\lbeverest\everest\FStar\bin\..\ulib\FStar.Pervasives.fst.checked is corrupt; will recheck D:\lbeverest\everest\FStar\ulib\FStar.Pervasives.fst (suppressing this warning for further modules) (Error 317) Expected D:\lbeverest\everest\FStar\ulib\FStar.Pervasives.fst to already be checked 1 error was reported (see above) make[1]: *** [Makefile:58: FStar.Kremlin.Endianness.fst.checked] Error 1 make[1]: Leaving directory '/cygdrive/d/lbeverest/everest/kremlin/kremlib' make: *** [Makefile:28: kremlib] Error 2 make: Leaving directory '/cygdrive/d/lbeverest/everest/kremlin' ================================================================================ FAILURE: build failed for kremlin We currently don't build an OCaml library for Hacl* lib like we do for FStar ulib. Do you have a use case where this would be more convenient than extracting the individual files you depend on? The first issue you encountered about inconsistent assumptions over interface FStar_Pervasives_Native is due to you re-extracting and re-compiling modules in the FStar namespace and linking them with fstarlib, which includes all of them. The solution is to compile only the modules not already included in fstarlib, which would also save you some time. Here's a simple example of how to build an OCaml binary from an FStar module that depends on Hacl* lib: module Test open FStar.All open Lib.IntTypes val test: unit -> ML uint32 let test () = IO.print_string "Hello world!\n"; u32 0 .PHONY: all test clean FSTAR_HOME ?= ../../FStar HACL_HOME ?= .. FSTAR_FLAGS = $(OTHERFLAGS) \ --cmi \ --cache_checked_modules \ --already_cached "'Prims+FStar+LowStar+Lib'" \ --include $(HACL_HOME)/lib FSTAR = $(FSTAR_HOME)/bin/fstar.exe $(FSTAR_FLAGS) ROOTS = Test.fst all: rm -f .depend && $(MAKE) .depend $(MAKE) test # 1. Generation of .ml files # - generate the F* dependency graph via `fstar --dep full` # - verify every F* file in parallel to generate .checked files # - extract each .checked file into a .ml file in parallel .depend: $(FSTAR) --dep full $(ROOTS) --extract '* -Prims -LowStar -FStar' > $@ include .depend %.checked: | .depend $(FSTAR) $< && \ touch $@ %.ml: | .depend $(FSTAR) --codegen OCaml \ --extract_module $(basename $(notdir $(subst .checked,,$<))) \ $(notdir $(subst .checked,,$<)) && \ touch $@ # 2. Compilation OCAMLOPT=OCAMLPATH="$(FSTAR_HOME)/bin" ocamlfind opt -package fstarlib -linkpkg %.cmx: $(OCAMLOPT) -c $< -o $@ Test.cmx: @echo 'let _ = test()' >> Test.ml $(OCAMLOPT) -c $< -o $@ test.exe: Test.cmx $(OCAMLOPT) -o test.exe $(subst .ml,.cmx,$(ALL_ML_FILES)) test: test.exe ./test.exe clean: rm -f *.ml *.cmi *.cmo *.cmx *.cmxa *.o *.a *~ *.checked *.exe This assumes you have already verified and generated .checked files for all modules in lib. It will verify the rest of the dependencies (in this case just Test), extract Test and the modules in lib to OCaml and build a binary from them and fstarlib. All this can be done in parallel using make -j. This is a sketch of how you would do it building an intermediate hacllib.cmxa: # 2. Compilation OCAMLOPT_=OCAMLPATH="$(FSTAR_HOME)/bin" ocamlfind opt -package fstarlib OCAMLOPT=$(OCAMLOPT_) -linkpkg %.cmx: $(OCAMLOPT) -c $< -o $@ Test.cmx: @echo 'let _ = test()' >> Test.ml $(OCAMLOPT) -c $< -o $@ LIB_FILES = $(filter Lib_%,$(ALL_ML_FILES)) OTHER_FILES = $(filter-out Lib_%,$(ALL_ML_FILES)) hacllib.cmxa: $(subst .ml,.cmx,$(LIB_FILES)) $(OCAMLOPT_) -a -o $@ Lib_IntTypes.cmx test.exe: hacllib.cmxa $(subst .ml,.cmx,$(OTHER_FILES)) $(OCAMLOPT) -o test.exe $< $(subst .ml,.cmx,$(OTHER_FILES)) It works for this simple example but it's only a sketch because some modules in lib have dependencies on KreMLin modules that aren't in the Lib namespace and won't be appended to hacllib.cmxa. We now produce obj/libhaclml.cmxa as part of the build. Assuming you have a successful build of HACL*, then you should be able to include obj/ to get the benefits of the extracted .ml files, and you should be able to link against libhaclml.cmxa to obtain the final executable. This is how we assemble libhaclml.cmxa: 1069 obj/libhaclml.cmxa: $(filter-out $(HACL_HOME)/obj/Meta_Interface.cmx,$(ALL_CMX_FILES)) 1070 # JP: doesn't work because a PPX is prepended for some reason 1071 #ocamlfind mklib -o haclml -package fstarlib -g -I $(HACL_HOME)/obj $(addprefix $(HACL_HOME)/obj/*.,cmo cmx ml o) 1072 ocamlfind opt -a -o $@ -package fstarlib -g -I $(HACL_HOME)/obj $^ To use libhaclml.cmxa as a client, you would need to link against both fstarlib and libhaclml and pass -linkpkg to ocamlfind opt in order to produce a final executable. Please reopen this if still relevant, thanks.
gharchive/issue
2019-04-25T07:30:21
2025-04-01T04:34:27.425817
{ "authors": [ "jamesbarne", "karthikbhargavan", "msprotz", "s-zanella" ], "repo": "hacl-star/hacl-star", "url": "https://github.com/hacl-star/hacl-star/issues/154", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2506479112
grouping test cases and categorisation Hi @haesleinhuepf I was going through the preprint and one thought I had was grouping the test cases under categories. We get an overall view of how well LLMs perform but lose the granularity on whether LLMs perform well/worse on certain tasks and not on others.. For example, with our test cases perhaps grouping can be like: Quantification/Measurements: test cases that measure or count can be under quantificaiton/measurements statistical analysis: t test, pairwise correlation Morphological operations: binary close, skeleton, expand labels etc.. It may give an idea of where we need more or less test cases as well. I remember you had a preprint on ontologies and standards for bioimage analysis. Perhaps that can be used as a reference. Cheers Pradeep Hey @pr4deepr , great idea! This categorization is obviously a subjective task. We could automate this and make it more objective using an LLM; a text-classification LLM. Do you by chance have experience with this? Cheers, Robert No, I do not. Would you put it through chatgpt for example to get a first pass? If I copied the readme in the main repo containing description of the current test cases and used the question in Chat-GPT (GPT-4o): I have some python functions and each of them perform a specific operation in bioimage analysis. Classify them into categories based on their function and where they will fit in the image analysis pipeline. 1. Image Preprocessing These functions focus on preparing the image by modifying its format, reducing noise, or performing other transformations to enhance analysis. rgb_to_grey_image_transform rotate_image_by_90_degrees crop_quarter_image transpose_image_axes remove_noise_edge_preserving subtract_background_tophat mask_image remove_labels_on_edges remove_small_labels expand_labels_without_overlap 2. Image Enhancement Functions that enhance specific features of the image, such as sharpening edges or enhancing contrast. detect_edges convolve_images deconvolve_image remove_noise_edge_preserving 3. Segmentation These functions segment images into meaningful regions or objects, usually through intensity-based methods. apply_otsu_threshold_and_count_postiive_pixels region_growing_segmentation workflow_segmentation_counting workflow_watershed_segmentation_correction_measurement workflow_segmentation_measurement_summary workflow_segment_measure_umap 4. Morphological Operations These functions apply morphological transformations to binary images for cleaning or refining shapes. binary_closing binary_skeleton label_binary_image_and_count_labels label_sequentially remove_labels_on_edges remove_small_labels 5. Quantification and Measurement Functions that quantify different properties of labeled regions or objects in the image. convex_hull_measure_area count_number_of_touching_neighbors measure_aspect_ratio_of_regions measure_intensity_of_labels measure_intensity_over_time measure_mean_image_intensity measure_pixel_count_of_labels measure_properties_of_regions extract_surface_measure_area measure_properties_of_regions count_objects_over_time count_overlapping_regions 6. Feature Extraction Functions that extract geometrical or intensity features from the image. fit_circle radial_intensity_profile convex_hull_measure_area measure_aspect_ratio_of_regions 7. File I/O Functions that handle loading, saving, or listing files. list_image_files_in_folder open_image_read_voxel_size open_image_return_dimensions open_nifti_image open_zarr 8. Statistical Analysis Functions for performing statistical analysis on datasets related to the images. bland_altman t_test pair_wise_correlation_matrix mean_std_column mean_squared_error 9. Pipeline/Workflow Automation These functions automate a sequence of steps, combining multiple image processing operations. workflow_batch_process_folder_count_labels workflow_batch_process_folder_measure_intensity workflow_segment_measure_umap workflow_segmentation_measurement_summary workflow_watershed_segmentation_correction_measurement 10. Miscellaneous Functions that don't fit directly into other categories but are useful for specific tasks. return_hello_world Awesome! I couldn't have done better ☺️ Now what do we do woth this? Shall we preserve it somewhere in the repository? I'm creating a branch with the new categories.. Plan will be to save this as a yaml file.. There are errors and repetitions above, so need to clean it up as well.. https://github.com/pr4deepr/human-eval-bia/tree/function_categorize repetitions above Well I think test-cases can be in multiple categories. Good point So, I've done an initial pass. Interesting findings: Morphological operations: GPT-4o >claude 3.5 Feature extraction: Claude3.5 >GPT-4o Statistical analysis: llama3 ==gpt3.5turbo >claude3.5 and GPT-4o The categorisation of course is important and if not done properly can misrepresent the results.. The function categorisation can be found here: https://github.com/pr4deepr/human-eval-bia/blob/function_categorize/demo/create_function_category_yaml.ipynb which saves is as a yaml file I can create a separate notebook for the data processing and graphing as its currently here: https://github.com/pr4deepr/human-eval-bia/blob/function_categorize/demo/summarize_by_case.ipynb.. Happy to create a PR, but wasn't sure if it should be to main.. Yes! I certainly need such a figure for talks, because showing the blue table for all test-cases doesn't fit on a slide. It could also be in the paper... Curious what @tischi says about this figure: What I'm a bit concerned about this the static list of categories in the other notebook. It could be a pain to maintain this mid/long term. Would it be possible to put them in a dataframe, and add some code that warns if a test-case is in no category? Or even better, code that uses gpt4-o to categorize tesr-cases that are in no category and then adding them to the dataframe? Regarding categorization, why not require some metadata tag to be present with each submitted test case? Maybe it's too late for this but we could add it to existing test cases as the numbers seem manageable. Doing it with an LLM would anyway require manual review of the outcome. Yet, we don't have any infrastructure for handling meta data of test-cases. I was hoping to fully automate this, so that minimal manual curation is necessary. In an earlier discussion, also categorizing code depending on its complexity was discussed. No matter how we do these things, I'd love to have a semi-automatic solution with minimal code/infrastructure to maintain. We can use the GPT4-o idea, but is there a way to have a seed or something similar to guarantee relatively similar responses? The categories change everytime I ask... OR we just need to really be specific on question we ask GPT.. The tagging could be left to the author of the test case given a choice of predefined categories. Then its should be a matter of reading the tags when compiling the results. If classification is automated with an LLM, the outcome is likely to change over time and with the LLM used. I think we would need a deterministic algorithm for this. @pr4deepr Exactly what I thought likely :) Ok, I leave the decision about this to you guys. Whatever works :-) I'm happy with solution from @jkh1 , i.e., having a few tags, and getting the author of new test cases to put those tags in their functions. We can have a few different tags for each category. This could be a requirement when submitting a new test cases. For existing functions, perhaps myself and @jkh1 could decide on categories & tags add them to existing test cases and decide where to define it submit a PR Cheers Pradeep put those tags in their functions Can you give an example how this could look like? Either in functions or in each notebook. I need to look at the code first. Will update it here Upon looking at the code again, I think we'll want to minimize any modifications to existing test functions & with creating yaml files for cases at this point. I propose we have all the categorisation information in a yaml file with: each function name as a key values are the categories. The categories can be: file_i_o image_preprocessing morphological_operations segmentation feature_extraction statistical_analysis workflow_automation oher example yaml file: create_umap feature_extraction workflow_segment_measure_umap segmentation workflow_automation feature_extraction I'm happy to go through existing test cases and create this yaml file.. When a test case PR is submitted, the yaml file will have to be modified to add the new function and category. The PR template will need to be modified. If the need arises we can expand the categories, but I feel like this should cover it. Yes, great idea! The PR template will need to be modified. We can also add some python code which tests if all test cases are in thus yaml.file. e.g. in create_cases.ipynb or as github worklfow. Sounds good to me. My intiial idea was to use notebook tags but I realized this may be more complicated to get at. I've made the necessary changes with commit history here: https://github.com/pr4deepr/human-eval-bia/commits/function_categorize/ Added a categorise functions yaml file Added a check in create_test_cases to verify all functions are present Code updated to plot by category updated PR template. wording may need changing If you are happy with this, I can open a PR. Let me know which branch you'd prefer. Awesome @pr4deepr , thanks for working on this! Yes, please send a PR!
gharchive/issue
2024-09-04T23:35:48
2025-04-01T04:34:27.488189
{ "authors": [ "haesleinhuepf", "jkh1", "pr4deepr" ], "repo": "haesleinhuepf/human-eval-bia", "url": "https://github.com/haesleinhuepf/human-eval-bia/issues/112", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
448852537
Loading progress in % Loading progress in % for multiple resources. Based off PR #6 made by @jiawenquan This is excellent It is a singleton class because being an Angular service.
gharchive/pull-request
2019-05-27T13:14:59
2025-04-01T04:34:27.490249
{ "authors": [ "haestflod", "jiawenquan" ], "repo": "haestflod/generic-3Dproduct-viewer", "url": "https://github.com/haestflod/generic-3Dproduct-viewer/pull/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
609888334
Add --grep option custom grep. why? dalfox is xss tool, but it can also be used to find other vulnerabilities. e.g ssti information leak etc.. so, i think need custom grep option it is last options for release :D Thats great bro, you should have that, but in case you dont, hope this helps for that or other project. patterns.json.txt And my SSTI playroom =] <%= 339 * 431 %> #{ 339 * 431 } #{set} ($run=339 * 431) $run ${339*431} <%=339*431%> ${{339*431}} {{339*431}} {{339*'3'}} [[339*431]] {{= 339*431 }} {#339*431} <# 339*431 > {@339*431} ${{"{{"}}339*431{{"}}"}} Hi @bsysop Thanks dude! Originally, I was going to let the user customize the response grep, but I think some pattern is okay to put it in a built-in pattern. I'll use it well! Thank you. Gooot it, as you wish bro, you could use it in other tool if you think is better! --grep argument input is the file? regex string..? and Log Format [G] Found via built-in grepping / payload: {339*431} , grep: 146109 1 line: 146109({"isSuccess":false,"errorMsg":"Parameter error! apps is null","error +> https://blahblha!~~~ [G] Found via custom grepping / payload: 'adf , grep: internal_domain://asdf 1 line: internal_domain://asdf~~({"isSuccess":false,"errorMsg":"Parameter error! apps is null","error +> https://blahblha!~~~ Thats awesome, i would love to test that and check new approaches! Function development is complete, only small pattern addition is required. follow issue below https://github.com/hahwul/dalfox/issues/60 Amazing mate!!!! Great job!
gharchive/issue
2020-04-30T12:07:03
2025-04-01T04:34:27.500782
{ "authors": [ "bsysop", "hahwul" ], "repo": "hahwul/dalfox", "url": "https://github.com/hahwul/dalfox/issues/58", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2349312796
Get master up to date master up to date for purpose of development Probably nothing bad will happen...
gharchive/pull-request
2024-06-12T17:20:38
2025-04-01T04:34:27.501694
{ "authors": [ "RichardStaszkiewicz" ], "repo": "haichangsi/WUT-TBD-PROJ1", "url": "https://github.com/haichangsi/WUT-TBD-PROJ1/pull/5", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
836047249
internal/graphics: Race condition at the vertices backend and copying at internal/graphicscommand Git bisect said c8b98f13fb06eee62d6eba2c37d891f86726e40d was the culprit. CC @wasedaigo (v *verticesBackend) slice was not concurrent safe. This should not be an issue in v2.0.0 or v1.x since this function was used only on browsers where there is one thread. OK this is not an easy issue: graphics.QuadVertex is called from the screen rendering (last=true) At the same time, graphics.QuadVertex is accessed from a goroutine in parallel The vertices backend is reset (head is set to 0) even though the vertices is not copied at the graphicscommand package. Then the vertices data is overwritten and broken. Found there is a potential race condition even with the restorable package. I think it is inevitable to use a simple allocation instead of using the backend, even on browsers... Reverting 9cb631e30f3cf7471d53ed9aa76afb42846b0918 should have the same effect. I'll backport this change. In v2.0 and v1.12, this issue was only on browsers. Also I'll (re)add mutex to vertexBackend.slice since there is no guarantee about context change in Wasm.
gharchive/issue
2021-03-19T14:18:38
2025-04-01T04:34:27.536414
{ "authors": [ "hajimehoshi" ], "repo": "hajimehoshi/ebiten", "url": "https://github.com/hajimehoshi/ebiten/issues/1546", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
286475817
examples/audio: Music should keep playing even the tab is background when IsRunnableInBackground is true Related to #272 When thinking examples/audio as a music player, it'd be nice to keep playing music This is no longer an issue. I don't know when this was fixed...
gharchive/issue
2018-01-06T10:39:06
2025-04-01T04:34:27.537526
{ "authors": [ "hajimehoshi" ], "repo": "hajimehoshi/ebiten", "url": "https://github.com/hajimehoshi/ebiten/issues/476", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2092603148
夜间模式下,部分页面的字符颜色显示异常 https://www.bilibili.com/v/popular/weekly/?num=251 這個頁面我不太想適配其實,之後寫個正則過濾掉防止影響 暫時先過濾了這頁面。。。過後適配
gharchive/issue
2024-01-21T14:11:56
2025-04-01T04:34:27.550650
{ "authors": [ "hakadao", "miracomangomanchuria" ], "repo": "hakadao/BewlyBewly", "url": "https://github.com/hakadao/BewlyBewly/issues/206", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1893168011
Hadhunna/adding metrics Contributing to the Azure SDK Features Added metrics to header. Will be used to track the number of requests and responses for each action. Other Changes Change version semantics to Major.Minor.Patch.PreRelease Can we add links to the work items in the description? Also, please run the following commands from C:\repos\azure-sdk-for-net: eng\scripts\Update-Snippets.ps1 eng\scripts\CodeChecks.ps1 -ServiceDirectory entra
gharchive/pull-request
2023-09-12T19:55:32
2025-04-01T04:34:27.561962
{ "authors": [ "HarmanDhunna", "hakimms" ], "repo": "hakimms/azure-sdk-for-net", "url": "https://github.com/hakimms/azure-sdk-for-net/pull/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
323423046
Allow for customized sudo env vaiables Whether or not this is a good idea, I want to do: halyard::allowed_env_variables: - DEBUG - PUPPET_ENV - PROFILE - SSH_AUTH_SOCK This let's me do that ❯ halyard Repo is unclean: /opt/halyard/repo 2018-05-15 19:50:48 Jareds-MacBook-Pro.local STARTING RUN 2018-05-15 19:50:54 Jareds-MacBook-Pro.local Notice: Compiled catalog for jareds-macbook-pro.local in environment production in 0.55 seconds 2018-05-15 19:50:55 Jareds-MacBook-Pro.local Notice: /Stage[main]/Halyard/Sudoers::Allowed_command[halyard_puppet]/File[/etc/sudoers.d/halyard_puppet]/content: content changed '{md5}8fd6def48b20bcf4bd829ac382b6dc46' to '{md5}7a6862829d4223eb8e6baf25c2f6f289' 2018-05-15 19:50:57 Jareds-MacBook-Pro.local Notice: Applied catalog in 2.74 seconds 2018-05-15 19:50:57 Jareds-MacBook-Pro.local ENDING RUN ❯ sudo cat /etc/sudoers.d/halyard_puppet # Allows halyard user to run puppet Cmnd_Alias HALYARD_PUPPET = /opt/halyard/repo/meta/halyard,/bin/sh -c /opt/halyard/repo/meta/halyard Defaults!HALYARD_PUPPET secure_path = /sbin:/bin:/usr/sbin:/usr/bin Defaults!HALYARD_PUPPET env_keep+=DEBUG Defaults!HALYARD_PUPPET env_keep+=PUPPET_ENV Defaults!HALYARD_PUPPET env_keep+=PROFILE Defaults!HALYARD_PUPPET env_keep+=SSH_AUTH_SOCK jaredledvina ALL=(root) NOPASSWD: HALYARD_PUPPET As for the 'why?', because I have my github SSH key on my yubikey, and my remote for halyard is my fork using ssh :) So the goal here is that when running puppet-run it'll run with your SSH agent so it can chat w/ github? @akerl - Yep, passing through SSH_AUTH_SOCK let's haylard, git pull successfully in my configuration.
gharchive/pull-request
2018-05-15T23:53:46
2025-04-01T04:34:27.594641
{ "authors": [ "akerl", "jaredledvina" ], "repo": "halyard/puppet-halyard", "url": "https://github.com/halyard/puppet-halyard/pull/6", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }