id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1936405649 | Need an Easy Way to Add Implied Automatic Modules to the Module Path (Parity with Maven)
tl;dr: I just want to add a few implied automatic modules to the module path without managing transitive dependencies.
To demonstrate, let me construct a very lightweight example module with three easily-met requirements:
The module includes a module-info.java file.
It uses Immutables to create @Value.Immutable interfaces (and generate their implementations).
It needs Google Guava as a dependency.
To repro, we can create a minimalistic Java file (note that the generated code will differ based on whether Guava is present or not):
package org.example;
import org.immutables.value.Value;
@Value.Immutable
public interface Empty {}
Here's the corresponding module-info.java:
module org.example {
requires static org.immutables.value;
requires com.google.errorprone.annotations;
requires jsr305;
}
(jsr305 is an implied automatic module generated by com.google.code.findbugs:jsr305.)
Let's try to get his code to build with Maven first, and then with Gradle.
Maven
I only have a passing knowledge of Maven, but it didn't take very long for me to generate a working pom.xml file:
<?xml version="1.0" encoding="UTF-8"?>
<project xmlns="http://maven.apache.org/POM/4.0.0"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<modelVersion>4.0.0</modelVersion>
<groupId>org.example</groupId>
<artifactId>maven-module</artifactId>
<version>1.0-SNAPSHOT</version>
<properties>
<maven.compiler.source>17</maven.compiler.source>
<maven.compiler.target>17</maven.compiler.target>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
</properties>
<dependencies>
<dependency>
<groupId>org.immutables</groupId>
<artifactId>value</artifactId>
<version>2.10.0</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>com.google.guava</groupId>
<artifactId>guava</artifactId>
<version>32.1.2-jre</version>
</dependency>
</dependencies>
</project>
Now granted, one could criticize Maven for including an implied automatic module on the module path, but this build logic just works.
Gradle
This won't build out-of-the-box; Gradle will complain that it can't find module jsr305. Now let's 1) temporarily get rid of the module-info.java file, and 2) just add the plugin to the build.gradle.kts:
plugins {
`java-library`
id("org.gradlex.extra-java-module-info") version "1.4.2"
}
repositories {
mavenCentral()
}
dependencies {
compileOnly("org.immutables:value-annotations:2.10.0")
annotationProcessor("org.immutables:value:2.10.0")
implementation("com.google.guava:guava:32.1.2-jre")
}
This actually will not build either:
Execution failed for task ':lib:compileJava'.
> Could not resolve all files for configuration ':lib:compileClasspath'.
> Failed to transform failureaccess-1.0.1.jar (com.google.guava:failureaccess:1.0.1) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}.
> Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.guava/failureaccess/1.0.1/1dcf1de382a0bf95a3d8b0849546c88bac1292c9/failureaccess-1.0.1.jar.
> Not a module and no mapping defined: failureaccess-1.0.1.jar
> Failed to transform jsr305-3.0.2.jar (com.google.code.findbugs:jsr305:3.0.2) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}.
> Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.code.findbugs/jsr305/3.0.2/25ea2e8b0c338a877313bd4672d3fe056ea78f0d/jsr305-3.0.2.jar.
> Not a module and no mapping defined: jsr305-3.0.2.jar
> Failed to transform j2objc-annotations-2.8.jar (com.google.j2objc:j2objc-annotations:2.8) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}.
> Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.j2objc/j2objc-annotations/2.8/c85270e307e7b822f1086b93689124b89768e273/j2objc-annotations-2.8.jar.
> Not a module and no mapping defined: j2objc-annotations-2.8.jar
I was expecting a solution where I could add this plugin and this short snippet, and the build would just work like it did with Maven.
extraJavaModuleInfo {
automaticModule("com.google.code.findbugs:jsr305", "jsr305")
}
What I actually got is a solution where I also have to do extra work to manage all my transitive dependencies—which will become a real pain for larger projects. Each dep added could potentially introduce new build errors.
Now perhaps there are some valid reasons for this design decision w.r.t. managing transitive dependencies. But if I have to do all this extra work just to get Java modules to work with Gradle, why not just use Maven—where things just work out-of-the-box.
For many real-world projects, we can realistically expect that you're going to have to take some dependencies that don't use named modules. (E.g., for Undertow and XNIO, you will have to consume an implied automatic module as well.)
This is unrelated to this plugin.
This plugin is about patching existing Jars to add a module-info.class so that all Jars are real modules and you can fully use the Java Module System as originally intended.
The behavior in Maven you describe is to put everything on the --module-path, no matter if the Jars are compatible or not. This may fail for certain setups (invalid names, split packages). In your example it works, but there are many constellations where it does not. And "it works" still means that you do not fully use the module system as there are Automatic Modules involved. The Automatic Modules mechanism in Java only exist as a kind of intermediate compatibility. A lot of the advantages of the Module System are lost once you involve automatic modules. So if you fully want to use the Module System, you should only use dependencies that are already real modules (Guava is not https://github.com/google/guava/issues/2970).
But if you really want the behavior of Maven you describe in Gradle, this is a Gradle core issue on the topic: https://github.com/gradle/gradle/issues/12630#issuecomment-1125105629
Feel free to share your argumentation there.
With current Gradle versions, you could re-configure tasks in your build to get the Maven behavior:
tasks.withType<JavaCompile>().configureEach {
doFirst {
options.compilerArgs.add("--module-path")
options.compilerArgs.add(classpath.asPath)
classpath = files()
}
}
tasks.withType<JavaExec>().configureEach {
doFirst {
jvmArguments.add("--module-path")
jvmArguments.add(classpath.asPath)
classpath = files()
}
}
It's ultimately your call since this is your project, but I would ask that you reconsider.
I'd ask this: what is more likely to attract users to this plugin?
One day, they woke up and had this burning desire to convert all JARs to real Java modules.
They ran into a (well-)known issue where Gradle cannot not find a module when it's a filename-based automodule.
My money is on (2), especially when even javax.inject:javax.inject:1 is a filename-based automodule.
It's also worth noting that the benefits of Java modules are not all-or-nothing. E.g., you can still benefit from strong encapsulation for your module's code, even if some of your dependencies have filename-based automodules.
This could very well be a case where the perfect (real modules for all JARs) is the enemy of the good (strong encapsulation for your code), especially if people decide to either not use Java modules or switch to Maven. (In my case, I just switched to Maven.)
(Also, I had some issues getting your workaround to work.)
@mikewacker, like @jjohannes mentioned, this is the way how Gradle works and treat non-modular JARs. This plugin in particular allows you to add the missing metadata either by defining the proper module-info.class or by adding the Automatic-Module-Name entry to the JAR Manifest. While the former is recommended to benefit from the module system the most, the latter will allow you to ("quickly") solve your problem and make Gradle place the JAR on the module path. The example is below.
extraJavaModuleInfo {
automaticModule("javax.inject:javax.inject", "javax.inject")
}
The main problem that I noted there is that you have to do this for all your transitive dependencies, not just the modules that have a requires in module-info.java. (See the error message from the original post.)
Maybe I misunderstood the request a litte bit. My point is that – from a technical perspective – this plugin cannot be changed to provide exactly the same behavior that Maven has to Gradle:
Gradle modifies JARs (when using this plugin)
Maven does not modify JARs
That being said, I could imagine that we add an option to add the Automatic-Module-Name entry automatically to all Jars that are not Modules. This option would use the name of the Jar file as Module Name (same as what Java does implicitly). It can fail at build time for invalid names and inform the user to explicitly define a name for the problematic Jar. I personally do not see much value in such an option. But maybe it is interesting as one step in migrating an existing project to Modules. Or just for experimentation.
I can imagine something like this:
extraJavaModuleInfo {
allPlainJarsAreAutomaticModules = true
}
(Happy about bettter suggestions for how to call this option.)
If that is what you are looking for @mikewacker, I can reopen this issue (and adjust title and description accordingly).
It's moreso that things just work in Maven. From that perspective, it doesn't really matter if that's accomplished by adding plain JARs to the module path, or by converting all plain JARs to (explicit) automatic modules.
What Maven does is that it includes plain JARs on the module path, but it generates a warning:
[WARNING] *********************************************************************************************************************************************************************************************************************
[WARNING] * Required filename-based automodules detected: [jsr305-3.0.2.jar, undertow-core-2.3.9.Final.jar, xnio-api-3.8.8.Final.jar, javax.inject-1.jar]. Please don't publish this project to a public artifact repository! *
[WARNING] *********************************************************************************************************************************************************************************************************************
(Note: This warning only lists plain JARs that the module directly requires in module-info.java. It doesn't list all the plain JARs that you transitively depend on.)
As a simple example, let's use one of the most common Java dependencies: Guava. Here is the error that you get if have a very simple build file with only the com.google.com.guava:guava:32.1.3-jre dep and an empty extraJavaModuleInfo{} section. (Let's assume that your module never requires a plain JAR in module-info.java.)
Execution failed for task ':lib:compileJava'.
> Could not resolve all files for configuration ':lib:compileClasspath'.
> Failed to transform failureaccess-1.0.1.jar (com.google.guava:failureaccess:1.0.1) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}.
> Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.guava/failureaccess/1.0.1/1dcf1de382a0bf95a3d8b0849546c88bac1292c9/failureaccess-1.0.1.jar.
> Not a module and no mapping defined: failureaccess-1.0.1.jar
> Failed to transform listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar (com.google.guava:listenablefuture:9999.0-empty-to-avoid-conflict-with-guava) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}.
> Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.guava/listenablefuture/9999.0-empty-to-avoid-conflict-with-guava/b421526c5f297295adef1c886e5246c39d4ac629/listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar.
> Not a module and no mapping defined: listenablefuture-9999.0-empty-to-avoid-conflict-with-guava.jar
> Failed to transform jsr305-3.0.2.jar (com.google.code.findbugs:jsr305:3.0.2) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}.
> Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.code.findbugs/jsr305/3.0.2/25ea2e8b0c338a877313bd4672d3fe056ea78f0d/jsr305-3.0.2.jar.
> Not a module and no mapping defined: jsr305-3.0.2.jar
> Failed to transform j2objc-annotations-2.8.jar (com.google.j2objc:j2objc-annotations:2.8) to match attributes {artifactType=jar, javaModule=true, org.gradle.category=library, org.gradle.libraryelements=jar, org.gradle.status=release, org.gradle.usage=java-api}.
> Execution failed for ExtraJavaModuleInfoTransform: /home/mike/.gradle/caches/modules-2/files-2.1/com.google.j2objc/j2objc-annotations/2.8/c85270e307e7b822f1086b93689124b89768e273/j2objc-annotations-2.8.jar.
> Not a module and no mapping defined: j2objc-annotations-2.8.jar
At this point, I'd just take the option to turn all plain JARs into automatic modules, rather than untangle this error.
Obviously, one of the big benefits of Java modules is strong encapsulation, i.e., limiting the packages that your module exports. If I want the benefits of strong encapsulation, but I also have an unavoidable dependency on a plain JAR (e.g., Undertow, javax.inject), this is where that feature becomes useful.
I just want to add those plain JARs to my module path so that the requires statement compiles; I don't want to manage all my transitive dependencies which are plain JARs.
This is in 1.6. Turn it on with this:
extraJavaModuleInfo {
deriveAutomaticModuleNamesFromFileNames = true
}
@mikewacker let me know if you run into any issues using this.
Thanks!
Seems to work with Guava and Immutables, but I ran into an issue with Dagger. I'll open a separate issue for that.
| gharchive/issue | 2023-10-10T22:52:10 | 2025-04-01T04:34:25.600349 | {
"authors": [
"iherasymenko",
"jjohannes",
"mikewacker"
],
"repo": "gradlex-org/extra-java-module-info",
"url": "https://github.com/gradlex-org/extra-java-module-info/issues/74",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2652416210 | SceneVariableSet: Notify scene objects that use time macros when time changes
Thought this was a clever solution (Making SceneVariableSet notify scene objects that use __from and __to macro when time range changes).
But it makes it so that SceneQueryRunner executes double queries when a query is using these macros (as SceneQueryRunner is already subscribing to time range).
Need a way to dedupe those updates.
Possible solutions
dedupe the run queries somehow (when they are called with the same time range)
Opt-in to being notified of time range "__from" and "__to" variables (or opt-out) in VariableDependencyConfig
Tested locally and it seems to work, and the code looks good. Maybe making it opt-in is safest for the moment?
@kaydelaney after thinking more about it I did not like this solution much, becomes a bit complex to have it be opt-in from SceneVariableSet (would need to have some kind of special handling for this variable in the propagation chain).
Also feels a bit messy that when time range change we propagate variable change updates for for __from and __to (causing double updates when a scene object depends on both).
Opted for a simpler feature in VariableDependencyConfig that can handle the opt-in & and the time range subscription.
| gharchive/pull-request | 2024-11-12T14:46:25 | 2025-04-01T04:34:25.816980 | {
"authors": [
"kaydelaney",
"torkelo"
],
"repo": "grafana/scenes",
"url": "https://github.com/grafana/scenes/pull/966",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2168005336 | Incorrect default value placeholders on host edit form
Defect description
Create ~/.ssh/config file with the following contents:
Host localhost
HostName 127.0.0.1
User override-user
Open goto and create a host with similar parameter, but leave Login field untouched.
Notice that the placeholder of Login field contains your actual login instead of override-user as it is defined in ~/.ssh/config.
Also blocks IMPROVEMENT-47 as goto will not display proper ssh key name in the status line.
| gharchive/issue | 2024-03-04T23:49:18 | 2025-04-01T04:34:25.827060 | {
"authors": [
"grafviktor"
],
"repo": "grafviktor/goto",
"url": "https://github.com/grafviktor/goto/issues/60",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1738070375 | docs: fix broken links
Fixes #425
I fixed all links that matched core.telegram.org/bots[^/] and were broken. I also checked links matching core.telegram.org/(?!bots), and none of them were broken.
Fan
tas
tic
| gharchive/pull-request | 2023-06-02T11:59:09 | 2025-04-01T04:34:25.856887 | {
"authors": [
"KnorpelSenf",
"roj1512"
],
"repo": "grammyjs/grammY",
"url": "https://github.com/grammyjs/grammY/pull/427",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
723700799 | Foundry 0.7.4 Multilevel toksen duplication with Furnace
When I tried to make a macro run with Multilevel tokens in 0.7.4 I realised that there are two of everything related to Multilevel tokens, as you can see from the pictures, and every option (cloning, teleport, triggers, etc.) is on by default.
When I set a macro for example in one of the posssible locations, it overwrites it with the combination of the two locations.
For example I write Macro1 in the first Macro region and Macro2 in the second, I update reopen Multilevel tab and both regions are overwritten as Macro1,Macro2.
It seems to be caused by The Furnace module, but these two worked together with 0.6.6.
Not sure if it needs to be fixed in your side or Furnaces side, but I will post this to both.
### ### ### __
See https://github.com/kakaroto/fvtt-module-furnace/issues/62
Closing as I believe this is now fixed in Furnace 2.3.1
| gharchive/issue | 2020-10-17T08:01:37 | 2025-04-01T04:34:25.919969 | {
"authors": [
"Webwra",
"grandseiken"
],
"repo": "grandseiken/foundryvtt-multilevel-tokens",
"url": "https://github.com/grandseiken/foundryvtt-multilevel-tokens/issues/27",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1696660513 | InitialState is null in App.razor OnInitializedAsync Method
Hi,
Thanks for the great example! I'm very close to getting this working but running across an issue where the InitialState is null in the App.razor when I start the app. I am using the BlazorServer App. I do initialize the InitialApplicationState class in the _Host.cshtml and the access token is populated when the app starts. However, when it gets to the OnInitializedAsync method in the App.razor, InitialState is null because it is not instantiated (object reference not set to an instance of an object error). Am I missing an instantiation somewhere?
Thanks!
Oops, forgot to add the param-InitialState="initialState" on the component. Works great now!
Thanks @COBowler300. I am glad you got it working.
| gharchive/issue | 2023-05-04T20:30:50 | 2025-04-01T04:34:25.931038 | {
"authors": [
"COBowler300",
"grantcolley"
],
"repo": "grantcolley/blazor-auth0",
"url": "https://github.com/grantcolley/blazor-auth0/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
526916507 | Crawler Error
Hi everyone, I'd love to use this tool to help me with my search for any type of housing, however, I'm getting a crawler error.
File "/Users/danijel/anaconda3/bin/wg-gesucht-crawler-cli", line 10, in <module> sys.exit(cli()) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 697, in main rv = self.invoke(ctx) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/cli.py", line 109, in cli wg_gesucht.search() File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/crawler.py", line 465, in search ad_list = self.fetch_ads(filters_to_check) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/crawler.py", line 289, in fetch_ads url_list.extend(self.process_filter_results(search_results)) File "/Users/danijel/anaconda3/lib/python3.7/site-packages/wg_gesucht/crawler.py", line 240, in process_filter_results post_date_link = result.find("td", {"class": "ang_spalte_datum"}).find("a")
So apparently they changed their source code, as I cant find the class ang_spalte_datum there any more. I just don't understand the logic well enough to find the new class they replaced it with.
Any help is appreciated, I'm really betting on this bot to find a place to live
Hey @sechsneun Answered this in the other issue, https://github.com/grantwilliams/wg-gesucht-crawler-cli/issues/10#issuecomment-568722522
Let me know if that helps?
| gharchive/issue | 2019-11-22T00:25:46 | 2025-04-01T04:34:25.933793 | {
"authors": [
"grantwilliams",
"sechsneun"
],
"repo": "grantwilliams/wg-gesucht-crawler-cli",
"url": "https://github.com/grantwilliams/wg-gesucht-crawler-cli/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1157504271 | No graph4nlp_demo folder created when git clone
I followed the instructions in the repo:
created environment (python 3.8)
git clone -b v0.5.5 https://github.com/graph4ai/graph4nlp.git
While this created a graph4nlp folder, it does not have a graph4nlp_demo folder. I had to separately clone to the graph4nlp_demo repo
graph4nlp and graph4nlp_demo are two separate repositories. You may need to git clone them separately to get both of them. Thanks.
| gharchive/issue | 2022-03-02T17:42:38 | 2025-04-01T04:34:25.935455 | {
"authors": [
"SaizhuoWang",
"vinven7"
],
"repo": "graph4ai/graph4nlp_demo",
"url": "https://github.com/graph4ai/graph4nlp_demo/issues/9",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
248602050 | CarbonLink cache-query return 0 datapoints in some cases
graphite-web 1.0.2 (0.9.15 not affected)
This issue occurs only in two simultaneous circumstances:
STORAGE_DIR in local_settings.py is defined via symbolic link
There is an element in metric name of fewer than 4 characters in length
Example
$ grep 'STORAGE_DIR =' /opt/graphite/webapp/graphite/local_settings.py
STORAGE_DIR = '/var/lib/graphite/whisper'
$ ls -l /var/lib/graphite/whisper
lrwxrwxrwx 1 root root 20 Jun 26 11:59 /var/lib/graphite/whisper -> /mnt/ssd/graphite/db
$ curl '127.0.0.1/render/?from=-2min&format=json&target=Env.HTTP.NumConnections' | jq .
[
{
"target": "Env.HTTP.NumConnections",
"datapoints": [
[
48,
1502180500
],
[
50,
1502180510
],
[
50,
1502180520
],
[
null,
1502180530
],
[
null,
1502180540
],
[
null,
1502180550
],
[
null,
1502180560
],
[
null,
1502180570
],
[
null,
1502180580
],
[
null,
1502180590
],
[
null,
1502180600
],
[
null,
1502180610
]
]
}
]
$ tail /var/log/graphite/cache.log | grep cache-query
2017-08-08,12:15:08.930 :: CarbonLink cache-query request for HTTP.NumConnections returned 0 datapoints
The difference:
Env.HTTP.NumConnections
HTTP.NumConnections
I found the issue in the code.
An absolute path to whisper file is truncated to the length of the full metric name. This is not correct because absolute_path ends up with a file extension (.wsp), but relative_fs_path is not.
>>> absolute_path='/var/lib/graphite/whisper/Env/HTTP/NumConnections.wsp'
>>> real_fs_path='/mnt/ssd/graphite/db/Env/HTTP/NumConnections.wsp'
>>> metric_path='Env.HTTP.NumConnections'
>>>
>>> relative_fs_path = metric_path.replace('.', os.sep)
>>> relative_fs_path
'Env/HTTP/NumConnections'
>>> absolute_path[:-len(relative_fs_path)]
'/var/lib/graphite/whisper/Env/'
>>> base_fs_path = os.path.dirname(absolute_path[:-len(relative_fs_path)])
>>> base_fs_path
'/var/lib/graphite/whisper/Env'
>>> real_base_fs_path = os.path.realpath(base_fs_path)
>>> real_base_fs_path
'/mnt/ssd/graphite/db/Env'
>>> relative_real_fs_path = real_fs_path[len(real_base_fs_path):].lstrip('/')
>>> relative_real_fs_path
'HTTP/NumConnections.wsp'
>>> fs_to_metric(relative_real_fs_path)
'HTTP.NumConnections'
In addition, it is not at all clear why the metric name is newly calculated, because it is originally contained in the variable metric_path. Or am I missing something?
All working as expect with this patch for me:
--- __init__.py 2017-08-08 09:00:16.000000000 +0300
+++ __init__-new.py 2017-08-08 09:44:21.000000000 +0300
@@ -6,15 +6,6 @@
EXPAND_BRACES_RE = re.compile(r'.*(\{.*?[^\\]?\})')
def get_real_metric_path(absolute_path, metric_path):
- # Support symbolic links (real_metric_path ensures proper cache queries)
- real_fs_path = os.path.realpath(absolute_path)
- if absolute_path != real_fs_path:
- relative_fs_path = metric_path.replace('.', os.sep)
- base_fs_path = os.path.dirname(absolute_path[:-len(relative_fs_path)])
- real_base_fs_path = os.path.realpath(base_fs_path)
- relative_real_fs_path = real_fs_path[len(real_base_fs_path):].lstrip('/')
- return fs_to_metric(relative_real_fs_path)
-
return metric_path
Ahem... I doubt that fix is that simple. I'm not aware of what get_real_metric_path() doing though. @iksaif @DanCech ?
I'm not clear on the exact purpose, but it seems that the problem would be solved by using:
relative_fs_path = metric_path.replace('.', os.sep) + '.wsp'
@DanCech yes, this should work too.
But is not clear what get_real_metric_path() purpose? In my case with a symbolic link, all work fine without this function. Based on the code, this function extracts the metric path by truncating the real path and the resulting value is exactly the same as the metric_path.
The propose is to support similar links. Between 0.9 and 1.x, the functionality was expanded to support symlinks beyond the filename in #1738. https://github.com/graphite-project/graphite-web/pull/1738
Ah, cool. Thanks, @cbowman0 ! Maybe you could make a fix then?
I'm hitting this issue as well. Have made a PR to fix it.
| gharchive/issue | 2017-08-08T05:08:17 | 2025-04-01T04:34:25.963143 | {
"authors": [
"DanCech",
"cbowman0",
"deniszh",
"leochen4891",
"zasca"
],
"repo": "graphite-project/graphite-web",
"url": "https://github.com/graphite-project/graphite-web/issues/2012",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1377688345 | Fixing typo in documentation (apache)
This PR will fix a typo error on apache configuration file
💚 All backports created successfully
Status
Branch
Result
✅
1.1.x
Questions ?
Please refer to the Backport tool documentation
| gharchive/pull-request | 2022-09-19T09:48:04 | 2025-04-01T04:34:25.966191 | {
"authors": [
"anthony-quiros",
"deniszh"
],
"repo": "graphite-project/graphite-web",
"url": "https://github.com/graphite-project/graphite-web/pull/2777",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
126435031 | Can not open file for read - although the file exists.
So all I did was:
$ gcc incbin/incbin.c -o incbin
$ ./incbin -o scripts.cpp src/scripts.rc
and got an error:
Ingwie@Ingwies-Macbook-Pro.local ~/W/IceTea $ file src/scripts.rc
src/scripts.rc: ASCII c program text
Ingwie@Ingwies-Macbook-Pro.local ~/W/IceTea $ ./out/incbin -o scripts.cpp ./src/scripts.rc
failed to open `./src/scripts.rc' for reading
What exactly could be the issue for this?
In fact, this is the whole script: https://github.com/IngwiePhoenix/IceTea/blob/master/build.sh (will be commenting out the line that calls incbin soon, its actually not even required on UNIX :). But its analogous to the windows variant.)
Fixed in 76a0059
| gharchive/issue | 2016-01-13T14:52:26 | 2025-04-01T04:34:25.968150 | {
"authors": [
"IngwiePhoenix",
"graphitemaster"
],
"repo": "graphitemaster/incbin",
"url": "https://github.com/graphitemaster/incbin/issues/19",
"license": "unlicense",
"license_type": "permissive",
"license_source": "bigquery"
} |
2559669321 | Drop eventuals in favor of Tokio watch + timers
@Theodus suggested that Eventuals are really hard to debug and we should probably ship a standard solution like the Gateway currently does by using tokio watch.
Hi, can I have this?
This is a larger refactor that will update multiple parts of the code.
I suggest you starting with only one: escrow_accounts or allocations and we can split the PR in multiple.
Actually, to start you can update the dispute manager which is the smallest Eventual in the code.
https://github.com/graphprotocol/indexer-rs/blob/main/common/src/attestations/dispute_manager.rs
sure, I'll send PR as soon as possible
| gharchive/issue | 2024-10-01T16:06:29 | 2025-04-01T04:34:25.970815 | {
"authors": [
"gusinacio",
"taslimmuhammed"
],
"repo": "graphprotocol/indexer-rs",
"url": "https://github.com/graphprotocol/indexer-rs/issues/333",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
787163452 | Graphql-Server + Graphene + Flask issues
I'm trying to setup flask with graphene and graphql-server. As this is just a test which I am running, I'm using the latest beta releases:
graphene: 3.0.0b7
graphql-server: 3.0.0b4
Note: I had to use the beta versions of both packages, otherwise I run into dependency issues with graphql-core
Now I found some example which use graphql_server.flask.GraphQLView for the flask views. However, I couldnt use the graphene schema directly, I had to use schema.graphql_schema instead.
Is this expected?
Are there some side effects of this usage?
The following code is a minimal working example:
from flask import Flask
from graphene import ObjectType, String, Schema
from graphql_server.flask import GraphQLView
class Query(ObjectType):
hello = String(name=String(default_value='stranger'))
goodbye = String()
def resolve_hello(root, info, name):
return 'Hello {}'.format(name)
def resolve_goodbye(root, info):
return 'See ya!'
schema = Schema(query=Query)
app = Flask(__name__)
app.add_url_rule('/graphql', view_func=GraphQLView.as_view(
'graphql',
schema=schema.graphql_schema, # TODO: Check what the consequences are of using the graphql_schema
graphiql=True,
graphiql_version='1.3.2'
))
app.add_url_rule('/graphql/batch', view_func=GraphQLView.as_view(
'graphql_batch',
schema=schema.graphql_schema, # TODO: Check what the consequences are of using the graphql_schema
batch=True
))
if __name__ == '__main__':
app.run(host='0.0.0.0')
Nevermind: Just found a note in the docs which exactly mentiones this...
Nevermind: Just found a note in the docs which exactly mentiones this...
Hi @jrast
ich steh auf dem Schlauch! I have the same issue but can work it out how the docs can help me. Can you show me what you did to solve this issue?
| gharchive/issue | 2021-01-15T20:09:00 | 2025-04-01T04:34:26.007787 | {
"authors": [
"JonathanBecks",
"jrast"
],
"repo": "graphql-python/graphql-server",
"url": "https://github.com/graphql-python/graphql-server/issues/79",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1105449619 | Please Add Mutation Example
I saw the examples for juniper but only for query and I don't see any mutation example
@alresarena2021 it's not much difference between declaring query and mutations. The only difference is that a mutation should be declared in a Mutation root object.
See an example here: https://graphql-rust.github.io/juniper/master/quickstart.html#schema-example
| gharchive/issue | 2022-01-17T06:58:33 | 2025-04-01T04:34:26.009545 | {
"authors": [
"alresarena2021",
"tyranron"
],
"repo": "graphql-rust/juniper",
"url": "https://github.com/graphql-rust/juniper/issues/1018",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2744815627 | [graphiql] Autocomplete doesn't work
Is there an existing issue for this?
[X] I have searched the existing issues
Current Behavior
A few days ago, GraphiQL was working correctly, but today, without making any significant changes, I've noticed the autocomplete has stopped working. GraphiQL doesn't build the queries as it did a few days ago. I'm using graphiql with Spring boot 3.4.0 and spring-boot-starter-graphql.
Expected Behavior
Autocomlete working!
Steps To Reproduce
With a spring boot project.
Run the application (without graphql errors)
Open graphiql and try to make a query
Environment
GraphiQL Version: I don't know
OS: Ubuntu 24.04.1
Browser: Google Chrome, Opera
Bundler:I don't know
react Version:I don't know
graphql Version:I don't know
Anything else?
Fixed! Thanks!
| gharchive/issue | 2024-12-17T12:31:31 | 2025-04-01T04:34:26.014113 | {
"authors": [
"xdelvalle"
],
"repo": "graphql/graphiql",
"url": "https://github.com/graphql/graphiql/issues/3839",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1325718206 | reduce CO2 usage by running less unneccesary GH actions workflows
because it's good for the earth 🌍 and the patience of contributors 😆
i wish you could use paths-ignore further down, to skip individual tasks
note:
i decided to upgrade actions/setup-node usage as well for similar reasons, it appears to be faster but brings some bugs, so i will finish this up tonight
a lower priority fix - how to configure workflows to only re-run when only that workflow file changed? probably by using these same configs
this makes me want to decouple cypress tests into seperate workflows, as they and vscode-graphql have seperate tests that should only need to be run together if graphql-language-service changes which is rare but usually a good thing haha
another possibility, I wonder if some of these extremely edge-case node and graphql-js version matrix tests could just be run on merge to main? that way, a release breaks if it should, but it would be rare. maybe a few times a year at most we have an issue with either throughout the entire monorepo
so you know what's fun? mocking various fetch polyfill libraries across node versions, where TextDecoder may or may not be present and more excitement
| gharchive/pull-request | 2022-08-02T11:34:54 | 2025-04-01T04:34:26.017067 | {
"authors": [
"acao"
],
"repo": "graphql/graphiql",
"url": "https://github.com/graphql/graphiql/pull/2626",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1713666322 | 🛑 GrauNeko.com is down
In a7d98c7, GrauNeko.com (https://graueneko.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: GrauNeko.com is back up in 86cf026.
| gharchive/issue | 2023-05-17T11:14:51 | 2025-04-01T04:34:26.034510 | {
"authors": [
"graueneko"
],
"repo": "graueneko/status-page",
"url": "https://github.com/graueneko/status-page/issues/155",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1177755579 | 6666 - Migrate portal & console & cypress to node 16.10
Issue
gravitee-io/issues#6666
Description
Migrate portal console cypress to node 16.10
Additional context
📚 View the storybook of this branch here
🚀 CI was able to deploy the build of this PR, so you can now try it directly here
Notes: The deployed app is linked to the management API of APIM master. (Same login and password as APIM master)
Could you rebase the branch to fix the build ? thx
Yes
| gharchive/pull-request | 2022-03-23T08:05:55 | 2025-04-01T04:34:26.118583 | {
"authors": [
"ThibaudAV",
"gcusnieux"
],
"repo": "gravitee-io/gravitee-api-management",
"url": "https://github.com/gravitee-io/gravitee-api-management/pull/1486",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2595192263 | feat: ps - rework executionPhase & policy to use FlowPhase
Issue
https://gravitee.atlassian.net/browse/APIM-7242
Description
We need to keep PS compatible with all APIM 4.3 to 4.5
The goal it's to allow APIM 4.6 to use new way to define policy compatibility into phase.
So we
add new FlowPhase instead of ExecutionPhase.
add new ApiProtocolType like ApiType but "more precise"
add policy flowPhaseCompatibility to define FlowPhase compatibility for each ApiProtocolType
the Policy studio's internal code has been changed accordingly.
But the external interface remains unchanged.
Additional context
🧪 Gravitee.io Automatic Prerelease @gravitee/ui-schematics
npm install @gravitee/ui-schematics@13.4.0-apim-7242-flowphase-f02a23a
yarn add @gravitee/ui-schematics@13.4.0-apim-7242-flowphase-f02a23a
🧪 Gravitee.io Automatic Prerelease @gravitee/ui-policy-studio-angular
npm install @gravitee/ui-policy-studio-angular@13.4.0-apim-7242-flowphase-f02a23a
yarn add @gravitee/ui-policy-studio-angular@13.4.0-apim-7242-flowphase-f02a23a
🧪 Gravitee.io Automatic Prerelease @gravitee/ui-particles-angular
npm install @gravitee/ui-particles-angular@13.4.0-apim-7242-flowphase-f02a23a
yarn add @gravitee/ui-particles-angular@13.4.0-apim-7242-flowphase-f02a23a
📚 View the storybook of this branch here
:tada: This PR is included in version 13.5.0 :tada:
The release is available on:
npm package (@latest dist-tag)
npm package (@latest dist-tag)
npm package (@latest dist-tag)
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/pull-request | 2024-10-17T16:18:40 | 2025-04-01T04:34:26.128423 | {
"authors": [
"ThibaudAV",
"graviteeio"
],
"repo": "gravitee-io/gravitee-ui-particles",
"url": "https://github.com/gravitee-io/gravitee-ui-particles/pull/439",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2409838050 | feat(dgeni): generate color palettes and render in design-land
PR Checklist
Please check if your PR fulfills the following requirements:
[x] The commit message follows our guidelines: https://github.com/graycoreio/daffodil/blob/develop/CONTRIBUTING.md#commit
[x] Tests for the changes have been added (for bug fixes / features)
[x] Docs have been added / updated (for bug fixes / features)
PR Type
What kind of change does this PR introduce?
[ ] Bugfix
[x] Feature
[ ] Code style update (formatting, local variables)
[ ] Refactoring (no functional changes, no api changes)
[ ] Build related changes
[ ] CI related changes
[ ] Documentation content changes
[ ] Other... Please describe:
What is the current behavior?
Fixes: #2905
What is the new behavior?
moves lots of docs stuff to @daffodil/documentation to allow design land to access it too
Does this PR introduce a breaking change?
[ ] Yes
[ ] No
Other information
@xelaint the styling in the design land color component definitely needs some work. I think what I have sufficiently demonstrates how to use the data available though so hopefully you ca take it from here
@griest024 is this PR dependent on #2909 ?
Yes
@griest024 can we remove $error and $daff-white from the autogenerated list? I realized they need to be reworked and should not be part of this color palettes documentation.
@griest024 can we remove $error and $daff-white from the autogenerated list? I realized they need to be reworked and should not be part of this color palettes documentation.
will do, should any single colors ever be part of the generation or only the palettes that contain multiple shades?
I just went ahead and only gened shaded palettes since the implementation is much cleaner
@griest024 does this need to be reworked based on the changes Damien made in the last few weeks?
@xelaint yes but afaik those changes aren't all merged yet so I will wait until they are to rebase this.
| gharchive/pull-request | 2024-07-15T23:56:46 | 2025-04-01T04:34:26.216871 | {
"authors": [
"griest024",
"xelaint"
],
"repo": "graycoreio/daffodil",
"url": "https://github.com/graycoreio/daffodil/pull/2912",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
712637057 | Feature: Add custom endpoints
Magento 2 modules can have their own api endpoints. This PR makes it possible to use those.
For example:
`$magento = new Magento();
$response = $magento->api('/suttonsilver-pricelists')->get('pricelist/12');`
Thanks!
| gharchive/pull-request | 2020-10-01T08:51:32 | 2025-04-01T04:34:26.218781 | {
"authors": [
"VincentBean",
"ahinkle"
],
"repo": "grayloon/magento-laravel-api",
"url": "https://github.com/grayloon/magento-laravel-api/pull/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1095932568 | eliminate flicker when zooming
When zooming in/out there is sometimes a brief moment when tiles have no tile data, so the map is repainted with the background colour. The visual effect is a brief moment when the whole map is blank, which is a disorienting effect.
Fixed with 06bdd9998cd0dbc9ba888fcd50be84284bb7ed1d and eecec083abce467a28ea79cba9188d551c6c98d9
| gharchive/issue | 2022-01-07T02:55:02 | 2025-04-01T04:34:26.306416 | {
"authors": [
"greensopinion"
],
"repo": "greensopinion/flutter-vector-map-tiles",
"url": "https://github.com/greensopinion/flutter-vector-map-tiles/issues/22",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1731789 | undefined method `route_instance_path' on adding comments
When trying to add admin notes, using active_admin 0.3.2 and rails 3.0.10, the following Exception is thrown on the redirect after posting the comment:
NoMethodError in Admin::CommentsController#show: undefined method 'route_instance_path' for nil:NilClass
Stack trace show the responsible code to be:
activeadmin (0.3.2) lib/active_admin/comments.rb:47
This seems to be caused by the changes made in commit 7c2b355bd01ae0bb76b1af1c240d9ea13b617f43, reverting the commit makes it work again.
Please tell if you need the full framework trace.
Hey, I'm experiencing the same error but in a different flavour.
When posting a comment in the comment panel inside the resource "show" i get
undefined method 'route_instance_path' for nil:NilClass
in
activeadmin (0.3.2) lib/active_admin/comments.rb:47:in _callback_before_79'`
I'm not using active_reload.
I'm using activeadmin 0.3.2 on Rails 3.1.1
I'm having the same problem described by @cosenmarco.
My environment is this: https://gist.github.com/1437310
I'll re-open the bug as it seems to affect others, but try adding some useful debug info as to why route_instance_path is not defined.
same problem.
NoMethodError in Admin::CommentsController#show
undefined method `route_instance_path' for nil:NilClass
when adding a comment. Thanks
| gharchive/issue | 2011-09-24T22:18:48 | 2025-04-01T04:34:26.311218 | {
"authors": [
"cosenmarco",
"fbuenemann",
"mauriziodemagnis",
"nafkot"
],
"repo": "gregbell/active_admin",
"url": "https://github.com/gregbell/active_admin/issues/528",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1466081822 | Should I use this library ?
Hi All
I'm considering using this library but is see that the since April no code was added .
Any thoughts?
Depends on your requirement. If you have a lot of complex operations then better not use this. I dont see much support here._
'd recommend https://github.com/table-library/react-table-library if you need a good tool and can accept to go with unpopular lib, which utilizes modern concepts such as composition over configuration, otherwise https://github.com/TanStack/table is a way to go - bit outdated but well maintained and reliable.
Do not recommend https://github.com/olifolkerd/tabulator despite author @olifolkerd claims that it is react-ready, it lags and hangs when used in react, bugs are ignored.
Also going to leave Material React Table as a possible alternative here. The official MUI X DataGrid and AG Grid suggested above are great suggestions too, but they might require paid licenses for some features.
Also going to leave Material React Table as a possible alternative here. The official MUI X DataGrid and AG Grid suggested above are great suggestions too, but they might require paid licenses for some features.
I was having some dependency issues with this library. After testing many of the suggested ones I found that sometimes they are not so flexible in terms of custom styles.
I used only a few part of this library, so I decided to implement a simple, compatible, drop-in pure MUI+React replacement
https://gist.github.com/jsmolina/00dcdcda7897a7c6b23d8657e2d25716
It's just a gist, I do not plan to create a whole library from it.
| gharchive/issue | 2022-11-28T09:24:21 | 2025-04-01T04:34:26.316376 | {
"authors": [
"Alex-Github-Account",
"KevinVandy",
"Shreeabcd",
"jsmolina",
"yaniv-upstream"
],
"repo": "gregnb/mui-datatables",
"url": "https://github.com/gregnb/mui-datatables/issues/1967",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
322127359 | Page is greater than the total available
Hi, I'm having some problems using this library.
When I'm on the second page, or ahead and I make a search that should return less pages than the one I'm on, I get the following error:
Provided options.page of 1is greater than the total available page length of0``
And here is my code:
`export default class test extends React.Component {
constructor(props) {
super(props);
this.changePage = this.changePage.bind(this);
this.state = {
columns: ["foo", "bar", "ho", "hey", "ho", "let's", "Go", ""],
data: this.tratarDados(ultimosCoisosStore.getUltimosCoisos()),
options: {
responsive: 'scroll',
selectableRows: false,
filter: false,
print: false,
download: false,
rowsPerPageOptions: [5,10,15],
page: 0
}
}
this.updateData = this.updateData.bind(this);
this.tratarDados = this.tratarDados.bind(this);
this.returnIcones = this.returnIcones.bind(this);
this.refreshList = this.refreshList.bind(this);
coisoDispatcher.handleAction({
type: 'UPDATE_ULTIMOS_PEDIDOS'
});
}
componentDidMount() {
ultimosCoisosStore.addChangeListener('ULTIMOS_COISOS_UPDATED', this.updateData);
coiso.addChangeListener('REFRESH_LIST_ULTIMOS_COISOS', this.refreshList);
}
componentWillUnmount() {
coiso.removeChangeListener('REFRESH_LIST_ULTIMOS_COISOS', this.refreshList);
ultimosCoisosStore.addChangeListener('ULTIMOS_COISOS_UPDATED', this.updateData);
}
tratarDados(data) {
let temp = [];
for (var x in data) {
let data_temp = Object.values(data[x]);
data_temp[1] = parseFloat(data_temp[1]);
//data_temp[2] = moment(data_temp[2]).format("DD/MM/YYYY HH:mm:ss");
let icones = this.returnIcones(data_temp[6], x, data_temp[7]);
data_temp.pop();
data_temp.push(icones);
temp.push(Object.values(data_temp));
}
return temp;
}
returnIcones(status, id, show_delete) {
if (status === 3 || status === 4) {
return (
<BotoesOk key={id} nf_id={id} status={status} ok={true} show_delete={show_delete}/>
)
}
else {
return (
<BotoesOk key={id} nf_id={id} status={status} ok={false} show_delete={show_delete}/>
)
}
}
refreshList() {
//this.setState({options.page: 0});
pdvDispatcher.handleAction({
type: 'UPDATE_ULTIMOS_COISOS'
});
}
updateData() {
let dados = this.tratarDados(ultimosCoisosStore.getUltimosCoisos());
this.setState({data: dados});
}
render() {
return (
<Rodal onClose={this.props.hideModal} className="ultimos-coisos-modal" visible={true} width={1200} >
<MUIDataTable key={Math.random()} title={"Últimos Coisos"} data={this.state.data} columns={this.state.columns} options={this.state.options} />
</Rodal>
)
}
}`
Could someone help me unsderstand what is happening?
It looks like there's a bug. I will look into this today
This issue should be resolved. Upgrade to version 2.0.0-beta-3
Hi Greg,
First of all, thx for this library, saves me tons of work. I'm a newbee to javascript so there could be a very simple explanation to this,, but i am experiencing this very same issue on version 2.0.0-beta-37. My code:
''''js
import * as React from 'react';
import MUIDataTable from "mui-datatables";
import { createMuiTheme, MuiThemeProvider } from '@material-ui/core/styles';
export class ClientList extends React.Component {
state = {
page: 0,
count: 100,
rowsPerPage: 10,
data: []
};
getMuiTheme = () => createMuiTheme({
overrides: {
MuiMenuItem: {
root: {
fontSize: '1.25rem'
}
}
, MuiChip: {
label: {
fontSize: '1.25rem'
}
}
, MuiInputLabel: {
root: {
fontSize: '1.5rem'
}
}
, MuiTooltip: {
tooltip: {
fontSize: '1.25rem'
}
}
, MuiInputBase: {
input: {
fontSize: '1.25rem'
}
}
, MuiSelect: {
selectMenu: {
fontSize: '1.25rem'
}
}
, MuiTypography: {
h6: {
fontSize: '2rem'
}
, caption: {
fontSize: '1.25rem'
}
}
, MUIDataTableHeadCell: {
root: {
fontSize: '1.25rem'
, fontWeight: 'bold'
}
}
, MUIDataTableBodyCell: {
root: {
fontSize: '1.25rem'
//,backgroundColor: "#FF0000"
}
}
}
})
componentDidMount() {
this.getClientListWrapper(0,10);
}
onChangeRowsPerPage = rowsPerPage => {
this.setState({ rowsPerPage });
};
// get data
//getData = () => {
// this.getClientList().then(data => {
// this.setState({ data });
// });
//}
//get list of clients
getClientList = () => {
return new Promise((resolve, reject) => {
fetch('api/Client/Index')
.then(function (response) { return resolve(response.json()); })
//.then(function (myJson) { console.log(JSON.stringify(myJson)); })
;
});
}
//route to client overview page
showClientOverview = (id) => {
this.props.history.push("/client/edit/" + id);
};
getClientListWrapper = (page, rowsPerPage) => {
this.getClientList().then(data => {
this.setState({
page: page,
rowsPerPage: rowsPerPage,
data
});
});
};
render() {
const columns = [{ name: "Id", options: { display: false } }, "Clientnummer", "Naam", "Airline"];
const { data, page } = this.state;
const data1 = data.map(item => { return [item.clientId, item.clientNumber, item.fullName, item.companyName] });
const options = {
filter: true
,filterType: 'dropdown'
,responsive: 'stacked'
,serverSide: false
,count: data1.length
,page: page
,rowsPerPage: this.state.rowsPerPage
,rowsPerPageOptions: [10, 15, 50, 100]
,onChangeRowsPerpage: this.onChangeRowsPerPage
,onRowClick: (rowData, rowMeta) => {
this.showClientOverview(rowData[0])
}
,onTableChange: (action, tableState) => {
//console.log(action, tableState);
// a developer could react to change on an action basis or
// examine the state as a whole and do whatever they want
switch (action) {
case 'changePage':
this.getClientListWrapper(tableState.page, tableState.rowsPerPage);
break;
}
}
//TODO: load labels from label table
,textLabels: {
body: {
noMatch: "Clienten laden...",
toolTip: "Sorteren",
}
, pagination: {
next: "Volgende Pagina",
previous: "Vorige Pagina",
rowsPerPage: "Aantal objecten per pagina:",
displayRows: "van",
}
, toolbar: {
search: "Zoek",
downloadCsv: "Download CSV",
print: "Print",
viewColumns: "Filter Kolommen",
filterTable: "Filter Tabel",
}
, filter: {
all: "Alles",
title: "Filters",
reset: "Verwijder Filters",
}
, viewColumns: {
title: "Toon Kolommen",
titleAria: "Toon/Verberg Kolommen",
}
, selectedRows: {
text: "geselecteerde rij(en)",
delete: "Verwijder",
deleteAria: "Verwijder Geselecteerde Rijen",
}
}
};
return (
<MuiThemeProvider theme={this.getMuiTheme()}>
<MUIDataTable title={"Team Vervoort - Clienten"} data={data1} columns={columns} options={options} />
</MuiThemeProvider>
);
}
}
''''
I have 21 clients in my database. When navigating to last page (2) there is one last client on that page. If i change rowsPerPage to 15 then, the page goes blank. Refreshing it defaults to first page, rowsPerPage = 10. However: changing rowsPerPage to 50 or 100 works perfectly. Any ideas?
| gharchive/issue | 2018-05-11T00:12:32 | 2025-04-01T04:34:26.324427 | {
"authors": [
"gregnb",
"rafinha90",
"skalma"
],
"repo": "gregnb/mui-datatables",
"url": "https://github.com/gregnb/mui-datatables/issues/48",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2116835737 | Consider implementing From where T: Into<#inner_type>?
Currently derive(From) produces https://github.com/greyblake/nutype/blob/c8e3f72346c4294fce2982c5f1ea71224552bbe3/nutype_macros/src/common/gen/traits.rs#L125-L134
Is there anything holding back the production of something like the following?
quote! {
impl<T: Into<#inner_type>> ::core::convert::From<T> for #type_name {
#[inline]
fn from(raw_value: T) -> Self {
Self::new(raw_value.into())
}
}
}
@schneiderfelipe Hi, thanks for bringing this.
I guess I had a thought about it, the reason I've decided not to implement it is clarity / explicitness.
But you may make me reconsider this.
What's you real life use case, where you think it would make things a bit easier?
Well, it would make newtypes behave more like the inner type. For instance, something like the following would work, but currently doesn't:
#[newtype(derive(From))]
struct MyVec(Vec<isize8>);
// This works
let my_vec: Vec<isize8> = [0, 1, 2].into();
// This does not work, requires a separate From impl
let my_vec: MyVec = [0, 1, 2].into();
@schneiderfelipe Sold! :)
@danma3x Would you be interested in addressing this one?
Sure!
@schneiderfelipe Sold! :)
@danma3x Would you be interested in addressing this one?
There's a problem with my proposed change: it produces conflicting implementations of the trait From, e.g.,
error[E0119]: conflicting implementations of trait `From<_>` for type `derives::test_trait_from_string::__nutype_Name__::Name`
--> test_suite/tests/string.rs:357:9
|
357 | #[nutype(derive(From))]
| ^^^^^^^^^^^^^^^^^^^^^^^
| |
| first implementation here
| conflicting implementation for `derives::test_trait_from_string::__nutype_Name__::Name`
|
= note: this error originates in the attribute macro `nutype` (in Nightly builds, run with -Z macro-backtrace for more info)
In particular, for the case of integers, things seem much worse:
error[E0277]: the trait bound `u32: From<i32>` is not satisfied
--> test_suite/tests/integer.rs:532:22
|
532 | let amount = Amount::from(350);
| ^^^^^^ the trait `From<i32>` is not implemented for `u32`
|
= help: the following other types implement trait `From<T>`:
<u32 as From<bool>>
<u32 as From<char>>
<u32 as From<u8>>
<u32 as From<u16>>
<u32 as From<NonZeroU32>>
<u32 as From<Ipv4Addr>>
= note: required for `i32` to implement `Into<u32>`
note: required for `traits::test_trait_from::__nutype_Amount__::Amount` to implement `From<i32>`
--> test_suite/tests/integer.rs:529:9
|
529 | #[nutype(derive(From))]
| ^^^^^^^^^^^^^^^^^^^^^^^ unsatisfied trait bound introduced here
530 | pub struct Amount(u32);
| ^^^^^^
= note: this error originates in the attribute macro `nutype` (in Nightly builds, run with -Z macro-backtrace for more info)
So the compiler seems to figure out that 350 should be a u32 just fine if there is only a single From implementation, but "falls back" to i32 and relies on a separate From<i32> implementation otherwise, which should not exist obviously.
I'm afraid my proposal can't be done 😞 @greyblake what do you think?
The changes I made are in the master branch of my fork https://github.com/schneiderfelipe/nutype/commit/cca70474cba07f55de89f00e88fff2d89126a17e
Considering nutypes with validation may be fallible, and From shouldn't support fallible conversions, shouldn't the trait implementation be TryFrom, not From?
Considering nutypes with validation may be fallible, and From shouldn't support fallible conversions, shouldn't the trait implementation be TryFrom, not From?
Conversions are fallible when there is validation. When there is no validation set, conversions are infallible.
If there is validation, nutype won't allow you to derive From.
Though you can still derive TryFrom if there is no validation.
Ok I gave a try for a generic From implementation, but it's not gonna work the way we want.
I you want to reproduce the example, I pushed my changes in generic-from-impl-demo, see also https://github.com/greyblake/nutype/commit/d73ddeb214f47d754fb2a0b80156682437301564 commit.
Consider the following example:
#[nutype(derive(Into, From))]
pub struct Amount(i32);
This won't compile:
error[E0119]: conflicting implementations of trait `From<Amount>` for type `Amount`
The expands into the following:
impl ::core::convert::From<Amount> for i32 {
#[inline]
fn from(value: Amount) -> Self {
value.into_inner()
}
}
impl<T> ::core::convert::From<T> for Amount
where
i32: ::core::convert::From<T>,
{
#[inline]
fn from(raw_value: T) -> Self {
let inner_value = i32::from(raw_value);
Self::new(inner_value)
}
}
Note, that Into<i32> for Amount is actually defined as From<Amount> for i32, to keep the symmetry. Into<i32> for Amount will got implemented automatically through blanket implementation. This is common practice / idiom of the Rust language.
But in this case it becomes also a source of a problem.
It's not very obvious to see, but Rust finds 2 conflicting implementation for for From<Amount> for Amount.
The std has reasonable implementation of impl<T> From<T> for T, meaning converting from type T to T should just return self.
Other path Rust could take is: Amount -> i32 ->Amount, because our generic implementation of From enable this.
This could be addressed by implementing Into trait as solely Into<i32> from Amount,
meaning we lose From<Amount> for i32 implementation.
Between these 2 available options, there is no clear winner, a trade off needs to be make.
In this regard, my personal preference would be the current status quo.
With that I am closing the issue.
Thanks for you attention, if you followed me!
@greyblake nicely explained! I'm afraid this won't ever be possible, except maybe if specialisation gets stabilised (rust-lang/rust#31844). Thank you for taking the time though, really appreciated!
| gharchive/issue | 2024-02-04T02:16:00 | 2025-04-01T04:34:26.347146 | {
"authors": [
"asasine",
"greyblake",
"schneiderfelipe"
],
"repo": "greyblake/nutype",
"url": "https://github.com/greyblake/nutype/issues/124",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
429605409 | obv implementation
Implemented obv. It is another volume based indicator.
Please let me know what do you think of this indicator.
Hey, thanks for the new contribution :)
Please let me know what do you think of this indicator.
I haven't heard about this indicator before. The fact that the indicator accumulates volume differences and may return very huge numbers look a little bit weird for me.
But I would like to add it to the library :)
I haven't heard about this indicator before. The fact that the indicator accumulates volume differences and may return very huge numbers look a little bit weird for me.
Quoting from investopedia
Despite being plotted on a price chart and measured numerically, the actual individual quantitative value of OBV is not relevant. The indicator itself is cumulative, while the time interval remains fixed by a dedicated starting point, meaning the real number value of OBV arbitrarily depends on the start date. Instead, traders and analysts look to the nature of OBV movements over time; the slope of the OBV line carries all of the weight of analysis.
Its a very good indicator for volume divergences.
@shreyasdeotare Thanks!
OBV is available in the newly released version 0.1.4: https://crates.io/crates/ta/0.1.4
| gharchive/pull-request | 2019-04-05T06:42:30 | 2025-04-01T04:34:26.351553 | {
"authors": [
"greyblake",
"shreyasdeotare"
],
"repo": "greyblake/ta-rs",
"url": "https://github.com/greyblake/ta-rs/pull/13",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1519823949 | Cannot Restore WAX Account
Description
I am unable to restore my WAX paper backup certificate using "Import from key certificate". Everything seems to work fine and then I get a spinning wheel of death at the "Updating accounts..." display. I'm not doing anything especially strange, just following the prompts, and then it gets stuck at that spinner. See details the Steps to Reproduce section.
Platform
Desktop (MacOS)
Steps To Reproduce
Happens using desktop version 1.3.8 on macOS Ventura 13.1 (22C65).
Created an EOS and WAX account on mobile. Backed up to paper certificate so I could migrate to desktop and remove them from mobile. Could not import either from the launch screen after fresh install. Kept getting stuck after all verified with a spinning wheel of death at "Updating accounts..."
After multiple attempts, finally got the EOS account to import by adding the EOS blockchain and going to "Import from key certificate" instead. When it worked correctly, it gave me a timer countdown and eventually worked just fine.
Then I tried to do the same for WAX, added the blockchain, tried to use "Import from key certificate" and tried Option #1 and Option #2 (I really want Option #2) but I always get the spinning wheel of death at "Updating account..." and it doesn't go away after waiting a very long time. I am on a slow (tethered hotspot) at the moment, but nothing too crazy, I'm able to browse websites, etc.
Opened up the developer console and I only see some UI related warning that seem irrelevant, and in the network tab I see repeated calls to get_table_rows but nothing that looks like an error. When I close the "Updating account..." window then everything goes away, but no account added, and no progress. So I can't restore this account (must have tried over 20 times by now and same results each time).
Relevant log output
I don't see any errors in the developer tools.
Anything else?
I was able to hit this endpoint just fine:
https://wax.greymass.com/
{
"server_version":"12902721",
"chain_id":"1064487b3cd1a897ce03ae5b6a865651747e2e152090f99c1d19d44e01aea5a4",
"head_block_num":222827194,
"last_irreversible_block_num":222826865,
"last_irreversible_block_id":"0d48117112baad3c44cf03bcf887e4c227255a1c76a6ec86dc5594812ff2d4f1",
"head_block_id":"0d4812ba91df59048a9da66bfbee81cfe9e2b9210934c140530275500606fe6f",
"head_block_time":"2023-01-05T01:02:08.000",
"head_block_producer":"sentnlagents",
"virtual_block_cpu_limit":762817,
"virtual_block_net_limit":1048576000,
"block_cpu_limit":200000,
"block_net_limit":1048576,
"server_version_string":"v2.0.12wax02",
"fork_db_head_block_num":222827194,
"fork_db_head_block_id":"0d4812ba91df59048a9da66bfbee81cfe9e2b9210934c140530275500606fe6f",
"server_full_version_string":"v2.0.12wax02-12902721dbd9c8a8d2734ede4f8978c1ce8c17b8"
}
Thanks, I'll respond privately.
Thanks for the email 👍
I think I've isolated what's going on. The account does indeed lack the required RAM to update the permissions, which is why it's getting stuck. We have added an error handler in our dev build so that it'll at least display this error message.
Since this is a pretty common scenario - we have also updated one of Anchor's services to also provide this RAM in instances where this happens on accounts trying to recover.
I believe if you try again (no need to update Anchor), you should be able to recover the account now.
1.3.9 is also now released which actually displays an error instead of just getting stuck at this point in the process:
https://github.com/greymass/anchor/releases/tag/v1.3.9
| gharchive/issue | 2023-01-05T01:05:49 | 2025-04-01T04:34:26.361618 | {
"authors": [
"aaroncox",
"raycardillo"
],
"repo": "greymass/anchor",
"url": "https://github.com/greymass/anchor/issues/1324",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
378818353 | How to get templateLayoutFactory
I need to create markers with custom layout. In issue #27 contextTypes were used to receive ymaps, however after the definiton of the same contextTypes structure I found out that ymaps in the context was empty. The second way I tried was to access ymaps with HOC. And again, incoming ymaps from that HOC hasn't templateLayoutFactory. Maybe I should somehow tell withYMaps to load that module, if it is a module and If not, then how can I access that thing?
Sorry, it was realy some kind of a dumb question after all.
Prefer to leave this issue here in case someone else will occur the same problem.
The main issue is that templateLayoutFactory is realy a module itself and it is not loaded by default. It you want it, then just add it it modules, that should be loaded or load full package, or if you need it on-demand, then just wrap your component in withYMaps, set second param true (component will wait until API is loaded) and as third param is an array, to which you should just add:
templateLayoutFactrory
to the modules list, that should be loaded.
There is also an example of how to do a HOC for templateLayotFactory in migration guide:
https://react-yandex-maps.now.sh/migration-guide#onapiavaliable-is-removed-from-the-library
It is kinda confusing, I know. We will move a bunch of those to FAQ at some point
@gribnoysup, thanks
| gharchive/issue | 2018-11-08T16:38:41 | 2025-04-01T04:34:26.367759 | {
"authors": [
"ShiiRochi",
"gribnoysup"
],
"repo": "gribnoysup/react-yandex-maps",
"url": "https://github.com/gribnoysup/react-yandex-maps/issues/119",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2476320341 | 🛑 B-K Lighting, Inc. is down
In d31d8be, B-K Lighting, Inc. (https://bklighting.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: B-K Lighting, Inc. is back up in 332532f after 26 minutes.
| gharchive/issue | 2024-08-20T18:52:57 | 2025-04-01T04:34:26.391389 | {
"authors": [
"gripfastistech"
],
"repo": "gripfastistech/status",
"url": "https://github.com/gripfastistech/status/issues/90",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1428621971 | Comfort Activity Switch Name
Looks like HK behavior is different with new iOS. Need to fix something to make switch name show activity name, and not "[ZONE] Comfort Activity".
I am still experiencing this issue on iOS 17.0.2, macOS 14.0 (23A344)
Carrier Infinity:
System Control Model # SYSTXCCWIC01-B
System Control Firmware Version CESR131626-04.47
I have had some luck with identifying labels but the hold function seems to change their behavior and it cannot be reverted within HomeKit. Using the Carrier Infinity app does work as expected.
The switches are significantly improved on 1.6.13 or later. Give it a try, and feel free to reopen (or make a new issue) if you're still experiencing issues.
| gharchive/issue | 2022-10-30T05:54:04 | 2025-04-01T04:34:26.397416 | {
"authors": [
"MatthewNrmOK",
"grivkees"
],
"repo": "grivkees/homebridge-carrier-infinity",
"url": "https://github.com/grivkees/homebridge-carrier-infinity/issues/385",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
980673270 | RangeInput track color prop docs
Signed-off-by: GurkiranSingh gurkiransinghk@gmail.com
Doc are added for the RangeInput track color prop.
https://github.com/grommet/grommet/pull/5275
Hah, @jcfilben -- Sorry, I had this open all day and made a bunch of comments overlapping with yours! Looks like we are on a similar train of thought. Apologies for the confusion.
| gharchive/pull-request | 2021-08-26T21:51:46 | 2025-04-01T04:34:26.399752 | {
"authors": [
"g4rry420",
"halocline"
],
"repo": "grommet/grommet-site",
"url": "https://github.com/grommet/grommet-site/pull/304",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1313712081 | Networking: Bad Gateway should lead to explicit pipeline failure
Current behavior:
If an endpoint isn't available, nginx responds with its default 502 response.
Expected behavior:
Rudra makes it explicity, that tests are failing, because the host cannot be resolved.
Similar issues with timeouts, etc. should also be taken into account.
Modified behavior to print out warning when all responses are 502s.
| gharchive/issue | 2022-07-21T19:34:13 | 2025-04-01T04:34:26.417905 | {
"authors": [
"grossamos"
],
"repo": "grossamos/rudra",
"url": "https://github.com/grossamos/rudra/issues/2",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
36931173 | by default ... "Delete workspace before build starts" should be enabled
I can't think of a good reason why "Delete workspace before build starts" should not be enabled.
It will prevent previous build artifacts/report on the master of prior builds
from being incorporated by the current artifacts/report running on the slave.
Not doing so can lead to false broken builds.
e.g its possible a prior build had generates a broken test report that exists on the master build.
It's rare but development may require previous report to generate differences against the current report. They can disable this field ... rather than forcing all most developments who create new builds to enable this parameter.
Not having this as a default causes a lot of confusion as any error logs based on workspace files will lead people and developers down the wrong path wondering what is wrong with their build setups.
I agree this should be a default value.
DotCi now by default cleans workspace before build per
https://github.com/groupon/DotCi/blame/master/src/main/java/com/groupon/jenkins/buildtype/dockercompose/BuildConfiguration.java#L165
shellCommands.add("chmod -R u+w . ; find . ! -path \"./deploykey_rsa.pub\" ! -path \"./deploykey_rsa\" -delete");
Interesting side note:
Docker user beware of volume mounting as root. It's recommend that docker-compose.override.yml be setup to create the userid of the slave agent. Because testing can result in creation of files and if image is running as root then those files will now be owned by root. However, the next build will lack permission to to perform the necessary find delete in the subsequent build, as the slave agent userid.
| gharchive/issue | 2014-07-01T21:00:57 | 2025-04-01T04:34:26.453377 | {
"authors": [
"danmconrad",
"tmack8001",
"vvitayau"
],
"repo": "groupon/DotCi",
"url": "https://github.com/groupon/DotCi/issues/46",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
212287833 | objc-tests: InteropTestsRemoteWithCronet test4MBResponsesAreAccepted failure
https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3190/consoleText
/jenkins/workspace/gRPC_pull_requests_macos/workspace_objc_macos_dbg_native/src/objective-c/tests/InteropTests.m:164: error: -[InteropTestsRemoteWithCronet test4MBResponsesAreAccepted] : ((error) == nil) failed: "Error Domain=io.grpc Code=3 "Received message larger than max (4194316 vs. 4194304)" UserInfo={io.grpc.TrailersKey={
}}" - Finished with unexpected error: Error Domain=io.grpc Code=3 "Received message larger than max (4194316 vs. 4194304)" UserInfo={io.grpc.TrailersKey={
Test Case '-[InteropTestsRemoteWithCronet test4MBResponsesAreAccepted]' failed (2.314 seconds).
This is keeping the macos master build red. Is the fix complex?
seen again in https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3314/testReport/junit/(root)/objc_macos_opt_native/objc_tests/
pr: https://github.com/grpc/grpc/pull/10076 against v1.2.x
seen again in https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3333/testReport/junit/(root)/objc_macos_dbg_native/objc_tests/
pr https://github.com/grpc/grpc/pull/9986 against v1.2.x
Seen again in https://grpc-testing.appspot.com/job/gRPC_pull_requests_macos/3346/testReport/junit/(root)/objc_macos_dbg_native/objc_tests/
pr #10087 against v1.2.x
| gharchive/issue | 2017-03-06T23:54:20 | 2025-04-01T04:34:26.595009 | {
"authors": [
"apolcyn",
"dgquintas",
"y-zeng"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/issues/10003",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
263996006 | Artifact build on Windows failed
https://sponge.corp.google.com/target?id=da133ab3-ab38-456d-b057-008de7e91964&target=github/grpc&searchFor=&show=ALL&sortBy=STATUS
same as https://github.com/grpc/grpc/issues/12872 which has been fixed already.
| gharchive/issue | 2017-10-09T19:05:37 | 2025-04-01T04:34:26.596474 | {
"authors": [
"adelez",
"jtattermusch"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/issues/12911",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
280645112 | New Failure: csharp.Grpc.Core.Tests.ContextPropagationTest
Test: csharp.Grpc.Core.Tests.ContextPropagationTest
Poll Strategy: None
URL: https://kokoro2.corp.google.com/job/grpc/job/ubuntu/job/master/job/grpc_basictests_multilang/691
https://sponge.corp.google.com/invocation?id=ca86c237-d2ae-4cb0-89b3-1dd41dd267a0&searchFor=
this should also be fixed by https://github.com/grpc/grpc/pull/13675.
| gharchive/issue | 2017-12-08T23:16:26 | 2025-04-01T04:34:26.598688 | {
"authors": [
"dgquintas",
"jtattermusch"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/issues/13690",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
377469077 | grpc-java qps worker: qps_driver often reports "Worker 1 could not be properly quit because Received RST_STREAM with error code 8"
The entire scenario passes, but qps_driver reports problem terminating java qps_woker.
It happens only for grpc-java.
The status reported by java worker seems to be GRPC_STATUS_RESOURCE_EXHAUSTED = 8
2018-11-05 13:28:44,247 START: qps_worker_java_0
2018-11-05 13:28:44,251 START: qps_worker_java_1
2018-11-05 13:28:44,254 START: qps_json_driver.java_protobuf_async_streaming_ping_pong_secure
2018-11-05 13:28:44,258 WAITING: 1 queued, 1 jobs running, 0 complete, 0 failed (load 1.00) next: qps_json_driver.quit @ 1.00 cpu
2018-11-05 13:29:32,423 ++ dirname tools/run_tests/performance/run_qps_driver.sh
+ cd tools/run_tests/performance/../../..
+ cmake/build/qps_json_driver '--scenarios_json={"scenarios": [{"name": "java_protobuf_async_streaming_ping_pong_secure", "warmup_seconds": 15, "benchmark_seconds": 30, "num_servers": 1, "server_config": {"async_server_threads": 1, "channel_args": [{"str_value": "latency", "name": "grpc.optimization_target"}], "security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "threads_per_cq": 0, "server_type": "ASYNC_SERVER"}, "client_config": {"security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "channel_args": [{"str_value": "latency", "name": "grpc.optimization_target"}], "async_client_threads": 1, "outstanding_rpcs_per_channel": 1, "rpc_type": "STREAMING", "payload_config": {"simple_params": {"resp_size": 0, "req_size": 0}}, "client_channels": 1, "threads_per_cq": 0, "load_params": {"closed_loop": {}}, "client_type": "ASYNC_CLIENT", "histogram_params": {"max_possible": 60000000000.0, "resolution": 0.01}}, "num_clients": 1}]}' --scenario_result_file=scenario_result.json
RUNNING SCENARIO: java_protobuf_async_streaming_ping_pong_secure
I1105 13:28:44.266914471 14417 driver.cc:276] Starting server on localhost:10400 (worker #0)
D1105 13:28:44.267247177 14417 ev_posix.cc:169] Using polling engine: epollex
D1105 13:28:44.267283569 14417 dns_resolver.cc:338] Using native dns resolver
D1105 13:28:44.267631541 14417 dns_resolver.cc:279] Start resolving.
I1105 13:28:44.268252223 14417 subchannel.cc:874] Connect failed: {"created":"@1541424524.268217537","description":"Failed to connect to remote host: OS Error","errno":111,"file":"/tmpfs/src/github/grpc/src/core/lib/iomgr/tcp_client_posix.cc","file_line":205,"os_error":"Connection refused","syscall":"connect","target_address":"ipv4:127.0.0.1:10400"}
I1105 13:28:44.268271388 14417 subchannel.cc:756] Subchannel 0x7f3b680036f0: Retry in 999 milliseconds
D1105 13:28:44.268297072 14417 dns_resolver.cc:259] In cooldown from last resolution (from 1 ms ago). Will resolve again in 999 ms
D1105 13:28:45.266394537 14429 dns_resolver.cc:279] Start resolving.
I1105 13:28:45.267204871 14429 subchannel.cc:715] Failed to connect to channel, retrying
I1105 13:28:45.267701347 14417 subchannel.cc:837] New connected subchannel at 0x55bda5385810 for subchannel 0x7f3b680036f0
I1105 13:28:46.097366545 14417 driver.cc:335] Starting client on localhost:10410 (worker #1)
D1105 13:28:46.097465420 14417 driver.cc:357] Client 0 gets 1 channels
D1105 13:28:46.097501750 14417 dns_resolver.cc:279] Start resolving.
I1105 13:28:46.098013094 14417 subchannel.cc:837] New connected subchannel at 0x7f3b68001bf0 for subchannel 0x7f3b68005d40
I1105 13:28:46.837003397 14417 driver.cc:378] Initiating
I1105 13:28:46.855068021 14417 driver.cc:399] Warming up
I1105 13:29:01.855263396 14417 driver.cc:405] Starting
I1105 13:29:01.871796201 14417 driver.cc:432] Running
I1105 13:29:31.855285081 14417 driver.cc:446] Finishing clients
I1105 13:29:31.883236617 14417 driver.cc:460] Received final status from client 0
I1105 13:29:31.883348844 14417 driver.cc:492] Finishing servers
I1105 13:29:31.888333061 14417 driver.cc:506] Received final status from server 0
I1105 13:29:31.889280898 14417 report.cc:82] QPS: 3526.7
I1105 13:29:31.890097805 14417 report.cc:122] QPS: 3526.7 (inf/server core)
I1105 13:29:31.890113801 14417 report.cc:127] Latencies (50/90/95/99/99.9%-ile): 246.0/400.5/444.7/520.8/652.1 us
I1105 13:29:31.890120604 14417 report.cc:137] Server system time: 0.00%
I1105 13:29:31.890126595 14417 report.cc:139] Server user time: 71.74%
I1105 13:29:31.890131975 14417 report.cc:141] Client system time: 0.00%
I1105 13:29:31.890143351 14417 report.cc:143] Client user time: 70.10%
I1105 13:29:31.890148851 14417 report.cc:148] Server CPU usage: 0.00%
I1105 13:29:31.890154007 14417 report.cc:153] Client Polls per Request: 0.00
I1105 13:29:31.890158888 14417 report.cc:155] Server Polls per Request: 0.00
I1105 13:29:31.890164413 14417 report.cc:160] Server Queries/CPU-sec: 4911.88
I1105 13:29:31.890169730 14417 report.cc:162] Client Queries/CPU-sec: 5031.00
+ '[' performance_test.performance_experiment_singlevm '!=' '' ']'
+ tools/run_tests/performance/bq_upload_result.py --bq_result_table=performance_test.performance_experiment_singlevm
Warning: Table performance_experiment_singlevm already exists
Successfully uploaded scenario_result.json to BigQuery.
2018-11-05 13:29:32,423 PASSED: qps_json_driver.java_protobuf_async_streaming_ping_pong_secure [time=48.2sec, retries=0:0]
2018-11-05 13:29:32,423 START: qps_json_driver.quit
2018-11-05 13:29:32,426 WAITING: ETA 48.2 sec; 0 queued, 1 jobs running, 1 complete, 0 failed (load 1.00)
2018-11-05 13:29:32,452 E1105 13:29:32.441533137 14537 driver.cc:550] Worker 0 could not be properly quit because Received RST_STREAM with error code 8
E1105 13:29:32.450617305 14537 driver.cc:550] Worker 1 could not be properly quit because Received RST_STREAM with error code 8
2018-11-05 13:29:32,452 FAILED: qps_json_driver.quit [ret=1, pid=14536, time=0.0sec]
E.g. here:
https://source.cloud.google.com/results/invocations/b3638c0b-37c1-4ddf-998f-561d98823f41/targets;collapsed=/grpc%2Fcore%2Fmaster%2Flinux%2Fgrpc_e2e_performance_singlevm/log
The worker log:
2018-11-05 13:29:35,801 ++ dirname tools/run_tests/performance/run_worker_java.sh
+ cd tools/run_tests/performance/../../..
+ cd ../grpc-java
+ benchmarks/build/install/grpc-benchmarks/bin/benchmark_worker --driver_port=10410
Nov 05, 2018 1:28:46 PM io.grpc.benchmarks.driver.LoadClient <init>
INFO: Client Config
server_targets: "localhost:37451"
client_type: ASYNC_CLIENT
security_params {
use_test_ca: true
server_host_override: "foo.test.google.fr"
}
outstanding_rpcs_per_channel: 1
client_channels: 1
async_client_threads: 1
rpc_type: STREAMING
load_params {
closed_loop {
}
}
payload_config {
simple_params {
}
}
histogram_params {
resolution: 0.01
max_possible: 6.0E10
}
channel_args {
name: "grpc.optimization_target"
str_value: "latency"
}
[GC (Allocation Failure) [PSYoungGen: 524800K->18141K(611840K)] 524800K->18165K(2010112K), 0.0229622 secs] [Times: user=0.14 sys=0.03, real=0.02 secs]
Nov 05, 2018 1:29:32 PM io.grpc.benchmarks.driver.LoadWorker$WorkerServiceImpl quitWorker
INFO: Received quitWorker request.
Nov 05, 2018 1:29:32 PM io.grpc.benchmarks.driver.LoadWorker main
INFO: DriverServer has terminated.
Heap
PSYoungGen total 611840K, used 102588K [0x00000000d5580000, 0x0000000100000000, 0x0000000100000000)
eden space 524800K, 16% used [0x00000000d5580000,0x00000000da7f7d10,0x00000000f5600000)
from space 87040K, 20% used [0x00000000f5600000,0x00000000f67b7498,0x00000000fab00000)
to space 87040K, 0% used [0x00000000fab00000,0x00000000fab00000,0x0000000100000000)
ParOldGen total 1398272K, used 24K [0x0000000080000000, 0x00000000d5580000, 0x00000000d5580000)
object space 1398272K, 0% used [0x0000000080000000,0x0000000080006000,0x00000000d5580000)
Metaspace used 20736K, capacity 20896K, committed 21248K, reserved 1069056K
class space used 2387K, capacity 2441K, committed 2560K, reserved 1048576K
2018-11-05 13:29:35,801 PASSED: qps_worker_java_1 [time=51.5sec, retries=0:0]
@carl-mastrangelo could the error GRPC_STATUS_RESOURCE_EXHAUSTED be reported by grpc-java because of [GC (Allocation Failure) [PSYoungGen: 524800K->18141K(611840K)] 524800K->18165K(2010112K), 0.0229622 secs] [Times: user=0.14 sys=0.03, real=0.02 secs]?
Would increasing java heap size help?
Btw, this is happening even for single channel ping pong scenarios like "java_generic_async_streaming_ping_pong_secure" or "java_protobuf_async_unary_ping_pong_secure", which seems odd, because with these scenarios, there's max 1 RPC in flight at any given time
I'm able to reproduce locally by running tools/run_tests/run_performance_tests.py -l java -r 'java_protobuf_unary_ping_pong_insecure' and it happens in 100% of cases it seems.
Based on some local experiments I have a suspicion that the quitWorker logic might be flawed:
https://github.com/grpc/grpc-java/blob/65bd38476f2007a65f249bb852c99069ab9f2c87/benchmarks/src/main/java/io/grpc/benchmarks/driver/LoadWorker.java#L254
The driverServer.shutdownNow(); is invoked right after the quitWorker RPC is finished and this seems to be causing the C++ qps_driver to receiver error code 8 (GRPC_STATUS_RESOURCE_EXHAUSTED) from that quitWorker invocation.
The GC (Allocation Failure) is a red herring. That just means Young was filled up, and it's doing a GC.
Looking through the Java code, we rarely use ResourceExhausted, so I don't think this would be generated locally. shutdown now terminates new incoming RPCs, but should let all existing RPCs finish gracefully. Shutdown commands are serialized along with writes on the socket.
Unfortunately I could only repro it once. I am reassigning back to you since I can't diagnose it.
| gharchive/issue | 2018-11-05T16:08:39 | 2025-04-01T04:34:26.607328 | {
"authors": [
"carl-mastrangelo",
"jtattermusch"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/issues/17101",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
456761153 | PHP: insecure environment read function 'getenv' used"
What version of gRPC and what language are you using?
PHP 7.2.19
PECL GRPC 1.21.3
What operating system (Linux, Windows,...) and version?
Alpine Linux 3.9.4
What runtime / compiler are you using (e.g. python version or version of gcc)
PHP
What did you do?
Update to 1.21.3 (not sure the what the version was before). Nothing helpful to describe here I'm afraid.
What did you expect to see?
Expected no warnings in log since default logging level is "ERROR" (and GRPC_VERBOSITY is set to "ERROR").
What did you see instead?
[17-Jun-2019 08:02:15] WARNING: [pool application] child 171 said into stderr: "D0617 08:02:15.289927347 171 env_linux.cc:71] Warning: insecure environment read function 'getenv' used"
Anything else we should know about your project / environment?
Search for this issue and found https://github.com/grpc/grpc/issues/8104 but this is outdated. The issue showed itself after updating to 1.21.3, but I'm not sure which version we came from. I found references of other people recently having this issue https://github.com/grpc/grpc/issues/18833#issuecomment-499518864. Maybe @villers has more insight into what is happening.
Also looked into some commits (I'm be no means an expert, just trying to be helpful). The following commit seems to have recently touched the logging levels: 90fbdc92f522af9f98297e08c0ed174361977d46
Does caused by this commit https://github.com/grpc/grpc/pull/18539/files ?
I am seeing this as well.
Any plans for fixing this?
I think this line blow is the cause
https://github.com/grpc/grpc/pull/18539/commits/3107cda311853ce57f3a67847ce15dd14a5c9fd5#diff-f4dd247c3672ef404b2cec6be6573a39R208
@kellegous could this be related to https://github.com/grpc/grpc/pull/18539 in your opinion? (as the author).
@Swahjak It's possible but I'm not sure how. That change uses getenv directly and doesn't use gpr_getenv which is what issues that warning. It could be possible that somewhere in grpc, the libc getenv is being replaced with gpr_getenv but I haven't looked to see if that's the case.
file /src/core/lib/gpr/env_linux.cc
line 71: gpr_log(GPR_DEBUG, "Warning: insecure environment read function '%s' used",
line 72: insecure_func_used);
When SET ENV GRPC_VERBOSITY=ERROR
Already print 'Warning: insecure environment read function 'getenv' used'
why?
I attempted to get a backtrace to the offending call to gpr_getenv this morning and, as best I can tell, this is what is emitting the warning.
Breakpoint 1, gpr_getenv (
name=name@entry=0x7ffff7781608 <g_env_str_grpc_verbosity> "GRPC_VERBOSITY")
at /tmp/pear/temp/grpc/src/core/lib/gpr/env_linux.cc:69
69 /tmp/pear/temp/grpc/src/core/lib/gpr/env_linux.cc: No such file or directory.
(gdb) bt
#0 gpr_getenv (name=name@entry=0x7ffff7781608 <g_env_str_grpc_verbosity> "GRPC_VERBOSITY")
at /tmp/pear/temp/grpc/src/core/lib/gpr/env_linux.cc:69
#1 0x00007ffff74f2598 in grpc_core::GlobalConfigEnv::GetValue (
this=this@entry=0x7ffff7781620 <g_env_grpc_verbosity>)
at /tmp/pear/temp/grpc/src/core/lib/gprpp/global_config_env.cc:75
#2 0x00007ffff74f28b4 in grpc_core::GlobalConfigEnvString::Get (
this=this@entry=0x7ffff7781620 <g_env_grpc_verbosity>)
at /tmp/pear/temp/grpc/src/core/lib/gprpp/global_config_env.cc:126
#3 0x00007ffff74ef573 in gpr_global_config_get_grpc_verbosity ()
at /tmp/pear/temp/grpc/src/core/lib/gpr/log.cc:78
#4 gpr_log_verbosity_init () at /tmp/pear/temp/grpc/src/core/lib/gpr/log.cc:78
#5 0x00007ffff74f2d69 in do_basic_init () at /tmp/pear/temp/grpc/src/core/lib/surface/init.cc:69
#6 0x00007ffff7fbf01a in ?? () from /lib/ld-musl-x86_64.so.1
#7 0x00005555563ab6a0 in core_globals ()
#8 0x00007ffff7fbef99 in pthread_mutexattr_settype () from /lib/ld-musl-x86_64.so.1
#9 0x00007ffff7784ee8 in g_init_mu ()
from /usr/local/lib/php/extensions/no-debug-non-zts-20180731/grpc.so
file: src/php/ext/grpc/php_grpc.c
PHP_RINIT_FUNCTION(grpc) {
if (!GRPC_G(initialized)) {
apply_ini_settings(TSRMLS_C);
grpc_init();
register_fork_handlers();
grpc_php_init_completion_queue(TSRMLS_C);
GRPC_G(initialized) = 1;
}
return SUCCESS;
}
RINIT function
before grpc_init()
apply_ini_settings() OR MINIT Function
where is it used getenv() ?
The warning should be removed after #19691 is merged. We tested this on master and verified the warning is gone. Please watch out for the next upcoming 1.23 release.
Fixed in 1.23.0RC1. 1.23.0 will be released in 2 weeks.
| gharchive/issue | 2019-06-17T06:19:58 | 2025-04-01T04:34:26.619755 | {
"authors": [
"Swahjak",
"axot",
"kellegous",
"lvzhihao",
"stanley-cheung",
"taka-oyama"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/issues/19366",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
467114561 | memory_usage_test fail under mac
D0711 10:54:11.677362000 140735725650816 ev_posix.cc:174] Using polling engine: poll
D0711 10:54:11.678205000 140735725650816 dns_resolver_ares.cc:485] Using ares dns resolver
D0711 10:54:11.692659000 140735725650816 test_config.cc:384] test slowdown factor: sanitizer=1, fixture=1, poller=1, total=1
D0711 10:54:11.692816000 140735725650816 test_config.cc:384] test slowdown factor: sanitizer=1, fixture=1, poller=1, total=1
D0711 10:54:11.693971000 140735725650816 ev_posix.cc:174] Using polling engine: poll
D0711 10:54:11.693988000 140735725650816 ev_posix.cc:174] Using polling engine: poll
D0711 10:54:11.694074000 140735725650816 dns_resolver_ares.cc:485] Using ares dns resolver
D0711 10:54:11.694077000 140735725650816 dns_resolver_ares.cc:485] Using ares dns resolver
I0711 10:54:11.694176000 140735725650816 server.cc:180] creating server on: [::]:27749
I0711 10:54:11.696370000 140735725650816 subchannel.cc:1076] New connected subchannel at 0x7ff5afe06160 for subchannel 0x7ff5afe04760
memory_usage_client(23863,0x7fff96ef4380) malloc: *** error for object 0x7ff5afe02150: pointer being freed was not allocated
*** set a breakpoint in malloc_error_break to debug
Fixed memory_usage_test failure under Linux by #19811. Not sure if it fixed the situation under Mac since there is another test failed #19819 .
| gharchive/issue | 2019-07-11T21:25:57 | 2025-04-01T04:34:26.625232 | {
"authors": [
"bigfacebear",
"yang-g"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/issues/19614",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
964126327 | C++ callback service OnDone called twice and app crahsed
What version of gRPC and what language are you using?
1.39.0, C++
What operating system (Linux, Windows,...) and version?
Windows, OS Version: 10.0.19043 N/A Build 19043
What runtime / compiler are you using (e.g. python version or version of gcc)
Microsoft (R) C/C++ Optimizing Compiler Version 19.16.27045 for x64
What did you do?
Can't disclose my source, will try to come up w/ a standalone UT, but this is the flow (verified OnDone called twice)
Create bidi stream
"classic" impl for OnReadDone behavior
void GrpcSession::OnReadDone(bool ok)
{
if (ok)
{
try
{
switch (_curReq.message_case())
{
// BIZ logic, dispatch requests yada yada
}
StartRead(&_curReq);
}
catch (const exception& ex)
{
FinishWrapper({ INTERNAL, ex.what() });
}
}
else
{
FinishWrapper(Status::OK);
}
}
void GrpcSession::OnWriteDone(bool ok)
{
if (ok)
{
Response* resp;
{
lock_guard<mutex> _{ _respMtx };
ASSERT(!_respQueue.empty(), "Queue must be not empty");
_respQueue.pop();
if (!_respQueue.empty())
{
resp = _respQueue.front().get();
}
else
{
resp = nullptr;
}
}
if (resp)
{
StartWrite(resp);
}
else
{
_respQueueEmpty.notify_all();
}
}
else
{
FinishWrapper({ INTERNAL, "OnWriteDone failed" });
}
}
void GrpcSession::FinishWrapper(const Status& st)
{
if (_finishRequested.test_and_set()) // this is atomic_flag
{
return;
}
if (!st.ok())
{
GS_ERROR("Completing session due to err=%s", st.error_message().c_str());
}
Finish(st);
}
Run client which "bombs" the service w/ messages
Close the client app abruptly
OnReadDone will invoke Finish (since ok is false)
OnDone is called
Outstanding write will fail, but I won't issue another call to Finish since FinishWarpper is idempotent
Library is initiating another OnDone
crash (since its seems ctx_ was destroyed at:
void CallOnDone() override {
reactor_.load(std::memory_order_relaxed)->OnDone();
grpc_call* call = call_.call();
auto call_requester = std::move(call_requester_);
if (ctx_->context_allocator() != nullptr) {
ctx_->context_allocator()->Release(ctx_);
}
this->~ServerCallbackReaderWriterImpl(); // explicitly call destructor
::grpc::g_core_codegen_interface->grpc_call_unref(call);
call_requester();
}
What did you expect to see?
What did you see instead?
Anything else we should know about your project / environment?
Your code rocks :) I LOVE the new callback approach, and would like to incorporate it deeper in our projects
My bad, seems calling Finish while there is an Write in the air is a big no no
| gharchive/issue | 2021-08-09T15:41:14 | 2025-04-01T04:34:26.631946 | {
"authors": [
"mosdav"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/issues/26947",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
269015630 | Bump version to 1.7.1
(mostly to deliver https://github.com/grpc/grpc/pull/13169).
[trickle] No significant performance differences
[microbenchmarks] No significant performance differences
Superseded by https://github.com/grpc/grpc/pull/13168
| gharchive/pull-request | 2017-10-27T07:43:09 | 2025-04-01T04:34:26.634028 | {
"authors": [
"grpc-testing",
"jtattermusch"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/13170",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
372594814 | In gRPC-C++ podspec, copy the certificate to bundle resources
In GRPC.podspec, the root certificates file from etc/roots.pem is copied to the bundle resources, so that it can be loaded at runtime for SSL to function properly. However, roots.pem is not preserved in gRPC-C++.podspec, so a project depending on that pod won't be able to establish SSL connections. Having the project maintain its own copy of the certificates file and keep it in sync with gRPC repo is undesirable.
****************************************************************
libgrpc.so
VM SIZE FILE SIZE
++++++++++++++ ++++++++++++++
[ = ] 0 0 [ = ]
****************************************************************
libgrpc++.so
VM SIZE FILE SIZE
++++++++++++++ ++++++++++++++
[ = ] 0 0 [ = ]
[trickle] No significant performance differences
Objective-C binary sizes
*****************STATIC******************
New size Old size
1,997,294 Total (=) 1,997,294
No significant differences in binary sizes
***************FRAMEWORKS****************
New size Old size
11,021,771 Total (>) 11,021,759
No significant differences in binary sizes
Corrupt JSON data (indicates timeout or crash):
bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.new: 10
bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.old: 10
[microbenchmarks] No significant performance differences
****************************************************************
libgrpc.so
VM SIZE FILE SIZE
++++++++++++++ ++++++++++++++
[ = ] 0 0 [ = ]
****************************************************************
libgrpc++.so
VM SIZE FILE SIZE
++++++++++++++ ++++++++++++++
[ = ] 0 0 [ = ]
[trickle] No significant performance differences
Objective-C binary sizes
*****************STATIC******************
New size Old size
2,015,303 Total (=) 2,015,303
No significant differences in binary sizes
***************FRAMEWORKS****************
New size Old size
11,115,426 Total (=) 11,115,426
No significant differences in binary sizes
Corrupt JSON data (indicates timeout or crash):
bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.new: 10
bm_call_create.BM_IsolatedFilter_ClientChannelFilter_NoOp_.counters.old: 10
[microbenchmarks] No significant performance differences
Basic Tests MacOS [opt] - #16201
Basic Tests Multi-language Linux - #15308
| gharchive/pull-request | 2018-10-22T16:29:55 | 2025-04-01T04:34:26.638176 | {
"authors": [
"grpc-testing",
"muxi",
"var-const"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/16962",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
539373246 | Backport #21487 to v1.26.x.
Reference: https://github.com/grpc/grpc/pull/21487
This fixes broken Ruby artifact builds on MacOS.
CC @veblush @apolcyn
Does this mean a patch only for Ruby?
Since v1.26.0 hasn't been tagged yet, I don't think this requires a patch at all.
Failures:
https://github.com/grpc/grpc/issues/20385
Going ahead and merging as this is purely a Ruby change and all relevant Ruby tests have already passed.
| gharchive/pull-request | 2019-12-17T23:52:38 | 2025-04-01T04:34:26.641006 | {
"authors": [
"gnossen",
"srini100"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/21502",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
597537124 | Removing obsolete C++ tutorial content
Contributes to https://github.com/grpc/grpc.io/issues/180.
cc @ejona86 @jtattermusch @srini100
One of the disadvantages of hosting doc pages as part of a repo is that, AFAIK, it isn't possible to create a redirect. So, this PR removes the page content but directs readers to the page's new home.
If any of you feel that we should just drop the file, I'm ok with that too. Let me know what you prefer.
| gharchive/pull-request | 2020-04-09T20:34:28 | 2025-04-01T04:34:26.642557 | {
"authors": [
"chalin"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/22633",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
605663648 | dont run resource_quota_server test case under epoll1
See internal b/151212019 for context.
based on @karthikravis, @yashykt and @yang-g opinions it's the best to just stop running the test case on epoll1 (which is the only poller when the test case is flaky)
the way to exclude a poller didn't exist in generate_tests.bzl, so I added it.
also unmarking the test as flaky (as it won't be run at all).
At some point we might get rid of epoll1 entirely.
This fix is pretty high impact as it should reduce the number of distinct items we're seeing in the flakiness dashboard basically in half:
| gharchive/pull-request | 2020-04-23T16:11:26 | 2025-04-01T04:34:26.644881 | {
"authors": [
"jtattermusch"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/22751",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
826979740 | Fix link to test file
@drfloob
@bradfol Sorry I missed this! If you'd like to sign the CLA and reopen the PR, I'll be happy to merge it.
| gharchive/pull-request | 2021-03-10T02:39:47 | 2025-04-01T04:34:26.645849 | {
"authors": [
"bradfol",
"drfloob"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/25666",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
868449400 | xds_end2end_test test infra: Eds Args refactoring and enhancing WaitForBackend
This change is
known issue #26128
| gharchive/pull-request | 2021-04-27T04:42:00 | 2025-04-01T04:34:26.647160 | {
"authors": [
"donnadionne"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/26093",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1226988091 | [promises] Convert lame client
@markdroth
I think remaining failures are unrelated.
| gharchive/pull-request | 2022-05-05T17:42:08 | 2025-04-01T04:34:26.648274 | {
"authors": [
"ctiller"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/29587",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1554134795 | GCP Observability Logging: Base64 Encode metadata, message and status-details
Tested on Stanley's CI
Stop this feed
On Mon, Jan 23, 2023, 6:52 PM Stanley Cheung @.***>
wrote:
@.**** approved this pull request.
—
Reply to this email directly, view it on GitHub
https://github.com/grpc/grpc/pull/32184#pullrequestreview-1266722899,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AVMQI7VRERDNJKH73HCBZ2DWT47V5ANCNFSM6AAAAAAUEQNCGI
.
You are receiving this because you are subscribed to this thread.Message
ID: @.***>
| gharchive/pull-request | 2023-01-24T01:31:11 | 2025-04-01T04:34:26.651642 | {
"authors": [
"Tette69",
"yashykt"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/32184",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1976796755 | [Python Misc] Revert change to print backtrace in server
Fix: https://github.com/grpc/grpc/issues/34853
In order to make debugging easier, we have begun printing backtraces in servers. However, this change has the unintended consequence of printing errors to stderr by default, which may not be expected by some users.
This PR reverts the change. We recommend that users set up a logging sink if they want to see errors. We will add this to our documentation later.
Hi there, any news on when this is going to be released? It's blocking Pulumi on Python 3.12 for everybody :-(
https://github.com/pulumi/pulumi/issues/14258
🎉 @XuanWang-Amos, @gnossen, do you know when this will be released?
@XuanWang-Amos, when will this be released? Thank you!
| gharchive/pull-request | 2023-11-03T19:19:35 | 2025-04-01T04:34:26.654265 | {
"authors": [
"XuanWang-Amos",
"justinvp",
"zyv"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/34877",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
132490048 | Add grpc_version.py to PYTHON-MANIFEST.in
Should fix #5152
Can one of the admins verify this patch?
Can one of the admins verify this patch?
this is ok to test
This makes most of the distribtests pass
https://grpc-testing.appspot.com/view/Artifacts/job/gRPC_distribtest/62/architecture=x64,language=python,platform=linux/console
| gharchive/pull-request | 2016-02-09T18:07:45 | 2025-04-01T04:34:26.656289 | {
"authors": [
"grpc-jenkins",
"jtattermusch",
"soltanmm"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/5155",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
142476855 | Update clang-format to 3.8
Pros:
we standardize on just one version of clang
the new version sorts includes
I mostly prefer the new decisions
Cons:
I only mostly prefer the new decisions
Churn
Let's bite the bullet on this. LGTM.
| gharchive/pull-request | 2016-03-21T21:29:42 | 2025-04-01T04:34:26.657986 | {
"authors": [
"ctiller",
"vjpai"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/5895",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
150154180 | Temporarily reduce throughput test depth to 1 RPC per channel
due to flake. Need to root-cause the flake.
LGTM.
Undo LGTM.
Looks like the test is still timing out:
TIMEOUT: qps_json_driver.cpp_single_channel_throughput_secure [pid=9036]
Yes, just realized that. It's become more urgent to root-cause this, and maybe it wasn't caused by this particular change.
BTW, note that this failed on the perf smoke test, which is a single-machine test. So, it's not about single-machine vs multi-machine.
Superseded by#6270
| gharchive/pull-request | 2016-04-21T18:27:42 | 2025-04-01T04:34:26.660241 | {
"authors": [
"jtattermusch",
"vjpai"
],
"repo": "grpc/grpc",
"url": "https://github.com/grpc/grpc/pull/6250",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
126461814 | Fatal error: travis-ci build Cannot read property 'contents' of undefined
this is similar to #330. here is travis-ci build-log and my repository. I am able to build locally , but i can't succeed travis-ci build. Can any one look at this?
I am encountering the same issue with my travis-ci build for an Angular + Firebase Application...
Gruntfile.js defines the task
imagemin: {
dist: {
files: [{
expand: true,
cwd: '<%= yeoman.app %>/images',
src: '{,*/}*.{png,jpg,jpeg,gif}',
dest: '<%= yeoman.dist %>/images'
}]
}
},
grunt build completes successfully locally...
Running "imagemin:dist" (imagemin) task
Minified 9 images (saved 77.32 kB)
Done, without errors.
Execution Time (2016-01-13 17:04:49 UTC)
imagemin:dist 1.6s 100%
Total 1.6s
grunt build errors out via travis-ci...
Warning: Running "imagemin:dist" (imagemin) task
Fatal error: Cannot read property 'contents' of undefined
Execution Time (2016-01-13 17:00:32 UTC)
loading tasks 9ms 2%
imagemin:dist 441ms 98%
Total 450ms Use --force to continue.
Aborted due to warnings.
I have tried to determine if there is a syntax issue in how I am declaring src in the task... however, that does not appear to be the case... as I can successfully complete the build locally without error using multiple combinations, including the example provided by Grunt for grunt-contrib-imagemin
Ref: https://github.com/gruntjs/grunt-contrib-imagemin
Any thoughts or suggestions?
As a follow-up... I have successfully moved passed this issue by swapping out "grunt-contrib-imagemin": "^1.0.0" for "grunt-image": "^1.2.1" ... and subsequently updating my Gruntfile.js by replacing imagemin tasks with image tasks...
Local builds and Travis builds are now succeeding with the same results.
Gruntfile.js
image: {
dist: {
files: [{
expand: true,
cwd: '<%= yeoman.app %>/images',
src: '{,*/}*.{png,jpg,jpeg,gif}',
dest: '<%= yeoman.dist %>/images'
}]
}
},
Sample build output below:
Running "image:dist" (image) task
✔ app/images/npmjs.png -> before=3.75 kB after=158 B reduced=3.59 kB(95.9%)
✔ app/images/bootstrap.png -> before=41.5 kB after=21.93 kB reduced=19.57 kB(47.2%)
✔ app/images/bowerjs.png -> before=17.71 kB after=6.65 kB reduced=11.06 kB(62.5%)
✔ app/images/nodejs.png -> before=8.45 kB after=2.93 kB reduced=5.51 kB(65.3%)
✔ app/images/starterlog.png -> before=34.53 kB after=10.58 kB reduced=23.95 kB(69.3%)
✔ app/images/gruntjs.png -> before=84.38 kB after=27.98 kB reduced=56.4 kB(66.8%)
✔ app/images/firebase.png -> before=32.03 kB after=12.19 kB reduced=19.83 kB(61.9%)
✔ app/images/fontawesome.png -> before=12.01 kB after=5.63 kB reduced=6.38 kB(53.1%)
✔ app/images/angular.png -> before=41.5 kB after=18.63 kB reduced=22.87 kB(55.1%)
Done, without errors.
Execution Time (2016-01-13 17:47:44 UTC)
image:dist 2.6s 99%
Total 2.6s
it works :+1: !!!! after removing "^" when specifying a version of vinyl-fs in package.json file , it builds successfully with this build-log. Thanks @juliusosokinas , @lorijoan , @ernestorocha , @FellowHobbyist and everybody else.
I am closing this issue now!
| gharchive/issue | 2016-01-13T16:44:56 | 2025-04-01T04:34:26.678513 | {
"authors": [
"FellowHobbyist",
"anbestephen"
],
"repo": "gruntjs/grunt-contrib-imagemin",
"url": "https://github.com/gruntjs/grunt-contrib-imagemin/issues/345",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
561854760 | Failed to upload state: AccessDenied (As root user)
Hi Terragrunt team,
I'm trying to use Terragrunt for a from scratch project I am doing and found out that I can't upload the state to the S3 bucket created with terragrunt initusing the root account. I think it could be related to #770 and #978.
I'm using latest version:
$ terragrunt -version
terragrunt version v0.21.11
$ terraform -version
Terraform v0.12.19
+ provider.aws v2.48.0
Output of terragrunt apply:
module.iam_amontalban.aws_iam_user.this[0]: Creating...
Failed to save state: failed to upload state: AccessDenied: Access Denied
status code: 403, request id: 992F1BEF89128A5B, host id: XvdXpf1q2kx7NIuD26MU+OaaIbtEjGMdZgI6IGqXZe0sxducyoj7d/bIe0PwixdiPatZNC+VbJI=
Error: Failed to persist state to backend.
Policy in the S3 bucket created with Terragrunt:
$ aws-vault exec acme_root -- aws s3api get-bucket-policy --bucket acme-bucket
{
"Policy": "{\"Version\":\"2012-10-17\",\"Statement\":[{\"Sid\":\"RootAccess\",\"Effect\":\"Allow\",\"Principal\":{\"AWS\":\"arn:aws:iam::123456789012:root\"},\"Action\":\"s3:*\",\"Resource\":\"arn:aws:s3:::acme-bucket\"}]}"
}
Current running user:
$ aws-vault exec acme_root -- aws sts get-caller-identity
{
"UserId": "123456789012",
"Account": "123456789012",
"Arn": "arn:aws:iam::123456789012:root"
}
Not even an aws s3 cp works:
$ aws-vault exec acme_root -- aws s3 cp variables.tf s3://acme-bucket --sse=aws:kms
upload failed: ./variables.tf to s3://acme-bucket/variables.tf An error occurred (AccessDenied) when calling the PutObject operation: Access Denied
Let me know if you need any other information.
Thanks!
Investigating this a little bit looks like the policy generated in #978 is wrong, it is:
{
"Version": "2012-10-17",
"Statement": [
{
"Sid": "RootAccess",
"Effect": "Allow",
"Principal": {
"AWS": "arn:aws:iam::123456789012:root"
},
"Action": "s3:*",
"Resource": "arn:aws:s3:::acme-bucket"
}
]
}
When it should be like:
{
"Version": "2012-10-17",
"Statement": [
{
"Sid": "RootAccess",
"Effect": "Allow",
"Principal": {
"AWS": "arn:aws:iam::123456789012:root"
},
"Action": "s3:*",
"Resource": [
"arn:aws:s3:::acme-bucket",
"arn:aws:s3:::acme-bucket/*"
]
}
]
}
After applying the above policy to the S3 bucket I was able to store the state.
Thanks for reporting! I think you may be right. Would you be up for a quick PR to fix this?
Hi @brikis98 I've added the fix for this issue, can you check it. Thanks.
Should be fixed in https://github.com/gruntwork-io/terragrunt/releases/tag/v0.21.13. Please check (binaries should show up in a few min).
| gharchive/issue | 2020-02-07T20:27:00 | 2025-04-01T04:34:26.685801 | {
"authors": [
"amontalban",
"brikis98",
"nicomfer"
],
"repo": "gruntwork-io/terragrunt",
"url": "https://github.com/gruntwork-io/terragrunt/issues/1038",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1500130125 | Migrate terraform to terragrunt for high number of modules
I have microservice non-prod env. where each team has its own env. which means a lot of envs, in terms of terrafrom right now I have two github repos:
The first one has root tf modules.
The Second one has reusable tf modules.
The first one contains 10 root modules each one of them calls just a single module that resides in the second repo, this module in the second repo calls 13 modules, so one of the modules in the second repo represents a proxy repo.
to recap, the first repo structure:
tf-env1.tf
tf-env2.tf
tf-env3.tf
....
tf-env10.tf
The second one has a structure:
api-gateway.tf
vpc.tf
rds.tf
ec2.tf
s3.tf
lambda.tf
lb.tf
SPECIAL_Module.tf
the call chain goes as:
in repo1 : tf-envX.tf
|
in repo2 : SPECIAL_Module.tf
|
in repo2 : api-gateway.tf,vpc.tf,rds.tf,s3.tf .... etc
As part of migrating to terragrunt in order to give the developers an ability to control their env. via managing their env. variables and free up the SRE team to just writing new terraform modules.
I'm thinking of getting rid of SPECIAL_Module.tf, since it has a lot of local controls on how terraform works, but the issue is if I remove SPECIAL_Module.tf terragrunt doesn't support more than one source on each hcl file.
Another option is to create an hcl for each module from the second repo, but this means I would manage # of modules * # of envs files., so it came to my mind to group the modules in the second repo into 3 or 4 groups.
One solution I thought about is to move SPECIAL_Module.tf to the first module and include it in each env., but this means a huge single terraform file would reside inside terragrunt repo. which I'm not sure if it's an abuse for terragrunt.
So is there a way to migrate from terraform to terragrunt with a high # of modules and a high # of envs?
hi,
ideas on how to adopt Terragrunt can be collected from examples:
https://github.com/gruntwork-io/terragrunt-infrastructure-live-example
https://terragrunt.gruntwork.io/docs/features/keep-your-terraform-code-dry/
Code from SPECIAL can be referenced in dependencies
https://terragrunt.gruntwork.io/docs/reference/config-blocks-and-attributes/#dependencies
The first link is very simple example use case, in the second one the # of modules are 3 with 3 env. which is a different case.
| gharchive/issue | 2022-12-16T12:27:47 | 2025-04-01T04:34:26.697384 | {
"authors": [
"aladdin55",
"denis256"
],
"repo": "gruntwork-io/terragrunt",
"url": "https://github.com/gruntwork-io/terragrunt/issues/2397",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
918019161 | Adapt count parser to tf 0.15.5
The plan output for no changes changed in tf 0.15.5, so we need to update the regex.
Thanks for review! Build failed, but upon investigation they are transient errors unrelated to this change set, so will go ahead and merge this in!
| gharchive/pull-request | 2021-06-10T22:49:36 | 2025-04-01T04:34:26.698675 | {
"authors": [
"yorinasub17"
],
"repo": "gruntwork-io/terratest",
"url": "https://github.com/gruntwork-io/terratest/pull/925",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1149724270 | 🛑 bretterhofer.at is down
In 7e8ce6b, bretterhofer.at (https://bretterhofer.at) was down:
HTTP code: 0
Response time: 0 ms
Resolved: bretterhofer.at is back up in 9cbb890.
| gharchive/issue | 2022-02-24T20:37:14 | 2025-04-01T04:34:26.713080 | {
"authors": [
"grzchr15"
],
"repo": "grzchr15/uptime",
"url": "https://github.com/grzchr15/uptime/issues/4314",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2061779015 | Active Identifier Unhandled Error thrown in read handler
I'm receiving the below error fairly frequently within the logs, and after awhile it the xbox plugin stops working until a restart. The below error seems to correlate to whenever I open home app, or run a random home shortcut and polling of the xbox occurs even when I'm not interacting with it.
Below is my config. Which is simple in nature.
"devices": [
{
"name": "Living Room Xbox Series X",
"host": "10.0.1.132",
"xboxLiveId": "REMOVEDID",
"webApiControl": true,
"getInputsFromDevice": true,
"filterGames": true,
"filterApps": true,
"filterSystemApps": true,
"filterDlc": true,
"inputsDisplayOrder": 1,
"sensorPower": false,
"sensorInput": false,
"sensorScreenSaver": false,
"webApiPowerOnOff": true,
"webApiRcControl": false,
"webApiVolumeControl": false,
"webApiToken": "REMOVEDTOKEN",
"webApiClientId": "REMOVEDCLIENTID",
"enableDebugMode": false,
"disableLogInfo": false,
"disableLogDeviceInfo": false,
"infoButtonCommand": "nexus",
"volumeControl": -1,
"enableRestFul": false,
"restFulDebug": false,
"enableMqtt": false,
"mqttDebug": false,
"mqttAuth": false
}
],
"_bridge": {
"username": "0E:D7:D1:31:C2:57",
"port": 42538
},
"platform": "XboxTv"
}
Log snippet:
01/01/2024, 14:42:43] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, set Power: ON
[01/01/2024, 14:42:53] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Connected.
[01/01/2024, 14:42:53] [homebridge-xbox-tv] -------- Living Room Xbox Series X --------'
[01/01/2024, 14:42:53] [homebridge-xbox-tv] Manufacturer: Microsoft
[01/01/2024, 14:42:53] [homebridge-xbox-tv] Model: Xbox Series X
[01/01/2024, 14:42:53] [homebridge-xbox-tv] Serialnr: REMOVED
[01/01/2024, 14:42:53] [homebridge-xbox-tv] Firmware: 10.0.25398
[01/01/2024, 14:42:53] [homebridge-xbox-tv] Locale: en-US
[01/01/2024, 14:42:53] [homebridge-xbox-tv] ----------------------------------
[01/01/2024, 14:42:59] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, set Power: OFF
[01/01/2024, 14:43:16] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Disconnected.
[01/01/2024, 14:44:16] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 14:44:17] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Current Media: STOP
[01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Target Media: STOP
[01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Mute: ON
[01/01/2024, 14:44:49] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Volume: 0
[01/01/2024, 14:44:49] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:44:49] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:45:22] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:45:27] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:45:43] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:52:24] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 14:52:24] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:52:30] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:52:32] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 14:52:32] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:52:34] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 14:52:35] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 14:53:31] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 14:53:31] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 16:01:52] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 16:01:52] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 16:12:38] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 16:12:39] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 16:13:04] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Power: OFF
[01/01/2024, 16:13:04] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 16:13:11] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[01/01/2024, 16:13:12] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
I am also getting this error, though I haven't managed to capture it when debug is enabled yet:
[01/01/2024, 12:11:13] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId')
Have now been able to capture the error with debug logging, hopefully the attached will help you @grzegorz914
homebridge.log.txt
this should be fixed with latest update 2.12.3
Thanks for your fantastic effort!
Thanks @grzegorz914 I have updated to 2.12.4 and will let you know if I see the error again.
Still seeing it in my logs on 2.12.4 unfortunately:
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Power: OFF
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Current Media: STOP
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Target Media: STOP
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Mute: ON
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Volume: 0
[02/01/2024, 17:21:30] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Power: OFF
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Current Media: STOP
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Target Media: STOP
[02/01/2024, 17:21:30] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Mute: ON
[02/01/2024, 17:21:30] [XboxTv] Device: 192.168.1.52 Xbox Series X, Volume: 0
[02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Power: OFF
[02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Current Media: STOP
[02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Target Media: STOP
[02/01/2024, 17:21:34] [homebridge-xbox-tv] This plugin threw an error from the characteristic 'Active Identifier': Unhandled error thrown inside read handler for characteristic: Cannot read properties of undefined (reading 'oneStoreProductId'). See https://homebridge.io/w/JtMGR for more info.
[02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Mute: ON
[02/01/2024, 17:21:34] [XboxTv] Device: 192.168.1.52 Xbox Series X, Volume: 0
Still seeing the error here too after updating
log.txt
So the steps to reproduce this are:
Restart Homebridge/child bridge while console is powered off (plugin successfully polls console and logs info)
Power on console (plugin still successfully polls console and logs info)
Power off the console (plugin polls console successfully for power/media state but errors getting input information).
Polling will continue to fail until either the console is powered back on or Homebridge/child bridge is restarted.
Please go to Homebridge/xboxTv/ folder and remove inputs, inputsNames and inputsTatgetVisibility files then restart plugin and let me know.
Updated to 2.12.5 and removed those files and that seems to have worked, thanks for your efforts @grzegorz914!
Updated and tested, everything is working great so far. Thanks @grzegorz914 ! Logs are staying clean, responsiveness of home actions and device state representation has improved back to similar duration prior to implementation of this plugin. Kudos, keep up the fantastic work. Plan on adding second xbox to this configuration and thoroughly testing so that I can roll out for the family. This is going to be game changer for them and my use case for automation of the AV systems.
| gharchive/issue | 2024-01-01T21:21:15 | 2025-04-01T04:34:26.748179 | {
"authors": [
"chris4prez",
"grzegorz914",
"insimbi-stuart"
],
"repo": "grzegorz914/homebridge-xbox-tv",
"url": "https://github.com/grzegorz914/homebridge-xbox-tv/issues/189",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2069225292 | 2.12.18 - Send connect request error: ReferenceError: tokenExist is not defined
Just updated to 2.12.18 before I was going to test and provide feedback on issues 191 - Enabling a second Xbox within the plugin failing
But before I could start testing, after update I'm getting the below errors and everything is no longer working. I blew away my entire configuration for child bridge, and single xbox, deleted home hub, xbox device and recreated all (performed auth, etc.) yet still receiving same error below without on/off working any more.
[1/7/2024, 12:48:36 PM] [homebridge-xbox-tv] Restarting child bridge...
[1/7/2024, 12:48:36 PM] Got SIGTERM, shutting down child bridge process...
[1/7/2024, 12:48:41 PM] [homebridge-xbox-tv] Child bridge process ended
[1/7/2024, 12:48:41 PM] [homebridge-xbox-tv] Process Ended. Code: 143, Signal: null
[1/7/2024, 12:48:48 PM] [homebridge-xbox-tv] Restarting Process...
[1/7/2024, 12:48:49 PM] [homebridge-xbox-tv] Launched child bridge with PID 52327
[1/7/2024, 12:48:49 PM] Registering platform 'homebridge-xbox-tv.XboxTv'
[1/7/2024, 12:48:49 PM] [homebridge-xbox-tv] Loaded homebridge-xbox-tv v2.12.18 child bridge successfully
[1/7/2024, 12:48:49 PM] Loaded 0 cached accessories from cachedAccessories.0E67955C5891.
[1/7/2024, 12:48:49 PM] Homebridge v1.7.0 (HAP v0.11.1) (homebridge-xbox-tv) is running on port 54592.
[1/7/2024, 12:48:57 PM] Living Room Xbox Series X 0EDB is running on port 43213.
[1/7/2024, 12:48:57 PM] Please add [Living Room Xbox Series X 0EDB] manually in Home app. Setup Code: 562-07-087
[1/7/2024, 12:49:00 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
[1/7/2024, 12:49:09 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
[1/7/2024, 12:49:20 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
[1/7/2024, 12:49:29 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
[1/7/2024, 12:49:39 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
[1/7/2024, 12:49:49 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
[1/7/2024, 12:49:59 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
[1/7/2024, 12:50:09 PM] [homebridge-xbox-tv] Device: 10.0.1.132 Living Room Xbox Series X, Send connect request error: ReferenceError: tokenExist is not defined
thanks issues resolved.
| gharchive/issue | 2024-01-07T17:56:53 | 2025-04-01T04:34:26.757876 | {
"authors": [
"chris4prez"
],
"repo": "grzegorz914/homebridge-xbox-tv",
"url": "https://github.com/grzegorz914/homebridge-xbox-tv/issues/194",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1844300158 | aria2的ARIANG_RPC_SECRET_AUTO默认参数建议设置为false
ARIANG_RPC_SECRET_AUTO false
如果设置为true,当远程访问aria2下载页面时,会在页面里默认设置为token的值,也就是任何人只要打开aria2页面就可以下载了,有风险。
当ARIANG_RPC_SECRET_AUTO设置为false,任何人打开aria2下载页面,如果不设置token的值,是无法下载的。
我自行测试,设置ARIANG_RPC_SECRET_AUTO ture时,在任何地方打开我的aria2就显示为连接状态。设置ARIANG_RPC_SECRET_AUTO false时,只有自己设置了token的浏览器才能进行下载,否则连不上。
当aria2远程下载暴露在公网时,别人不知道你的token值,也无法操作。
新版已默认设置为false。
| gharchive/issue | 2023-08-10T02:26:31 | 2025-04-01T04:34:26.769015 | {
"authors": [
"gshang2017",
"youland"
],
"repo": "gshang2017/docker",
"url": "https://github.com/gshang2017/docker/issues/202",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
54047692 | Filtering parameters similar to rails' filter_parameters
It would be useful to filter out sensitive details from logs at a global level - similar to how rails does this could also be inherited from the rails configuration
That's an application-level concern, not a logging one. Rails builds it into the framework and utilises those methods rather than their raw equivalent where appropriate.
| gharchive/issue | 2015-01-12T12:22:07 | 2025-04-01T04:34:26.770535 | {
"authors": [
"gshutler",
"stephenbinns"
],
"repo": "gshutler/hatchet",
"url": "https://github.com/gshutler/hatchet/issues/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1365167126 | Updating readme to be more accurate.
Signed-off-by: David Tippett dtip@amazon.com
This commit updates some of the readme to better reflect what is being run at the Wikimedia Foundation and Snagajob.
Just saw Flaxsearches PR so closing as duplicate XD
| gharchive/pull-request | 2022-09-07T20:47:13 | 2025-04-01T04:34:26.771836 | {
"authors": [
"dtaivpp"
],
"repo": "gsingers/opensearch-learning-to-rank-base",
"url": "https://github.com/gsingers/opensearch-learning-to-rank-base/pull/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2152902701 | Defer animation on list items until scrolled into view
Similar to https://pub.dev/packages/flutter_staggered_animations I would like to be able to only animate a list's items once they appear.
Regardless of scrolling, or items changing, they should only animate for the first time.
Sorry for the delayed response on this.
Interesting! I'll have to look at how they made that work. Worst case, I anticipate it would be pretty easy for the author of flutter_staggered_animations to create an animation type that would accept any flutter_animate instance.
Keeping open and will look into it when time permits. Feel free to take a stab at it in the meantime.
Okay, I've been giving this a bunch of thought, and have the start of a flexible API in mind that I'm fairly sure can be implemented:
foo.animate(hiddenBehavior: myBehavior);
HiddenBehavior(behavior, visibleFraction=0.5, resetOnHide=false, keepAlive=true)
// behaviors:
HiddenBehavior.PLAY // always play
HiddenBehavior.SKIP // skip to end if initially hidden
HiddenBehavior.PAUSE // remain paused at beginning if initially hidden
HiddenBehavior.WAIT // pause until visible
This would facilitate all of the behaviors I can think of:
current behavior (play on creation)
only play visible items, new items scroll in already at the end (or start) of their animation
play each item only when it first scrolls in
play every time it scrolls back in
Feedback is very welcome. I'm not 100% sure I love the naming of HiddenBehavior, but I think it's a bit more semantically intuitive than VisibilityBehavior (which is perhaps the more expected choice).
| gharchive/issue | 2024-02-25T19:17:56 | 2025-04-01T04:34:26.775388 | {
"authors": [
"gskinner",
"harkairt"
],
"repo": "gskinner/flutter_animate",
"url": "https://github.com/gskinner/flutter_animate/issues/131",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1829374758 | Update mlSearcher
фичи по тайм-ауту одной карты, рестарту сервера для каждой карты
новые карты из ветки Анны
убрал фичу с выбором количества сохраняемых агентов. пусть все сохраняются, в следующий раз не будет возможности забыть выставить правильное значение
фикс табличек: теперь там лежат кортежи <coverage, test_count, error_count, steps_count>
Что-то линтер ругается.
| gharchive/pull-request | 2023-07-31T14:52:45 | 2025-04-01T04:34:26.783571 | {
"authors": [
"emnigma",
"gsvgit"
],
"repo": "gsvgit/VSharp",
"url": "https://github.com/gsvgit/VSharp/pull/69",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1610239306 | support loading a collection which contains everything
Allow loading a collection of precomputed output. The input is a path to a collection that is a dictionary with keys X, Y and samples.
@emulate(collection="test.pkl")
def test(a, b):
pass
Codecov Report
Patch coverage: 52.94% and project coverage change: +23.20 :tada:
Comparison is base (27ce372) 11.16% compared to head (a18e472) 34.36%.
:mega: This organization is not using Codecov’s GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more
Additional details and impacted files
@@ Coverage Diff @@
## master #14 +/- ##
===========================================
+ Coverage 11.16% 34.36% +23.20%
===========================================
Files 9 9
Lines 412 419 +7
===========================================
+ Hits 46 144 +98
+ Misses 366 275 -91
Impacted Files
Coverage Δ
galilei/emulator.py
55.46% <52.94%> (+40.59%)
:arrow_up:
galilei/experimental.py
67.56% <0.00%> (+67.56%)
:arrow_up:
galilei/backends/sklearn.py
90.90% <0.00%> (+90.90%)
:arrow_up:
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
:umbrella: View full report at Codecov.
:loudspeaker: Do you have feedback about the report comment? Let us know in this issue.
| gharchive/pull-request | 2023-03-05T15:14:55 | 2025-04-01T04:34:26.811617 | {
"authors": [
"codecov-commenter",
"guanyilun"
],
"repo": "guanyilun/galilei",
"url": "https://github.com/guanyilun/galilei/pull/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
165726894 | sm2 加解密问题
老师,你好
我用sm2做加解密测试的时候,
加密可以出来结果
解密的时候失败了,这是什么问题
API的调用正确吗,还是其他的什么问题
int test_sm2_enc()
{
EC_KEY *key1;
EC_GROUP *group1;
int nid,ret;
key1 = EC_KEY_new();
if(!key1)
return 0;
int nid = NID_sm2p256v1;
group1 = EC_GROUP_new_by_curve_name(nid);
if(group1 == NULL)
return 0;
ret = EC_KEY_set_group(key1,group1);
if(ret != 1)
return 0;
ret = EC_KEY_generate_key(key1);
//EC_KEY_print_fp(stdout,key1,0);
/*---- ----*/
printf("----加密测试----\n");
char tmp_buf[48] = {0};
memset(tmp_buf,'1',48);
char out_buf[48+128] = {0};
int buf_size = 48 +128;
if(!SM2_encrypt_with_recommended(out_buf,&buf_size,
(const unsigned char *)tmp_buf,sizeof(tmp_buf),key1)) {
printf("SM2 Encryp ERROR !!!\n");
return -1;
}
else
{
printf("SM2 Encrypt Result[%d] :\n",buf_size);
//print_hex(out_buf,buf_size);
printf("----测试成功----\n\n");
}
printf("----解密测试----\n");
unsigned char msg[128] = {0};
size_t msglen = 0;
if(!SM2_decrypt_with_recommended(msg,&msglen,out_buf,buf_size,key1))
{
printf("解密失败!!!\n");
return -1;
}else {
printf("SM2 Decrypt Result[%d] :\n",msglen);
//print_hex(msg,msglen);
printf("----测试成功----\n\n");
}
return 0;
}
另外想问一个国密SSL的问题
现在代码树中定义了SSL_kECC 密钥交换算法,我想问这是什么算法
从文档上看和SSL_kRSA算法应该是差不多(用非对称密钥加密预主密钥),但是似乎openssl里面没有这种类似的算法
是不是要自己实现
----加密测试----
SM2 Encrypt Result[145] :
04 1F D6 11 66 1E 53 60 A0 C2 98 E4 78 08 7D 50 E6 4C 98 CC 0C 3E 68 7B 5C F5 D9 3E B3 CE 38 CA E0 D2 99 02 EF 2A 68 18 3F 80 2D 76 93 2C DD F5 3C 87 54 62 F0 C2 87 51 47 EC B9 E5 DA 63 57 1E 78 15 AE 1B 36 83 F1 58 0B 1C 8A 58 3F F8 EE 30 0A 6D 3C 25 E1 BC AE E3 52 CA CA 6C F4 31 E6 FD 32 34 A7 C7 24 EE 6A 07 91 02 9B 73 1F DC 8E BD 90 91 26 02 CB 7C 35 F2 B9 94 83 8A 5F C1 B7 F8 A4 62 F6 32 E0 B1 F3 F4 55 C5 39 9F DA EA 80 48 54
----测试成功----
----解密测试----
解密失败!!!
另外cryto/sm2/sm2test.c 编译之后运行也是出错的
sm2 sign p256 passed
sm2 sign b257 passed
error: sm2test.c 359
139772977178256:error:3406A06A:SM2 routines:SM2_decrypt:error:sm2_enc.c:520:
sm2 enc p256 failed
SM2 test failed
通过关注项目得到的信息是目前版本提供了完整的加解密示例
是不是我没找对示例代码的位置
sm2test.c可以看做是SM2加密签名的示例,涵盖了SM2内部接口和EVP接口的调用。
建议:
项目目前似乎没有.gitignore文件,个人感觉.gitignore在开发过程中比较重要,我从openssl项目中拷贝了.gitignore过来稍作改动可以参考一下
基于第一点,发现在最新一次更新中居然有109个文件的改动,其中大部分是.save文件,如果是必要的,最好单独提交一次commit,如果不是,.gitignore 中加入Makefile.sava或者*.save能避免不必要的更新
提议更新再频繁一点,比如一个功能做一次更新,一次更新夹杂太多的文件,容易淹没了真正需要关注的更改(比如最新一次更新),不利于其他开发者理解项目的进展情况,比如通过阅读代码我发现对于国密SSL的支持,本项目已经做了相当的改造,但是从commit 信息找到相关的信息就比较困难,可以将commit 细化,分类 (当然可能会增加工作量,我也只是建议)
最后,对于SM2加解密失败的问题,示例文件 sm2test.c 单独拿出来编译,也是解密会失败的,显然没有做过良好的测试,我做了简单的修改后可以正常用于测试了
提交了pull request 望采纳
https://github.com/guanzhi/GmSSL/pull/35
非常感谢您的建议!
.gitgnore我们后续我们会加上去
更新频率由目前开发方式导致暂时无法加快,并且由于暑期的影响近期频率会降低,不过我们正在准备社区化,期望在有更多贡献者加入后会形成对开发者比较友好的更新方式
处于和上面相同的原因,没有提前通知我们的pull request难以直接合并,不过我们会在理解完代码之后将您的代码并入,并通过主页和AUTHORS文件等方式注明您的贡献。
老师,目前我在做国密SSL的改造,想问几个关于密钥交换算法的问题
1.openssl中目前的密钥交换算法SSL_kECDHe 和 SSL_kEECH 算法有什么区别
2.GmSSL新定义的SSL_kECC算法又是怎么回事
3.GM/T 0024文档中描述的ECDHE 和ECC 密钥交换和以上的算法的对应关系是什么
SSL_kECC预计实现国密SSL的ECC-SM3-SM4套件,SSL_kECDHe和SSL_kEECH分别对应ECDHE-XXX、ECDH-XXX等SSL标准套件。
我是想问,国密套件中的ECDHE算法,比如(ECDHE_SM4_SM3),ECDHE对应的是SSL_kECDHe吗,如果是SSL_kECDHe的话,那按照openssl的实现
等于国密套件中的ECDHE,ECC和RSA密钥交换都没有ServerKeyExchange消息,但是GM/T 0024又描述了这几种消息(6.4.4.3),我觉得比较迷惑,还望解答一下
GM标准中的ECDHE对应双证书SM2密钥交换,ECC对应SM2公钥加密,这在SSL协议中没有对应的算法,因此不是SSL_kECDHe,需要定义新的SSL_kXXXXX。
就是说
1、ECC_SM4_SM3 对应的是SM2公钥加密,这在SSL协议中没有,需要自己定义实现
2、ECDHE_SM4_SM3 对应的【双证书SM2密钥交换】,不是SSL_kECDHe,也不是SSL_kEECDH,同样需要自己定义并实现?那这个【双证书SM2密钥交换】应该怎么理解?我应该查哪个相关的资料
SM2公钥加密密钥交换可以拿RSA密钥交换来作为参考,容易理解,
但是这个双证书是个什么概念呢,有什么可参考的呢
我这么理解:
服务器取服务器证书中的公钥,通过ServerKeyExchange发送
客户端取客户端证书中的公钥,通过ClientKeyExchange发送
通过SM2密钥交换算法,协商出预主密钥,实现密钥交换
全程完全使用证书中的公钥来做密钥交换,不会生成临时密钥对?(SSL_kECDHe,SSL_kEECDH都生成了临时密钥对)
是应该这样实现吗
关老师,我学习openssl时间不长,问题有点缺乏逻辑,还望包含
1.服务器都各有2对证书密钥(加密证书,签名证书),叫做双证书?
ECDHE算法,服务器,客户端都有两对SM2证书密钥
2.在密钥交换和消息认证中,无论是RSA,ECC,ECDHE都区分加密证书和签名证书,都需要双证书,不只是ECDHE_XXX套件需要双证书?
我开发过程中完全漏了双证书, 那等于在国密ssl的实现中,双证书才是重点,国密算法GmSSL项目都完全支持好了,快吐血了
服务器都各有2对证书密钥(加密证书,签名证书),这应该叫做单证书,单证书是指证书中的密钥仅支持加密或签名操作中的一种,我的理解。
在 2016年7月25日,18:59,XiaoTian <notifications@github.commailto:notifications@github.com> 写道:
服务器都各有2对证书密钥(加密证书,签名证书),叫做双证书
老师,我还想最后确认两个问题
1 无论是 RSA_XXX,ECC_XXX,ECDHE_XXX,在国密标准里服务器都需要使用双证书?
2 【双证书SM2密钥交换】,客户端需要有双证书吗,客户端的交换密钥是来自客户端的加密证书还是使用临时密钥?
老师
关于【国密SM2双证书的密钥交换】,我使用了服务器的加密密钥和客户端的临时密钥的方式实现(可以成功建立连接),这样做是合理的吗,还是客户端和服务器都必须要使用双证书
国密套件中的ECC_XXX 和 RSA_XXX套件实现也是要使用双证书吗(目前单证书已经实现了)
我要在下周结束SSL开发,还望老师帮忙确认一下这两个问题
http://www.cacrnet.org.cn/upload/fckeditor/File/2014/罗俊ssl vpn技术规范宣贯最终版0725挂网.pdf
其中有对12个密码套件使用的密钥和证书的说明,这个应该是比较权威的解释了。
太感谢了,正是我需要的信息!!!!
正式产品开发的话,建议去新华书店买一份纸质的规范,京东上应该也有。
网上的电子版不保证与实际发布的纸质版完全一致。
掌晓愚
在 2016年8月2日,17:27,xiyayadamaozai <notifications@github.commailto:notifications@github.com> 写道:
太感谢了,正是我需要的信息!!!!
―
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHubhttps://github.com/guanzhi/GmSSL/issues/33#issuecomment-236852005, or mute the threadhttps://github.com/notifications/unsubscribe-auth/AL-bho9aZ13rZYY0_qyN_xaeWsDpqKKdks5qbw1sgaJpZM4JNJLZ.
你好,老师,现在GmSSL编译的能编译支持arm64的么?你们给的编译后只支持x86_64的.
支持的,我测试过的,arm-linux 下编译运行都没问题
我是想把GmSSL编译成静态库用于iOS开发,是在mac环境下编译的.发现不支持arm64的.
iOS模拟器和arm32支持吗?
你好,在mac环境下编译的我发现只支持x86_64,也就是说只支持iOS模拟器,对于iOS开发真机arm64是不支持的。
发自 网易邮箱大师
在2016年08月13日 05:32,Zhi Guan 写道:
iOS模拟器和arm32支持吗?
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub, or mute the thread.
你可以把编译错误信息帖上来看看
另外请新开一个issue
你好,这是iOS开发真机报错:
ld: warning: ld: warning: ignoring file /Users/a123456/Desktop/SM2 SM4/libcrypto.a, file was built for archive which is not the architecture being linked (arm64): /Users/a123456/Desktop/SM2 SM4/libcrypto.aignoring file /Users/a123456/Desktop/SM2 SM4/libssl.a, file was built for archive which is not the architecture being linked (arm64): /Users/a123456/Desktop/SM2 SM4/libssl.a
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(CharacterKeyboard.o)) was built for newer iOS version (9.0) than being linked (7.0)
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(NumberKeyboardThird.o)) was built for newer iOS version (9.0) than being linked (7.0)
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(KeyboardTool.o)) was built for newer iOS version (9.0) than being linked (7.0)
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(Keyboard.o)) was built for newer iOS version (9.0) than being linked (7.0)
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(XYKeyboard.o)) was built for newer iOS version (9.0) than being linked (7.0)
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(NumberKeyboardFirst.o)) was built for newer iOS version (9.0) than being linked (7.0)
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(NumberKeyboardSecond.o)) was built for newer iOS version (9.0) than being linked (7.0)
ld: warning: object file (/Users/a123456/Desktop/SM2 SM4/zhlcs/Classes/SDK/XYKeyboard/libXYKeyboard.a(UIView+LVExtension.o)) was built for newer iOS version (9.0) than being linked (7.0)
Undefined symbols for architecture arm64:
"_EVP_sm3", referenced from:
_test_sm2_enc in sm2Plugin.o
"_EC_POINT_point2hex", referenced from:
_test_sm2_enc in sm2Plugin.o
"_EC_KEY_new", referenced from:
-[sm2Plugin sm2:] in sm2Plugin.o
"_EC_POINT_point2oct", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_KDF_get_x9_63", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_CRYPTO_malloc", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_BN_new", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_GROUP_get_order", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EVP_DigestInit_ex", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_POINT_new", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_BN_CTX_new", referenced from:
_SM2_do_encrypt in sm2_enc.o
-[sm2Plugin sm2:] in sm2Plugin.o
_test_sm2_enc in sm2Plugin.o
"_EVP_MD_CTX_create", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_KEY_get0_group", referenced from:
_SM2_do_encrypt in sm2_enc.o
-[sm2Plugin sm2:] in sm2Plugin.o
_test_sm2_enc in sm2Plugin.o
"_EC_GROUP_get_cofactor", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_KEY_get0_public_key", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_GROUP_get_degree", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_POINT_hex2point", referenced from:
-[sm2Plugin sm2:] in sm2Plugin.o
"_BN_rand_range", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_OpenSSLDie", referenced from:
_SM2_do_encrypt in sm2_enc.o
_test_sm2_enc in sm2Plugin.o
"_BN_free", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EVP_DigestUpdate", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EVP_DigestFinal_ex", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_POINT_mul", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_POINT_free", referenced from:
_SM2_CIPHERTEXT_VALUE_free in sm2_enc.o
"_BN_CTX_free", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_BN_num_bits", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_POINT_is_at_infinity", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EVP_MD_CTX_destroy", referenced from:
_SM2_do_encrypt in sm2_enc.o
"_EC_KEY_new_by_curve_name", referenced from:
-[sm2Plugin sm2:] in sm2Plugin.o
"_CRYPTO_free", referenced from:
_SM2_CIPHERTEXT_VALUE_free in sm2_enc.o
ld: symbol(s) not found for architecture arm64
clang: error: linker command failed with exit code 1 (use -v to see invocation)
| gharchive/issue | 2016-07-15T07:24:05 | 2025-04-01T04:34:26.846407 | {
"authors": [
"LiTianjue",
"conezxy",
"guanzhi",
"sinv",
"tongyu123"
],
"repo": "guanzhi/GmSSL",
"url": "https://github.com/guanzhi/GmSSL/issues/33",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2116687355 | 🛑 Sefaz - Rio Grande do Sul is down
In 42ce968, Sefaz - Rio Grande do Sul (https://onboardapi.guichepass.com.br/sefaz?code=2) was down:
HTTP code: 500
Response time: 157 ms
Resolved: Sefaz - Rio Grande do Sul is back up in 8cd33f9 after 1 hour, 13 minutes.
| gharchive/issue | 2024-02-03T21:41:26 | 2025-04-01T04:34:26.894957 | {
"authors": [
"suporte-gpass"
],
"repo": "guichevirtual/statuspage",
"url": "https://github.com/guichevirtual/statuspage/issues/1814",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1651005455 | css style border-radius not work
I tried with to generate image with html and css style like below command:
+++HTML `
<meta charset="UTF-8">
<body>
<img src="${basics.image}" alt="HTML5 Icon" width="100" height="100" style="border-radius: 50%;">
</body>
`+++
I expected it will generate an image with the circle shape, BUT it didn't.
If you're sure your HTML itself is correct, please read up on the limitations of HTML altChunk in MS Word. It is possible that what you are trying to do is not supported by MS Word.
| gharchive/issue | 2023-04-02T16:13:38 | 2025-04-01T04:34:26.907188 | {
"authors": [
"jjhbw",
"minuth"
],
"repo": "guigrpa/docx-templates",
"url": "https://github.com/guigrpa/docx-templates/issues/304",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2465271471 | [BUG] autoFocus not triggering keyboard on iOS devices
Version of the library: 1.2.4
Expected Behavior
When the OTPInput component is rendered with autoFocus, the first slot should be focused, and the keyboard should automatically appear on iOS devices.
Current Behavior
When using the OTPInput component with the autoFocus prop set to true, the first slot is correctly focused on iOS devices (both web and native), but the keyboard does not automatically appear. Users must tap on the focused slot to make the keyboard appear.
Steps to Reproduce
Create a new project using the input-otp library
Implement the OTPInput component with the autoFocus prop set to true
Run the project on an iOS device (either in a web browser or as a native app)
Observe the behavior when the OTPInput component is rendered
I've created a CodeSandbox to demonstrate this issue
I also noticed this problem today on iOS devices
How awkward. No idea why the keyboard wouldn't be triggering...
Hey @MorganMeirFitussi I just found out that won't work with iOS. It won't work with a normal <input />, too. See this is the only workaround and it's not pretty https://gist.github.com/cathyxz/73739c1bdea7d7011abb236541dc9aaa?permalink_comment_id=4412959#gistcomment-4412959
Even tho there's a workaround available for that, it shouldn't be responsibility of input-otp's but more of a general knowledge on iOS inputs.
| gharchive/issue | 2024-08-14T08:54:38 | 2025-04-01T04:34:26.917327 | {
"authors": [
"ETOPS7",
"MorganMeirFitussi",
"guilhermerodz"
],
"repo": "guilhermerodz/input-otp",
"url": "https://github.com/guilhermerodz/input-otp/issues/61",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
31942025 | Use nodechecker.com to figure out if a module has passing tests
We can sort them a bit lower if the tests are failing or don't exist
Example: http://nodechecker.com/#info/gulp
We may want to host our own since the interval at which they scan may not be fast enough and we can filter out only gulpplugin/gulpfriendly modules
Closing this because the plugin page has been rewritten and integrated with our main website: https://github.com/gulpjs/gulpjs.github.io
| gharchive/issue | 2014-04-22T03:01:11 | 2025-04-01T04:34:26.936809 | {
"authors": [
"contra",
"phated"
],
"repo": "gulpjs/plugins",
"url": "https://github.com/gulpjs/plugins/issues/76",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1305724521 | Local user?
Hi,
I was wondering if it is possible to use a local user to set the alarm, i.e not one linked to a hie-connect account with a simple pin code?
As I just want HA to be able to set not but not able to unset the alarm for safety.
Thanks
James
Hi,
Unfortunately local users dont have ISAPI api access. So It can't be used in this integration.
But I'll check a way to disable disarm functionality.
Or, It can be forced to ask a password (with a complexity of your choice) when disarm is clicked at HA.
Both needs some development.
Let me know which will help you most. Then I can start development of it.
I vote for keycode to use alarm control panel integration and alarm-panel which is included in hassio:
https://www.home-assistant.io/dashboards/alarm-panel/
Hi
Thanks.
I would go for a key code which we can set to disarm would be most useful for the masses.
Having looking at the user options, I can set a arm only account up on Hik-connect so this solves the need to disable the disarm function in first instances. I will trial over the weekend/next week with a separate Hik—connect account and update here for other peoples reference.
Hi,
So, I'll start implementing the keycode soon.
Hello,
Thanks you so much for having made the HACS integration of ax pro. I have switched my 100 devices to Home assistant last weekend because of that!
Please don't make the keycode mandatory. The current integration is great. We can arm or disarm with geolacation for example. If you implement the keycode, please make it optional.
Thank you.
I used the following script to test
`from hikaxpro import HikAxPro
axpro = HikAxPro("IP_ADDRESSS", "HIK_CONNECT_USERNAME", "PASSWORD")
res = axpro.zone_status()
print(res) `
| gharchive/issue | 2022-07-15T08:12:21 | 2025-04-01T04:34:26.951106 | {
"authors": [
"DejanBukovec",
"gunkutzeybek",
"nicolasv55",
"obrien-james"
],
"repo": "gunkutzeybek/hikaxpro_hacs",
"url": "https://github.com/gunkutzeybek/hikaxpro_hacs/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
852473587 | Please share your processed data set (via Google or Baidu Cloud Disk), thank you
Please share your processed data set (via Google or Baidu Cloud Disk), thank you.For some reasons, it is more difficult to download data from the original site. I will only use the data for research
Sure, I will release it soon. The datasets are more than 400 GB, so it will take a while to upload.
Thank you very much.
Data is released.
I cannot access preprocessed data...
Is Google or Baidu Cloud Disk available?
thx thx thx...
Hi Kingaza,
I’m trying to resolve this problem. The dataset is too big (around 400 GB). I need find an appropriate place to host the dataset.
Best,
Pengfei
From: kingaza @.>
Sent: Wednesday, June 9, 2021 1:44 AM
To: guopengf/FL-MRCM @.>
Cc: guopengf @.>; State change @.>
Subject: Re: [guopengf/FL-MRCM] Please share your processed data set (via Google or Baidu Cloud Disk), thank you (#3)
I cannot access preprocessed data...
Is Google or Baidu Cloud Disk available?
thx thx thx...
—
You are receiving this because you modified the open/close state.
Reply to this email directly, https://github.com/guopengf/FL-MRCM/issues/3#issuecomment-857394856 view it on GitHub, or https://github.com/notifications/unsubscribe-auth/AHUEVDZ76JIGWR4343DLZJLTR35SJANCNFSM42Q5DA3Q unsubscribe. https://github.com/notifications/beacon/AHUEVD3VXZPAHEYEI663JDTTR35SJA5CNFSM42Q5DA32YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOGMNM5KA.gif
thx again~
| gharchive/issue | 2021-04-07T14:31:37 | 2025-04-01T04:34:26.978562 | {
"authors": [
"guopengf",
"kingaza",
"zym1130232"
],
"repo": "guopengf/FL-MRCM",
"url": "https://github.com/guopengf/FL-MRCM/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1570102754 | 点击“提交”希望能有个图标表示提交成功,不然界面没有变化,一直感觉没有提交成功
希望大佬后续能加个图标,比如加个对号什么的,有个提示
想要這個功能 +1
| gharchive/issue | 2023-02-03T16:26:20 | 2025-04-01T04:34:26.979904 | {
"authors": [
"Donge-wx",
"trevim99"
],
"repo": "guopenghui/obsidian-language-learner",
"url": "https://github.com/guopenghui/obsidian-language-learner/issues/95",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
338501665 | 能不能尊重一下我的劳动成果?
应该署名参考我的github吧。。。这个分类和里面的内容和我的一模一样。
https://github.com/AweiLoveAndroid/CommonDevKnowledge/blob/master/interview/summary.md
不好意思,之前没看到你这个repo,从掘金搜了一下面试题集,整理了一下问题,写了下答案。
@AweiLoveAndroid 整合内容的目的只是为了让别人宣传你的名字吗?
我认为分享的意义是帮助他人,我知道你的初心不是不想帮助他人,但如果原文链接对你非常重要,我建议你以后不要公开发表,因为你不能保证所有人都不会犯错。
谢绝撕逼,世界和平。我只是搜集了各种面试题,并把自己博客的内容作为答案放在了里面,本来是自己备忘用的,因为是MIT协议,被鸿洋发到公众号了。我把所有的网页浏览记录都加上了。
另外有两点:
1 我只是搜集问题,问题的答案来源于我的博客。
2 不要要搞个大新闻,在知乎上发个阿里人抄袭你是什么意思,这是我入职阿里之前写的东西,不要动不动就把公司扯上。
| gharchive/issue | 2018-07-05T09:45:58 | 2025-04-01T04:34:26.984228 | {
"authors": [
"AweiLoveAndroid",
"guoxiaoxing",
"tcqq"
],
"repo": "guoxiaoxing/android-interview",
"url": "https://github.com/guoxiaoxing/android-interview/issues/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
312579365 | Try to handle the case where a element is missing an r attribute
I've run into an Excel spreadsheet where some <c> elements do not have address references. It looks like it's some kind of short hand where it's implicitly the next cell, and Excel opens the spreadsheet fine. Here's a pretty-printed snippet from the spreadsheet:
<sheetData>
<row r="1" spans="1:1" ht="18" customHeight="1">
<c r="A1" s="1" t="s">
<v>0</v>
</c>
</row>
<row r="2" spans="1:1" ht="12.75" customHeight="1">
<c r="A2" s="2" t="s">
<v>1</v>
</c>
</row>
<row r="3" spans="1:1" ht="12.75" customHeight="1">
<c r="A3" s="2" t="s">
<v>2</v>
</c>
</row>
<row r="5" spans="1:12" ht="12.75" customHeight="1">
<c r="A5" s="4" t="s">
<v>51</v>
</c>
<c s="4" t="s">
<v>52</v>
</c>
<c s="4" t="s">
<v>53</v>
</c>
<c s="4" t="s">
<v>54</v>
</c>
...
Exceljs breaks with:
(node:3101) TypeError: Cannot read property 'match' of undefined
at Object.decodeAddress (/Users/andreaslind/work/exceljs/lib/utils/col-cache.js:93:21)
at value.cells.forEach.cellModel (/Users/andreaslind/work/exceljs/lib/doc/row.js:325:46)
at Array.forEach (<anonymous>)
at module.exports.set model [as model] (/Users/andreaslind/work/exceljs/lib/doc/row.js:319:17)
at model.rows.forEach.rowModel (/Users/andreaslind/work/exceljs/lib/doc/worksheet.js:595:17)
at Array.forEach (<anonymous>)
at module.exports._parseRows (/Users/andreaslind/work/exceljs/lib/doc/worksheet.js:592:16)
at module.exports.set model [as model] (/Users/andreaslind/work/exceljs/lib/doc/worksheet.js:606:10)
at value.worksheets.forEach.worksheetModel (/Users/andreaslind/work/exceljs/lib/doc/workbook.js:197:23)
at Array.forEach (<anonymous>)
(The line numbers might be off, as I'm currently running an older version, unfortunately)
I'm not sure the enclosed solution is per spec. Maybe someone else knows more?
And yes, it's actually the same spreadsheet that caused https://github.com/guyonroche/exceljs/pull/536 -- I'll follow up if I'm able to find out how it was generated.
I just learned that the Excel file in question is an export from Xero, which is a piece of accounting software.
| gharchive/pull-request | 2018-04-09T15:32:53 | 2025-04-01T04:34:27.026480 | {
"authors": [
"papandreou"
],
"repo": "guyonroche/exceljs",
"url": "https://github.com/guyonroche/exceljs/pull/537",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
164661792 | almost ready
基本写完了
实现了内置的校验函数
datetime 和 date可能会有bug
unicode,dict,list这3个不用,dict,list在schema.py里面实现了,unicode去掉,不打算支持python2。
还差个枚举enum
add_validater,remove_validater也不需要,去掉。
| gharchive/pull-request | 2016-07-09T10:45:25 | 2025-04-01T04:34:27.028571 | {
"authors": [
"guyskk",
"vibiu"
],
"repo": "guyskk/validater",
"url": "https://github.com/guyskk/validater/pull/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2533041182 | 会员视频无法下载
https://www.bilibili.com/bangumi/play/ep835859?spm_id_from=333.999.0.0
例如上面的,显示解析错误
登录的账号是 会员账号吗 😂
我不是会员账号没试过下载会员视频
肯定是大会员账号,不只是会员视频,就像是这种普通视频,只要地址是这种格式,就解析错误。
https://www.bilibili.com/bangumi/play/ep835674?spm_id_from=333.999.0.0&from_spmid=666.25.episode.0
非会员的这种格式已于v0.0.9修复, 会员的视频 我这边就无法测了有空的话麻烦帮忙测下😂
会员视频解析错误:Error: TypeError: Cannot read properties of undefined (reading 'duration')
肯定是大会员账号,不只是会员视频,就像是这种普通视频,只要地址是这种格式,就解析错误。
bilibili.com/bangumi/play/ep835674?spm_id_from=333.999.0.0&from_spmid=666.25.episode.0
如果用非会员的连接弹窗后 点会员的那一集下载呢 会报错吗
肯定是大会员账号,不仅仅是会员视频,就像是这种普通视频,只要地址是这种格式,就解析错误。
bilibili.com/bangumi/play/ep835674?spm_id_from=333.999.0.0&from_spmid=666.25.episode.0
如果用非会员的连接弹窗后点会员的那一套下载呢会报错吗
会卡死
试下用这个版本 看看 https://github.com/gxr404/BilibiliVideoDownloadFork/releases/tag/v0.0.10
我这边借了个号 试了下 没啥问题
试下用这个版本看看https://github.com/gxr404/BilibiliVideoDownloadFork/releases/tag/v0.0.10
我借了个号试了下没啥问题
是可以下载了,但是地址不对,显示的多P视频就不是视频本身
https://www.bilibili.com/bangumi/play/ep835914?spm_id_from=333.999.0.0
没懂啥意思 该会员视频链接 是个合集 含其他多p视频 所以填入 会员视频和 非会员视频 都会弹窗,如果只需下其中某一集 弹窗里选择即可
没懂啥意思会员视频链接是个合集含其他多p视频所以填入会员视频的和非会员视频都会弹窗,如果只需下其中一个集弹窗里选择即可
需要的视频就不在多p列表,你打开看一下,要下载是右侧列表里的,而不是上面的正片
你是指 要下载的 红框内的 而不是 蓝框内的?
这个有点特殊😂
你是指要下载的红框内的而不是蓝框内的? 这个有点特殊😂
对,应该是B站隐藏了下方视频的真实地址,上面总挂着3个正片
可以下载了,但是还有一点小瑕疵,
https://www.bilibili.com/bangumi/play/ep835859?spm_id_from=333.999.0.0
这个视频有4K版本,但是上面的可选分辨率最高只有1080P高码率
大概看了下
如果使用 仅支持1080P高码率的视频链接 打开弹窗 选项只有1080P高码率
如果使用 支持4k的视频链接 打开弹窗 就有 选项4k
OK,完美解决,感谢大神!
如果选择4k下载 视频仅支持1080高码率 则 会自动降级到1080高码率(降级功能一直都有),
只是现在v0.0.11版本显示有问题 没显示视频真实的分辨率,
在v0.0.12修复了
| gharchive/issue | 2024-09-18T08:14:58 | 2025-04-01T04:34:27.090052 | {
"authors": [
"gxr404",
"youzhui"
],
"repo": "gxr404/BilibiliVideoDownloadFork",
"url": "https://github.com/gxr404/BilibiliVideoDownloadFork/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2471768721 | 🛑 Undertale社区档案馆 - UTCARC is down
In bbae30c, Undertale社区档案馆 - UTCARC (https://arc.utcwiki.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Undertale社区档案馆 - UTCARC is back up in 90289c7 after 11 minutes.
| gharchive/issue | 2024-08-18T03:24:24 | 2025-04-01T04:34:27.104013 | {
"authors": [
"gzombiejun"
],
"repo": "gzombiejun/upptime",
"url": "https://github.com/gzombiejun/upptime/issues/592",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1315688650 | Require conda package
[ ] Recipe to build conda package
[ ] Step to build conda package
[ ] Publish conda package to h2oai channel
superseded by https://github.com/h2oai/authn-py/issues/72
| gharchive/issue | 2022-07-23T15:51:08 | 2025-04-01T04:34:27.116405 | {
"authors": [
"Mathanraj-Sharma",
"zoido"
],
"repo": "h2oai/authn-py",
"url": "https://github.com/h2oai/authn-py/issues/31",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
199415331 | Error on gradle
I got this error when I want to use this library and put it on gradle:
Error:Could not GET 'https://repos.zeroturnaround.com/nexus/content/repositories/zt-public-releases/com/zeroturnaround/jrebel/android/jr-android-gradle/maven-metadata.xml'. Received status code 401 from server: Unauthorized
Enable Gradle 'offline mode' and sync project
What should I do to resolve it?
Hi. That is not an issue of this library.
It seems that the maven-metadata.xml file is protected with basic auth. I guess the file is used by JRabel plugin, so try disabling it. Thanks.
| gharchive/issue | 2017-01-08T10:50:49 | 2025-04-01T04:34:27.250139 | {
"authors": [
"h6ah4i",
"yasaman93"
],
"repo": "h6ah4i/android-advancedrecyclerview",
"url": "https://github.com/h6ah4i/android-advancedrecyclerview/issues/335",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
509225401 | Define list of packages to download
As part of https://github.com/habitat-sh/habitat/issues/6902 we want to provide some pre-defined lists of packages.
Note: the files here use the '#' comment character and whitespace for clarity; and won't work with habitat until issue https://github.com/habitat-sh/habitat/issues/7040 is addressed.
Background: Historically we provided a tarball 'LATEST.tgz' that contained all the packages in core
for all targets. That amounted to about 15GB, and was both too much and too little, in that many of
the packages weren't needed, and for many patterns (effortless) other origins were needed.
This is an initial attempt to provide some starter lists of packages.
The basic naming pattern is TASK_ARCH_CHANNEL. This due to a limitation of the input file
format. The simple newline separated list of package idents doesn't allow for specification of
channel or target architecture inline, so we're using a file naming convention to
The current tasks are
builder (setting up an on prem builder)
core_deps (a reduced starter set from core with common build time deps)
core_full (everything for a particular architecture)
effortless (starter set for the effortless pattern)
Each is broken out by the architecture and channel required; to complete some two downloads, once
from stable and once from unstable will be required.
For example, to get the complete effortless infrastructure for linux,
hab pkg download --download-directory download_pkgs --channel=unstable --target x86_64-linux --file quickstart_lists/effortless_x86_64-windows_unstable
hab pkg download --download-directory download_pkgs --channel=stable --target x86_64-linux --file quickstart_lists/effortless_x86_64-windows_stable
Closes #6902
Signed-off-by: Mark Anderson mark@chef.io
A few minor comments, overall looking good @markan
| gharchive/pull-request | 2019-10-18T17:45:51 | 2025-04-01T04:34:27.285655 | {
"authors": [
"chefsalim",
"markan"
],
"repo": "habitat-sh/on-prem-builder",
"url": "https://github.com/habitat-sh/on-prem-builder/pull/210",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1897276538 | 📓 Пользователь может авторизоваться
Пользовательская история и сценарии
Функция:
Я, как зарегистрированный пользователь
Хочу иметь возможность войти в систему
Чтобы начать пользоваться функционалом платформы
Сценарий: Успешная авторизация пользователя
Допустим существует зарегистрированный пользователь с логином 'test@example.com' и паролем '12345'
Когда пользователь осуществляет авторизацию с использованием логина 'test@example.com' и пароля '12345'
Тогда пользователю выдается токен авторизации
Сценарий: Неудачная авторизация пользователя
Допустим существует зарегистрированный пользователь с логином 'test@example.com' и паролем '12345'
Когда пользователь осуществляет авторизацию с использованием логина 'test@example.com' и пароля '54321'
Тогда пользователь получает ошибку авторизации
Дополнительная информация
No response
Функционал авторизации делаю выключенным, так как пока у нас нет развернутого Kratos на стендах
| gharchive/issue | 2023-09-14T20:40:37 | 2025-04-01T04:34:27.287347 | {
"authors": [
"picolino"
],
"repo": "habralab/garnet-team",
"url": "https://github.com/habralab/garnet-team/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2414882311 | cuteOneLineFlower-emma
Generates a three-petal colored flower. Uses random seed value to move the position of the flower, stem, and leaves, as well as change the shape of the petals slightly. This is my first coding project and I'm a beginner coder :)
[x] I have read the steps to getting a blot
[x] I am submitting art that...
[x] is algorithmically generated (will change each time the program is run)
[x] is drawable on a blot (fits in the work area & doesn't overlap too much)
[x] is original (not copied from somewhere else)
[ ] Optional, if you used a tutorial or based your art on something else, please include the link here:
[ ] Optional, if you remixed this from something else, mention it here:
Seeing as this hasn't been updated in over a month, I'm going to close the PR for now. You're free to make a new one in the future with the same artwork if you'd like! Reach out to me @ alexren on the slack if you have any questions!
| gharchive/pull-request | 2024-07-18T00:11:18 | 2025-04-01T04:34:27.292981 | {
"authors": [
"emma-x1",
"qcoral"
],
"repo": "hackclub/blot",
"url": "https://github.com/hackclub/blot/pull/710",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
646777284 | update scrappy user on profile change
Updates github and website on profile change
This will update the profile picture every time anything in the user's profile changes instead of only when the profile picture is updated.
Cool, this looks great! Merging now, will test once merged and fix any bugs that arise. Thanks so much 🙌
| gharchive/pull-request | 2020-06-27T22:12:54 | 2025-04-01T04:34:27.297820 | {
"authors": [
"MatthewStanciu",
"saharshy29"
],
"repo": "hackclub/scrappy",
"url": "https://github.com/hackclub/scrappy/pull/32",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2486220562 | [common] 해커톤 앱 제출
제목
팀 앱 제출 1차
팀 이름
common
팀 리포지토리
https://github.com/hackersground-kr/hg-common-hackergroud
IntelliJ 닫기 버튼을 눌러 초기 화면으로 어떻게 가는지 모르겠어요.
/invalid
| gharchive/issue | 2024-08-26T08:04:39 | 2025-04-01T04:34:27.306073 | {
"authors": [
"clcok",
"yeseong0412"
],
"repo": "hackersground-kr/hackers-ground",
"url": "https://github.com/hackersground-kr/hackers-ground/issues/951",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2011254203 | Texture the player prefab
The player prefabs needs to be UV-Mapped and get some textures
Done, at least partially. What remains is improvements.
| gharchive/issue | 2023-11-26T23:07:54 | 2025-04-01T04:34:27.306857 | {
"authors": [
"Fueredoriku",
"toberge"
],
"repo": "hackerspace-ntnu/Red-Planet-Rampage",
"url": "https://github.com/hackerspace-ntnu/Red-Planet-Rampage/issues/336",
"license": "BSD-4-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
266612187 | BUY TICKETS INDIVIDUALLY
No way to distinguish if ppl buy multiple tickets so changed policy to only allowing individual purchases.
Is this enforced in the event brite?
No, people can order multiple tickets. I could modify the Eventbrite to only allow one ticket
| gharchive/pull-request | 2017-10-18T19:22:13 | 2025-04-01T04:34:27.382213 | {
"authors": [
"casey-chow",
"dfan97"
],
"repo": "hackprinceton/static",
"url": "https://github.com/hackprinceton/static/pull/22",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
545080314 | Error in opening the audio file
As you have mentioned in the document, I replaced the path of the mp3 file. But it is showing the following error:
Traceback (most recent call last):
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media\codecs\wave.py", line 59, in init
self._wave = wave.open(file)
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\wave.py", line 510, in open
return Wave_read(f)
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\wave.py", line 164, in init
self.initfp(f)
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\wave.py", line 131, in initfp
raise Error('file does not start with RIFF id')
wave.Error: file does not start with RIFF id
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "voice_based_email_for_blind.py", line 26, in
music = pyglet.media.load(ttsname, streaming = False)
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media_init_.py", line 143, in load
raise first_exception
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media_init_.py", line 133, in load
loaded_source = decoder.decode(file, filename, streaming)
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media\codecs\wave.py", line 111, in decode
return StaticSource(WaveSource(filename, file))
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\site-packages\pyglet\media\codecs\wave.py", line 61, in init
raise WAVEDecodeException(e)
pyglet.media.codecs.wave.WAVEDecodeException: file does not start with RIFF id
Please help me to solve this issue .
@exist-world It seems like you have an invalid or unsupported audio file. Recall that only FLAC, AIFF, and RIFF WAV files are supported.
If it's one of those, and it's still not working, please let me know.
May I know your os?
Try this. Read the file with librosa, then convert it back to a temporary .wav file. Then read it back with the 'wave' package.
https://stackoverflow.com/questions/25672289/failed-to-open-file-file-wav-as-a-wav-due-to-file-does-not-start-with-riff-id/57349558#57349558
thanks for ur suggestion i will try and tell
My os is windows 10
Now that audio file is working...but choosing option 1 gives me following error...
Traceback (most recent call last):
File "voice_pro.py", line 107, in
mail.login('emailID','pswrd') #login part
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\smtplib.py", line 730, in login
raise last_exception
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\smtplib.py", line 721, in login
initial_response_ok=initial_response_ok)
File "C:\Users\ELCOT\AppData\Local\Programs\Python\Python37\lib\smtplib.py", line 642, in auth
raise SMTPAuthenticationError(code, resp)
smtplib.SMTPAuthenticationError: (535, b'5.7.8 Username and Password not accepted. Learn more at\n5.7.8 https://support.google.com/mail/?p=BadCredentials i17sm4151845pfr.67 - gsmtp')
And chosing option 2 gives me the following error
Traceback (most recent call last):
File "voice_pro.py", line 89, in
if int(text) == 1:
ValueError: invalid literal for int() with base 10: 'Tu'
@exist-world bro, Here [mail.login('emailID','pswrd') #login part] You have to put your email id and password for login . After that if any error occurs then please tell me.
yeah ..thank you !..Now 1 st option is working but in victim id i gave receiver mail id...it shows mail sent But receiver didnt get that mail
And giving 2 nd option shows me the following error
Traceback (most recent call last):
File "voice_pro.py", line 89, in
if int(text) == 1:
ValueError: invalid literal for int() with base 10: 'Tu'
Help me to solve this one..!
@exist-world Its because of 'Tu' not '2'. I mean script recognise the 'Tu' not 2 and that's why it fetch the error.
Just edit it and add [ if int(text) == 2 or text == "tu": ] in the second if statement not the first one.
Now, I really don't know why u have not received the mail in inbox. Have you checked in spam folder? If not then use your primary email id in both as a sender and receiver email id.
No i dont use any of that kind..its okay...but whenever i use option 2 its seems like its checking for option 1 only...its shows folln error
You said : Tu
Traceback (most recent call last):
File "voice_pro.py", line 90, in
if int(text)==1 or text == 'one':
ValueError: invalid literal for int() with base 10: 'Tu'
@exist-world Sorry for late response. Paste those with folln codes:
if text == '1' or text == 'One' or text == 'one':
elif text == '2' or text == 'tu' or text == 'two' or text == 'Tu' or text == 'to' or text == 'To' :
Soon I will update the latest one. As I told you before it was almost 3 years old project.
Okay ..pls upload the latest one as soon as possible...Thanks in advance
i have one doubt...you gave mail id of the receiver as static in code...for getting dynamic input from the user for mail id how to use ?? Please help quickly to solve this out..
@afuafu21 It arises some minor problem. I have added this feature in the latest one. But the problem is I have not enough time to test it. So, Its getting a bit late to upload. And there are another problem, if someone's email id is alphanumeric then the problem occurs, till now it should be in alphabet only.
Ohh Its okay...When u will upload the latest one.please upload soon or if u
don't mind send it to my mail...Because this month 18 is the last date for
my project submission and review....if u upload soon it is so helpful for
me....because of your contribution only i did this much in that project
Thanks a lot...
On Thu, Mar 5, 2020, 10:03 AM Sayak Naskar notifications@github.com wrote:
@afuafu21 https://github.com/afuafu21 It arises some minor problem. I
have added this feature in the latest one. But the problem is I have not
enough time to test it. So, Its getting a bit late to upload. And there are
another problem, if someone's email id is alphanumeric then the problem
occurs, till now it should be in alphabet only.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/hacky1997/voice-based-email-for-blind/issues/5?email_source=notifications&email_token=AOU64DHYZ75QVF2P7C4HO2TRF4TRDA5CNFSM4KCPYZU2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEN3VMVA#issuecomment-595023444,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AOU64DFXQNAT7TVKYMEEOY3RF4TRDANCNFSM4KCPYZUQ
.
@afuafu21 Please provide your email id. I have not tested it yet.
Okay Thank you this is my mail id-mini20pro20@gmail.com
On Fri, Mar 6, 2020, 4:30 PM Sayak Naskar notifications@github.com wrote:
@afuafu21 https://github.com/afuafu21 Please provide your email id. I
have not tested it yet.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/hacky1997/voice-based-email-for-blind/issues/5?email_source=notifications&email_token=AOU64DHTJUFKAGJRIG2UQIDRGDJWDA5CNFSM4KCPYZU2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEOA64MA#issuecomment-595717680,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/AOU64DFMGXSGFUXFRD6ABPDRGDJWDANCNFSM4KCPYZUQ
.
Issue has been resolved.
| gharchive/issue | 2020-01-03T17:14:41 | 2025-04-01T04:34:27.407192 | {
"authors": [
"afuafu21",
"exist-world",
"hacky1997"
],
"repo": "hacky1997/voice-based-email-for-blind",
"url": "https://github.com/hacky1997/voice-based-email-for-blind/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
437048585 | Hacl* library package
How do I include the Hacl* library in the lib folder when extracting my code to OCaml?
Currently I am stuck with this, which can extract successfully for F* codes only using the F* libraries.
fstar.exe --z3cliopt 'timeout=600000' --use_hints --use_hint_hashes --odir out --codegen OCaml <modulename>.fst
OCAMLPATH="../../fstar/bin" ocamlfind opt -package fstarlib -linkpkg -g out/<modulename>.ml -o test.exe
After I perform the fstar --codegen OCaml --include lib on my file while including the Hacl* library, I used the following Makefile to get the .cmxa hacl library.
FILES= \
FStar_All.ml \
FStar_BitVector.ml \
FStar_Calc.ml \
FStar_Exn.ml \
FStar_Heap.ml \
FStar_List_Tot.ml \
FStar_List_Tot_Base.ml \
FStar_List_Tot_Properties.ml \
FStar_Math_Lemmas.ml \
FStar_Math_Lib.ml \
FStar_Mul.ml \
FStar_Pervasives.ml \
FStar_Pervasives_Native.ml \
FStar_PredicateExtensionality.ml \
FStar_Preorder.ml \
FStar_PropositionalExtensionality.ml \
FStar_Seq.ml \
FStar_Seq_Base.ml \
FStar_Seq_Properties.ml \
FStar_ST.ml \
FStar_StrongExcludedMiddle.ml \
FStar_TSet.ml \
FStar_UInt.ml \
FStar_UInt8.ml \
FStar_UInt16.ml \
FStar_UInt32.ml \
FStar_UInt64.ml
OBJS=$(FILES:.ml=.o)
CMX=$(FILES:.ml=.cmx)
all: hacllib.cmxa
hacllib.cmxa:
OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -thread -w -58 -c $(FILES)
OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -a $(CMX) -o hacllib.cmxa
clean:
rm -f *.cmi *.cmo *.cmx *.cmxa *.o *.a *~
However, there was this error:
Error: The files fStar_BitVector.cmi and fStar_Seq_Base.cmi
make inconsistent assumptions over interface FStar_Pervasives_Native
make: *** [Makefile:36: hacllib.cmxa] Error 2
Can anyone enlighten me on whether I am on the right track and how to fix this?
You appear to have inconsistent state between fstar and hacl. Can you clean
both repos and recompile from scratch?
On Mon, Apr 29, 2019, 10:56 jamesbarne notifications@github.com wrote:
After I perform the fstar --codegen OCaml --include lib on my file while
including the Hacl* library, I used the following Makefile to get the .cmxa
hacl library.
FILES=
FStar_All.ml
FStar_BitVector.ml
FStar_Calc.ml
FStar_Exn.ml
FStar_Heap.ml
FStar_List_Tot.ml
FStar_List_Tot_Base.ml
FStar_List_Tot_Properties.ml
FStar_Math_Lemmas.ml
FStar_Math_Lib.ml
FStar_Mul.ml
FStar_Pervasives.ml
FStar_Pervasives_Native.ml
FStar_PredicateExtensionality.ml
FStar_Preorder.ml
FStar_PropositionalExtensionality.ml
FStar_Seq.ml
FStar_Seq_Base.ml
FStar_Seq_Properties.ml
FStar_ST.ml
FStar_StrongExcludedMiddle.ml
FStar_TSet.ml
FStar_UInt.ml
FStar_UInt8.ml
FStar_UInt16.ml
FStar_UInt32.ml
FStar_UInt64.ml
OBJS=$(FILES:.ml=.o)
CMX=$(FILES:.ml=.cmx)
all: hacllib.cmxa
hacllib.cmxa:
OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -thread -w -58 -c $(FILES)
OCAMLPATH="../../../Fstar/bin" ocamlfind opt -package fstarlib -a $(CMX) -o hacllib.cmxa
clean:
rm -f *.cmi *.cmo *.cmx *.cmxa *.o *.a *~
However, there was this error:
Error: The files fStar_BitVector.cmi and fStar_Seq_Base.cmi
make inconsistent assumptions over interface FStar_Pervasives_Native
make: *** [Makefile:36: hacllib.cmxa] Error 2
Can anyone enlighten me on whether I am on the right track and how to fix
this?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/project-everest/hacl-star/issues/154#issuecomment-487505105,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ABFUVS4C32ULZXFOZK3FGH3PS2Z3LANCNFSM4HIKOBBA
.
I recomplied the fstar and hacl from scratch but am still facing the same problem.
When attempting to recompile kremlin too, the build failed with a similar error regarding the FStar_Pervasives_Native module.
Rebuilding kremlin
Running: build_kremlin
make: Entering directory '/cygdrive/d/lbeverest/everest/kremlin'
ocamlbuild -I src -I lib -I parser -I kremlib -use-menhir -use-ocamlfind -classic-display -menhir "menhir --infer --explain" Kremlin.native Tests.native
ln -sf Kremlin.native krml
make -C kremlib
make[1]: Entering directory '/cygdrive/d/lbeverest/everest/kremlin/kremlib'
D:/lbeverest/everest/FStar/bin/fstar.exe --record_hints --use_hints --use_two_phase_tc true --odir .extract --cache_checked_modules --cmi --already_cached 'FStar -FStar.Kremlin.Endianness LowStar' FStar.Kremlin.Endianness.fst && \
touch FStar.Kremlin.Endianness.fst.checked
D:\lbeverest\everest\FStar\ulib\FStar.Pervasives.fst(0,0-0,0): (Warning 241) Unable to load D:\lbeverest\everest\FStar\bin\..\ulib\FStar.Pervasives.fst.checked since checked file D:\lbeverest\everest\FStar\bin\..\ulib\FStar.Pervasives.fst.checked is corrupt; will recheck D:\lbeverest\everest\FStar\ulib\FStar.Pervasives.fst (suppressing this warning for further modules)
(Error 317) Expected D:\lbeverest\everest\FStar\ulib\FStar.Pervasives.fst to already be checked
1 error was reported (see above)
make[1]: *** [Makefile:58: FStar.Kremlin.Endianness.fst.checked] Error 1
make[1]: Leaving directory '/cygdrive/d/lbeverest/everest/kremlin/kremlib'
make: *** [Makefile:28: kremlib] Error 2
make: Leaving directory '/cygdrive/d/lbeverest/everest/kremlin'
================================================================================
FAILURE: build failed for kremlin
We currently don't build an OCaml library for Hacl* lib like we do for FStar ulib.
Do you have a use case where this would be more convenient than extracting the individual files you depend on?
The first issue you encountered about inconsistent assumptions over interface FStar_Pervasives_Native is due to you re-extracting and re-compiling modules in the FStar namespace and linking them with fstarlib, which includes all of them. The solution is to compile only the modules not already included in fstarlib, which would also save you some time.
Here's a simple example of how to build an OCaml binary from an FStar module that depends on Hacl* lib:
module Test
open FStar.All
open Lib.IntTypes
val test: unit -> ML uint32
let test () =
IO.print_string "Hello world!\n";
u32 0
.PHONY: all test clean
FSTAR_HOME ?= ../../FStar
HACL_HOME ?= ..
FSTAR_FLAGS = $(OTHERFLAGS) \
--cmi \
--cache_checked_modules \
--already_cached "'Prims+FStar+LowStar+Lib'" \
--include $(HACL_HOME)/lib
FSTAR = $(FSTAR_HOME)/bin/fstar.exe $(FSTAR_FLAGS)
ROOTS = Test.fst
all:
rm -f .depend && $(MAKE) .depend
$(MAKE) test
# 1. Generation of .ml files
# - generate the F* dependency graph via `fstar --dep full`
# - verify every F* file in parallel to generate .checked files
# - extract each .checked file into a .ml file in parallel
.depend:
$(FSTAR) --dep full $(ROOTS) --extract '* -Prims -LowStar -FStar' > $@
include .depend
%.checked: | .depend
$(FSTAR) $< && \
touch $@
%.ml: | .depend
$(FSTAR) --codegen OCaml \
--extract_module $(basename $(notdir $(subst .checked,,$<))) \
$(notdir $(subst .checked,,$<)) && \
touch $@
# 2. Compilation
OCAMLOPT=OCAMLPATH="$(FSTAR_HOME)/bin" ocamlfind opt -package fstarlib -linkpkg
%.cmx:
$(OCAMLOPT) -c $< -o $@
Test.cmx:
@echo 'let _ = test()' >> Test.ml
$(OCAMLOPT) -c $< -o $@
test.exe: Test.cmx
$(OCAMLOPT) -o test.exe $(subst .ml,.cmx,$(ALL_ML_FILES))
test: test.exe
./test.exe
clean:
rm -f *.ml *.cmi *.cmo *.cmx *.cmxa *.o *.a *~ *.checked *.exe
This assumes you have already verified and generated .checked files for all modules in lib. It will verify the rest of the dependencies (in this case just Test), extract Test and the modules in lib to OCaml and build a binary from them and fstarlib. All this can be done in parallel using make -j.
This is a sketch of how you would do it building an intermediate hacllib.cmxa:
# 2. Compilation
OCAMLOPT_=OCAMLPATH="$(FSTAR_HOME)/bin" ocamlfind opt -package fstarlib
OCAMLOPT=$(OCAMLOPT_) -linkpkg
%.cmx:
$(OCAMLOPT) -c $< -o $@
Test.cmx:
@echo 'let _ = test()' >> Test.ml
$(OCAMLOPT) -c $< -o $@
LIB_FILES = $(filter Lib_%,$(ALL_ML_FILES))
OTHER_FILES = $(filter-out Lib_%,$(ALL_ML_FILES))
hacllib.cmxa: $(subst .ml,.cmx,$(LIB_FILES))
$(OCAMLOPT_) -a -o $@ Lib_IntTypes.cmx
test.exe: hacllib.cmxa $(subst .ml,.cmx,$(OTHER_FILES))
$(OCAMLOPT) -o test.exe $< $(subst .ml,.cmx,$(OTHER_FILES))
It works for this simple example but it's only a sketch because some modules in lib have dependencies on KreMLin modules that aren't in the Lib namespace and won't be appended to hacllib.cmxa.
We now produce obj/libhaclml.cmxa as part of the build. Assuming you have a successful build of HACL*, then you should be able to include obj/ to get the benefits of the extracted .ml files, and you should be able to link against libhaclml.cmxa to obtain the final executable.
This is how we assemble libhaclml.cmxa:
1069 obj/libhaclml.cmxa: $(filter-out $(HACL_HOME)/obj/Meta_Interface.cmx,$(ALL_CMX_FILES))
1070 # JP: doesn't work because a PPX is prepended for some reason
1071 #ocamlfind mklib -o haclml -package fstarlib -g -I $(HACL_HOME)/obj $(addprefix $(HACL_HOME)/obj/*.,cmo cmx ml o)
1072 ocamlfind opt -a -o $@ -package fstarlib -g -I $(HACL_HOME)/obj $^
To use libhaclml.cmxa as a client, you would need to link against both fstarlib and libhaclml and pass -linkpkg to ocamlfind opt in order to produce a final executable.
Please reopen this if still relevant, thanks.
| gharchive/issue | 2019-04-25T07:30:21 | 2025-04-01T04:34:27.425817 | {
"authors": [
"jamesbarne",
"karthikbhargavan",
"msprotz",
"s-zanella"
],
"repo": "hacl-star/hacl-star",
"url": "https://github.com/hacl-star/hacl-star/issues/154",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2506479112 | grouping test cases and categorisation
Hi @haesleinhuepf
I was going through the preprint and one thought I had was grouping the test cases under categories. We get an overall view of how well LLMs perform but lose the granularity on whether LLMs perform well/worse on certain tasks and not on others..
For example, with our test cases perhaps grouping can be like:
Quantification/Measurements: test cases that measure or count can be under quantificaiton/measurements
statistical analysis: t test, pairwise correlation
Morphological operations: binary close, skeleton, expand labels
etc..
It may give an idea of where we need more or less test cases as well.
I remember you had a preprint on ontologies and standards for bioimage analysis. Perhaps that can be used as a reference.
Cheers
Pradeep
Hey @pr4deepr ,
great idea! This categorization is obviously a subjective task. We could automate this and make it more objective using an LLM; a text-classification LLM. Do you by chance have experience with this?
Cheers,
Robert
No, I do not.
Would you put it through chatgpt for example to get a first pass?
If I copied the readme in the main repo containing description of the current test cases and used the question in Chat-GPT (GPT-4o):
I have some python functions and each of them perform a specific operation in bioimage analysis. Classify them into categories based on their function and where they will fit in the image analysis pipeline.
1. Image Preprocessing
These functions focus on preparing the image by modifying its format, reducing noise, or performing other transformations to enhance analysis.
rgb_to_grey_image_transform
rotate_image_by_90_degrees
crop_quarter_image
transpose_image_axes
remove_noise_edge_preserving
subtract_background_tophat
mask_image
remove_labels_on_edges
remove_small_labels
expand_labels_without_overlap
2. Image Enhancement
Functions that enhance specific features of the image, such as sharpening edges or enhancing contrast.
detect_edges
convolve_images
deconvolve_image
remove_noise_edge_preserving
3. Segmentation
These functions segment images into meaningful regions or objects, usually through intensity-based methods.
apply_otsu_threshold_and_count_postiive_pixels
region_growing_segmentation
workflow_segmentation_counting
workflow_watershed_segmentation_correction_measurement
workflow_segmentation_measurement_summary
workflow_segment_measure_umap
4. Morphological Operations
These functions apply morphological transformations to binary images for cleaning or refining shapes.
binary_closing
binary_skeleton
label_binary_image_and_count_labels
label_sequentially
remove_labels_on_edges
remove_small_labels
5. Quantification and Measurement
Functions that quantify different properties of labeled regions or objects in the image.
convex_hull_measure_area
count_number_of_touching_neighbors
measure_aspect_ratio_of_regions
measure_intensity_of_labels
measure_intensity_over_time
measure_mean_image_intensity
measure_pixel_count_of_labels
measure_properties_of_regions
extract_surface_measure_area
measure_properties_of_regions
count_objects_over_time
count_overlapping_regions
6. Feature Extraction
Functions that extract geometrical or intensity features from the image.
fit_circle
radial_intensity_profile
convex_hull_measure_area
measure_aspect_ratio_of_regions
7. File I/O
Functions that handle loading, saving, or listing files.
list_image_files_in_folder
open_image_read_voxel_size
open_image_return_dimensions
open_nifti_image
open_zarr
8. Statistical Analysis
Functions for performing statistical analysis on datasets related to the images.
bland_altman
t_test
pair_wise_correlation_matrix
mean_std_column
mean_squared_error
9. Pipeline/Workflow Automation
These functions automate a sequence of steps, combining multiple image processing operations.
workflow_batch_process_folder_count_labels
workflow_batch_process_folder_measure_intensity
workflow_segment_measure_umap
workflow_segmentation_measurement_summary
workflow_watershed_segmentation_correction_measurement
10. Miscellaneous
Functions that don't fit directly into other categories but are useful for specific tasks.
return_hello_world
Awesome! I couldn't have done better ☺️ Now what do we do woth this? Shall we preserve it somewhere in the repository?
I'm creating a branch with the new categories..
Plan will be to save this as a yaml file..
There are errors and repetitions above, so need to clean it up as well..
https://github.com/pr4deepr/human-eval-bia/tree/function_categorize
repetitions above
Well I think test-cases can be in multiple categories.
Good point
So, I've done an initial pass.
Interesting findings:
Morphological operations: GPT-4o >claude 3.5
Feature extraction: Claude3.5 >GPT-4o
Statistical analysis: llama3 ==gpt3.5turbo >claude3.5 and GPT-4o
The categorisation of course is important and if not done properly can misrepresent the results..
The function categorisation can be found here: https://github.com/pr4deepr/human-eval-bia/blob/function_categorize/demo/create_function_category_yaml.ipynb which saves is as a yaml file
I can create a separate notebook for the data processing and graphing as its currently here: https://github.com/pr4deepr/human-eval-bia/blob/function_categorize/demo/summarize_by_case.ipynb..
Happy to create a PR, but wasn't sure if it should be to main..
Yes! I certainly need such a figure for talks, because showing the blue table for all test-cases doesn't fit on a slide. It could also be in the paper... Curious what @tischi says about this figure:
What I'm a bit concerned about this the static list of categories in the other notebook. It could be a pain to maintain this mid/long term. Would it be possible to put them in a dataframe, and add some code that warns if a test-case is in no category? Or even better, code that uses gpt4-o to categorize tesr-cases that are in no category and then adding them to the dataframe?
Regarding categorization, why not require some metadata tag to be present with each submitted test case? Maybe it's too late for this but we could add it to existing test cases as the numbers seem manageable. Doing it with an LLM would anyway require manual review of the outcome.
Yet, we don't have any infrastructure for handling meta data of test-cases. I was hoping to fully automate this, so that minimal manual curation is necessary. In an earlier discussion, also categorizing code depending on its complexity was discussed. No matter how we do these things, I'd love to have a semi-automatic solution with minimal code/infrastructure to maintain.
We can use the GPT4-o idea, but is there a way to have a seed or something similar to guarantee relatively similar responses? The categories change everytime I ask... OR we just need to really be specific on question we ask GPT..
The tagging could be left to the author of the test case given a choice of predefined categories. Then its should be a matter of reading the tags when compiling the results. If classification is automated with an LLM, the outcome is likely to change over time and with the LLM used. I think we would need a deterministic algorithm for this.
@pr4deepr Exactly what I thought likely :)
Ok, I leave the decision about this to you guys. Whatever works :-)
I'm happy with solution from @jkh1 , i.e., having a few tags, and getting the author of new test cases to put those tags in their functions. We can have a few different tags for each category.
This could be a requirement when submitting a new test cases.
For existing functions, perhaps myself and @jkh1 could
decide on categories & tags
add them to existing test cases and decide where to define it
submit a PR
Cheers
Pradeep
put those tags in their functions
Can you give an example how this could look like?
Either in functions or in each notebook. I need to look at the code first. Will update it here
Upon looking at the code again, I think we'll want to minimize any modifications to existing test functions & with creating yaml files for cases at this point.
I propose we have all the categorisation information in a yaml file with:
each function name as a key
values are the categories.
The categories can be:
file_i_o
image_preprocessing
morphological_operations
segmentation
feature_extraction
statistical_analysis
workflow_automation
oher
example yaml file:
create_umap
feature_extraction
workflow_segment_measure_umap
segmentation
workflow_automation
feature_extraction
I'm happy to go through existing test cases and create this yaml file..
When a test case PR is submitted, the yaml file will have to be modified to add the new function and category. The PR template will need to be modified.
If the need arises we can expand the categories, but I feel like this should cover it.
Yes, great idea!
The PR template will need to be modified.
We can also add some python code which tests if all test cases are in thus yaml.file. e.g. in create_cases.ipynb or as github worklfow.
Sounds good to me. My intiial idea was to use notebook tags but I realized this may be more complicated to get at.
I've made the necessary changes with commit history here:
https://github.com/pr4deepr/human-eval-bia/commits/function_categorize/
Added a categorise functions yaml file
Added a check in create_test_cases to verify all functions are present
Code updated to plot by category
updated PR template. wording may need changing
If you are happy with this, I can open a PR. Let me know which branch you'd prefer.
Awesome @pr4deepr , thanks for working on this! Yes, please send a PR!
| gharchive/issue | 2024-09-04T23:35:48 | 2025-04-01T04:34:27.488189 | {
"authors": [
"haesleinhuepf",
"jkh1",
"pr4deepr"
],
"repo": "haesleinhuepf/human-eval-bia",
"url": "https://github.com/haesleinhuepf/human-eval-bia/issues/112",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
448852537 | Loading progress in %
Loading progress in % for multiple resources.
Based off PR #6 made by @jiawenquan
This is excellent
It is a singleton class because being an Angular service.
| gharchive/pull-request | 2019-05-27T13:14:59 | 2025-04-01T04:34:27.490249 | {
"authors": [
"haestflod",
"jiawenquan"
],
"repo": "haestflod/generic-3Dproduct-viewer",
"url": "https://github.com/haestflod/generic-3Dproduct-viewer/pull/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
609888334 | Add --grep option
custom grep.
why?
dalfox is xss tool, but it can also be used to find other vulnerabilities.
e.g
ssti
information leak
etc..
so, i think need custom grep option
it is last options for release :D
Thats great bro, you should have that, but in case you dont, hope this helps for that or other project.
patterns.json.txt
And my SSTI playroom =]
<%= 339 * 431 %>
#{ 339 * 431 }
#{set} ($run=339 * 431) $run
${339*431}
<%=339*431%>
${{339*431}}
{{339*431}}
{{339*'3'}}
[[339*431]]
{{= 339*431 }}
{#339*431}
<# 339*431 >
{@339*431}
${{"{{"}}339*431{{"}}"}}
Hi @bsysop
Thanks dude! Originally, I was going to let the user customize the response grep, but I think some pattern is okay to put it in a built-in pattern. I'll use it well! Thank you.
Gooot it, as you wish bro, you could use it in other tool if you think is better!
--grep argument input is the file? regex string..?
and
Log Format
[G] Found via built-in grepping / payload: {339*431} , grep: 146109
1 line: 146109({"isSuccess":false,"errorMsg":"Parameter error! apps is null","error
+> https://blahblha!~~~
[G] Found via custom grepping / payload: 'adf , grep: internal_domain://asdf
1 line: internal_domain://asdf~~({"isSuccess":false,"errorMsg":"Parameter error! apps is null","error
+> https://blahblha!~~~
Thats awesome, i would love to test that and check new approaches!
Function development is complete, only small pattern addition is required. follow issue below
https://github.com/hahwul/dalfox/issues/60
Amazing mate!!!! Great job!
| gharchive/issue | 2020-04-30T12:07:03 | 2025-04-01T04:34:27.500782 | {
"authors": [
"bsysop",
"hahwul"
],
"repo": "hahwul/dalfox",
"url": "https://github.com/hahwul/dalfox/issues/58",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2349312796 | Get master up to date
master up to date for purpose of development
Probably nothing bad will happen...
| gharchive/pull-request | 2024-06-12T17:20:38 | 2025-04-01T04:34:27.501694 | {
"authors": [
"RichardStaszkiewicz"
],
"repo": "haichangsi/WUT-TBD-PROJ1",
"url": "https://github.com/haichangsi/WUT-TBD-PROJ1/pull/5",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
836047249 | internal/graphics: Race condition at the vertices backend and copying at internal/graphicscommand
Git bisect said c8b98f13fb06eee62d6eba2c37d891f86726e40d was the culprit.
CC @wasedaigo
(v *verticesBackend) slice was not concurrent safe. This should not be an issue in v2.0.0 or v1.x since this function was used only on browsers where there is one thread.
OK this is not an easy issue:
graphics.QuadVertex is called from the screen rendering (last=true)
At the same time, graphics.QuadVertex is accessed from a goroutine in parallel
The vertices backend is reset (head is set to 0) even though the vertices is not copied at the graphicscommand package.
Then the vertices data is overwritten and broken.
Found there is a potential race condition even with the restorable package.
I think it is inevitable to use a simple allocation instead of using the backend, even on browsers...
Reverting 9cb631e30f3cf7471d53ed9aa76afb42846b0918 should have the same effect.
I'll backport this change. In v2.0 and v1.12, this issue was only on browsers. Also I'll (re)add mutex to vertexBackend.slice since there is no guarantee about context change in Wasm.
| gharchive/issue | 2021-03-19T14:18:38 | 2025-04-01T04:34:27.536414 | {
"authors": [
"hajimehoshi"
],
"repo": "hajimehoshi/ebiten",
"url": "https://github.com/hajimehoshi/ebiten/issues/1546",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
286475817 | examples/audio: Music should keep playing even the tab is background when IsRunnableInBackground is true
Related to #272
When thinking examples/audio as a music player, it'd be nice to keep playing music
This is no longer an issue. I don't know when this was fixed...
| gharchive/issue | 2018-01-06T10:39:06 | 2025-04-01T04:34:27.537526 | {
"authors": [
"hajimehoshi"
],
"repo": "hajimehoshi/ebiten",
"url": "https://github.com/hajimehoshi/ebiten/issues/476",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2092603148 | 夜间模式下,部分页面的字符颜色显示异常
https://www.bilibili.com/v/popular/weekly/?num=251
這個頁面我不太想適配其實,之後寫個正則過濾掉防止影響
暫時先過濾了這頁面。。。過後適配
| gharchive/issue | 2024-01-21T14:11:56 | 2025-04-01T04:34:27.550650 | {
"authors": [
"hakadao",
"miracomangomanchuria"
],
"repo": "hakadao/BewlyBewly",
"url": "https://github.com/hakadao/BewlyBewly/issues/206",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1893168011 | Hadhunna/adding metrics
Contributing to the Azure SDK
Features
Added metrics to header. Will be used to track the number of requests and responses for each action.
Other Changes
Change version semantics to Major.Minor.Patch.PreRelease
Can we add links to the work items in the description?
Also, please run the following commands from C:\repos\azure-sdk-for-net:
eng\scripts\Update-Snippets.ps1
eng\scripts\CodeChecks.ps1 -ServiceDirectory entra
| gharchive/pull-request | 2023-09-12T19:55:32 | 2025-04-01T04:34:27.561962 | {
"authors": [
"HarmanDhunna",
"hakimms"
],
"repo": "hakimms/azure-sdk-for-net",
"url": "https://github.com/hakimms/azure-sdk-for-net/pull/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
323423046 | Allow for customized sudo env vaiables
Whether or not this is a good idea, I want to do:
halyard::allowed_env_variables:
- DEBUG
- PUPPET_ENV
- PROFILE
- SSH_AUTH_SOCK
This let's me do that
❯ halyard
Repo is unclean: /opt/halyard/repo
2018-05-15 19:50:48 Jareds-MacBook-Pro.local STARTING RUN
2018-05-15 19:50:54 Jareds-MacBook-Pro.local Notice: Compiled catalog for jareds-macbook-pro.local in environment production in 0.55 seconds
2018-05-15 19:50:55 Jareds-MacBook-Pro.local Notice: /Stage[main]/Halyard/Sudoers::Allowed_command[halyard_puppet]/File[/etc/sudoers.d/halyard_puppet]/content: content changed '{md5}8fd6def48b20bcf4bd829ac382b6dc46' to '{md5}7a6862829d4223eb8e6baf25c2f6f289'
2018-05-15 19:50:57 Jareds-MacBook-Pro.local Notice: Applied catalog in 2.74 seconds
2018-05-15 19:50:57 Jareds-MacBook-Pro.local ENDING RUN
❯ sudo cat /etc/sudoers.d/halyard_puppet
# Allows halyard user to run puppet
Cmnd_Alias HALYARD_PUPPET = /opt/halyard/repo/meta/halyard,/bin/sh -c /opt/halyard/repo/meta/halyard
Defaults!HALYARD_PUPPET secure_path = /sbin:/bin:/usr/sbin:/usr/bin
Defaults!HALYARD_PUPPET env_keep+=DEBUG
Defaults!HALYARD_PUPPET env_keep+=PUPPET_ENV
Defaults!HALYARD_PUPPET env_keep+=PROFILE
Defaults!HALYARD_PUPPET env_keep+=SSH_AUTH_SOCK
jaredledvina ALL=(root) NOPASSWD: HALYARD_PUPPET
As for the 'why?', because I have my github SSH key on my yubikey, and my remote for halyard is my fork using ssh :)
So the goal here is that when running puppet-run it'll run with your SSH agent so it can chat w/ github?
@akerl - Yep, passing through SSH_AUTH_SOCK let's haylard, git pull successfully in my configuration.
| gharchive/pull-request | 2018-05-15T23:53:46 | 2025-04-01T04:34:27.594641 | {
"authors": [
"akerl",
"jaredledvina"
],
"repo": "halyard/puppet-halyard",
"url": "https://github.com/halyard/puppet-halyard/pull/6",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.