added
stringdate 2025-04-01 04:05:38
2025-04-01 07:14:06
| created
timestamp[us]date 2001-10-09 16:19:16
2025-01-01 03:51:31
| id
stringlengths 4
10
| metadata
dict | source
stringclasses 2
values | text
stringlengths 0
1.61M
|
|---|---|---|---|---|---|
2025-04-01T04:35:34.882824
| 2019-03-22T17:51:43
|
424330376
|
{
"authors": [
"mightyguava",
"ryanhall07"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11019",
"repo": "square/misk",
"url": "https://github.com/square/misk/pull/862"
}
|
gharchive/pull-request
|
Only start vitess in TESTING
Docker compose starts vitess in DEVELOPMENT now
Pretty sure this will break things. It's similar to the issue in https://github.com/square/misk/pull/814, that VitessScaleSafetyChecks installed by HibernateTestingModule has a dependency on StartVitessService.
Pretty sure this will break things. It's similar to the issue in #814, that VitessScaleSafetyChecks installed by HibernateTestingModule has a dependency on StartVitessService.
@mightyguava I don't think so, since HibernateTestingModule is still installed in tests. This is different from #814 since it's not trying to remove vitess in tests, only in local development
Ah ok. I thought HibernateTestingModule was being installed for local dev too...
|
2025-04-01T04:35:34.921178
| 2015-10-11T18:13:24
|
110873286
|
{
"authors": [
"rburgst",
"swankjesse"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11020",
"repo": "square/okhttp",
"url": "https://github.com/square/okhttp/issues/1919"
}
|
gharchive/issue
|
redirect for PROFIND request issues GET request
see code in com.squareup.okhttp.internal.http.HttpEngine#followUpRequest
case HTTP_MULT_CHOICE:
case HTTP_MOVED_PERM:
case HTTP_MOVED_TEMP:
case HTTP_SEE_OTHER:
// Does the client allow redirects?
if (!client.getFollowRedirects()) return null;
String location = userResponse.header("Location");
if (location == null) return null;
HttpUrl url = userRequest.httpUrl().resolve(location);
// Don't follow redirects to unsupported protocols.
if (url == null) return null;
// If configured, don't follow redirects between SSL and non-SSL.
boolean sameScheme = url.scheme().equals(userRequest.httpUrl().scheme());
if (!sameScheme && !client.getFollowSslRedirects()) return null;
// Redirects don't include a request body.
Request.Builder requestBuilder = userRequest.newBuilder();
if (HttpMethod.permitsRequestBody(userRequest.method())) {
requestBuilder.method("GET", null);
requestBuilder.removeHeader("Transfer-Encoding");
requestBuilder.removeHeader("Content-Length");
requestBuilder.removeHeader("Content-Type");
}
Thus, any http body is lost and out of a PROFIND request I get a GET request instead.
IMHO we need a more flexible solution for coming up with a redirected request. For PROPFIND, the redirected request should again be PROPFIND and the HTTP body should be maintained.
@rburgst does this fix the problem?
https://github.com/square/okhttp/pull/1901
yep, thanks!
seems to be back with okhttp-2.6.0 final
Problem seems to be in HttpEngine.java line 938
case HTTP_MULT_CHOICE:
case HTTP_MOVED_PERM:
case HTTP_MOVED_TEMP:
case HTTP_SEE_OTHER:
// Does the client allow redirects?
if (!client.getFollowRedirects()) return null;
String location = userResponse.header("Location");
if (location == null) return null;
HttpUrl url = userRequest.httpUrl().resolve(location);
// Don't follow redirects to unsupported protocols.
if (url == null) return null;
// If configured, don't follow redirects between SSL and non-SSL.
boolean sameScheme = url.scheme().equals(userRequest.httpUrl().scheme());
if (!sameScheme && !client.getFollowSslRedirects()) return null;
// Redirects don't include a request body.
Request.Builder requestBuilder = userRequest.newBuilder();
if (HttpMethod.permitsRequestBody(userRequest.method())) {
>>> HERE IS THE PROBLEM
requestBuilder.method("GET", null);
requestBuilder.removeHeader("Transfer-Encoding");
requestBuilder.removeHeader("Content-Length");
requestBuilder.removeHeader("Content-Type");
}
|
2025-04-01T04:35:34.932156
| 2020-03-31T20:27:56
|
591402884
|
{
"authors": [
"mopsalarm",
"yschimke"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11021",
"repo": "square/okhttp",
"url": "https://github.com/square/okhttp/issues/5917"
}
|
gharchive/issue
|
Infinite caching in DnsOverHttps
If you are using DnsOverHttps with post=false, url=https://<IP_ADDRESS>/dns-query and a custom OkHttpClient with a Cache set, responses are cached forever, and no validation regarding dns ttl are performed.
Cloudflare does not send a Cache-Control header, and DnsRecordCodec discards the ttl value from the response.
I am seeing cache-control in responses. But it's short e.g. 26 seconds.
21:44:45.939 <-- 200 https://cloudflare-dns.com/dns-query?ct=application/dns-udpwireformat&dns=AAABAAABAAAAAAAAB2h0dHBiaW4Db3JnAAABAAE (170ms)
21:44:45.939 date: Tue, 31 Mar 2020 20:44:45 GMT
21:44:45.939 content-type: application/dns-udpwireformat
21:44:45.939 content-length: 128
21:44:45.939 access-control-allow-origin: *
21:44:45.940 cache-control: max-age=26
21:44:45.940 expect-ct: max-age=604800, report-uri="https://report-uri.cloudflare.com/cdn-cgi/beacon/expect-ct"
21:44:45.940 vary: Accept-Encoding
21:44:45.940 server: cloudflare
21:44:45.940 cf-ray: 57cce902fac2e624-LHR
FWIW the behaviour implemented is deliberately inline with the spec
As a result, DoH servers need to carefully consider the HTTP caching
metadata they send in response to GET requests (responses to POST
requests are not cacheable unless specific response header fields are
sent; this is not widely implemented and is not advised for DoH).
In particular, DoH servers SHOULD assign an explicit HTTP freshness
lifetime (see Section 4.2 of [RFC7234]) so that the DoH client is
more likely to use fresh DNS data. This requirement is due to HTTP
caches being able to assign their own heuristic freshness (such as
that described in Section 4.2.2 of [RFC7234]), which would take
control of the cache contents out of the hands of the DoH server.
Hoffman & McManus Standards Track [Page 8]
RFC 8484 DNS Queries over HTTPS (DoH) October 2018
The assigned freshness lifetime of a DoH HTTP response MUST be less
than or equal to the smallest TTL in the Answer section of the DNS
response. A freshness lifetime equal to the smallest TTL in the
Answer section is RECOMMENDED. For example, if a HTTP response
carries three RRsets with TTLs of 30, 600, and 300, the HTTP
freshness lifetime should be 30 seconds (which could be specified as
"Cache-Control: max-age=30"). This requirement helps prevent expired
RRsets in messages in an HTTP cache from unintentionally being
served.
While the Spec allows for overriding cache in requests, the use of a configured Cache is deliberate to avoid hammering the server, and correctly implement caching based on TTLs.
Can you provide more detail to reproduce the issue, everything I can see seems to be working correctly in my test client.
Happy to continue the investigation, but I can't see what you are seeing.
FWIW Caching appears to work also
$ ./okurl --debug --dns=DNSOVERHTTPS --cache /tmp/a.cache https://api.twitter.com/robots.txt
22:15:02.897 -Dorg.conscrypt.native.workdir: /var/folders/gj/tk6xp0gx7kx_0b41p6vlyg6r00pfj6/T
22:15:03.117 url https://api.twitter.com/robots.txt
22:15:03.118 Request Request{method=GET, url=https://api.twitter.com/robots.txt, headers=[User-Agent:okurl/dev], tags={class com.baulsupp.okurl.credentials.Token=TokenSet(name=default)}}
22:15:03.178 Dns (api.twitter.com): /<IP_ADDRESS>, /<IP_ADDRESS>, /<IP_ADDRESS>, /<IP_ADDRESS>
22:15:03.453 Q10004 scheduled after 5 s : OkHttp api.twitter.com ping
22:15:03.458 >> CONNECTION 505249202a20485454502f322e300d0a0d0a534d0d0a0d0a
22:15:03.459 >> 0x00000000 6 SETTINGS
22:15:03.460 >> 0x00000000 4 WINDOW_UPDATE
Just to be clear here, I am talking about caching of the inner response from <IP_ADDRESS> for dns lookup, not caching of the outer response from twitter.com.
I think it is a flaw in some cute logic to try to force using the cache if possible, over making an additional request to the server.
https://github.com/square/okhttp/blob/ff0dee1c34fc06084357c93dc480f6ad89fef24a/okhttp-dnsoverhttps/src/main/java/okhttp3/dnsoverhttps/DnsOverHttps.kt#L108
I think it is a flaw in some cute logic to try to force using the cache if possible, over making an additional request to the server.
https://github.com/square/okhttp/blob/ff0dee1c34fc06084357c93dc480f6ad89fef24a/okhttp-dnsoverhttps/src/main/java/okhttp3/dnsoverhttps/DnsOverHttps.kt#L108
Yes, thats pretty much it. OkHttp also runs into this line: https://github.com/square/okhttp/blob/master/okhttp/src/main/java/okhttp3/internal/cache/CacheStrategy.kt#L190
I'm still not convinced it's a bug
https://github.com/square/okhttp/blob/ff0dee1c34fc06084357c93dc480f6ad89fef24a/okhttp/src/main/java/okhttp3/CacheControl.kt#L228-L234
The code is checking for 504, and uses maxAge, not maxStale
https://github.com/square/okhttp/blob/ff0dee1c34fc06084357c93dc480f6ad89fef24a/okhttp-dnsoverhttps/src/main/java/okhttp3/dnsoverhttps/DnsOverHttps.kt#L185-L202
|
2025-04-01T04:35:34.957201
| 2016-04-12T02:36:24
|
147615104
|
{
"authors": [
"JakeWharton",
"iNoles"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11022",
"repo": "square/sqlbrite",
"url": "https://github.com/square/sqlbrite/issues/105"
}
|
gharchive/issue
|
QueryToListOperator didn't called onCompleted
Is there is a reason that you don't want to called the onCompleted methods?
SqlBrite streams are infinite. They never complete.
|
2025-04-01T04:35:34.959310
| 2015-07-14T18:02:53
|
95002897
|
{
"authors": [
"squarefeet",
"webtech1980"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11023",
"repo": "squarefeet/ShaderParticleEngine",
"url": "https://github.com/squarefeet/ShaderParticleEngine/issues/73"
}
|
gharchive/issue
|
Issue when camera is positioned at { x:0, y:0, z:0 }
Hello and thanks for the great particle engine. For some reason when I position my camera at { x:0, y:0, z:0 } the particles are not rendered. This only seems to happen if I'm using the latest version of the particle engine. I have tested with THREE.js r70 and r71.
What position are you emitters set to?
It varies. After updating to the new release my particles weren't being rendered. So I stated playing around with the examples you provided. Your basic example has the camera set to camera.position.z = 50 and the emitter at { 0, 0, 0 }. That works fine but if you switch it around and put the emitter at { 0, 0, 50 } and the camera at { 0, 0, 0 } that's when i'm seeing the issue.
Ah, found the issue, thanks for letting me know about it - I doubt I would have come across it myself! I've just pushed v0.8.2 that should fix it for you.
|
2025-04-01T04:35:34.969158
| 2017-10-27T22:45:38
|
269256806
|
{
"authors": [
"squidboylan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11024",
"repo": "squidboylan/libdolphin",
"url": "https://github.com/squidboylan/libdolphin/issues/1"
}
|
gharchive/issue
|
Add hitbox data support
Plans for this are:
Have a separate hitbox class
Each Player will get 4 hitboxes as specified in the google drive doc
Do testing to ensure memory locations are correct
This is done
|
2025-04-01T04:35:35.012648
| 2019-07-03T16:23:03
|
463844214
|
{
"authors": [
"ajnavarro",
"erizocosmico",
"kuba--",
"ncordon",
"smacker",
"smola"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11025",
"repo": "src-d/gitbase",
"url": "https://github.com/src-d/gitbase/issues/918"
}
|
gharchive/issue
|
Top 10 repositories by commit count from HEAD example
I think there may be something wrong with the example: Top 10 repositories by commit count from HEAD in the docs.
For example, filtering for the repo [naming-suggestion](https://github.com/src-d/naming-suggestion= it would show count 4, when it ought to be 2:
SELECT repository_id, num_files FROM (
SELECT COUNT(f.*) num_files, f.repository_id
FROM ref_commits r
NATURAL JOIN commit_files cf
NATURAL JOIN files f
WHERE r.ref_name = 'HEAD'
AND repository_id = 'github.com/src-d/naming-suggestion'
) AS T
I do not know if it is an error or not, but both commits the aforementioned repo has are marked to have ref_name = HEAD while only the latest one should (?):
SELECT *
FROM ref_commits r
NATURAL JOIN commit_files cf
NATURAL JOIN files f
WHERE r.ref_name = 'HEAD'
AND repository_id = 'github.com/src-d/naming-suggestion'
If we replace ref_commits with the table refs results show correctly (HT @smacker )
If you take a look into:
SELECT * FROM refs;
you will see it points to the latest commit per reference, but:
SELECT * FROM ref_commits;
It contains also historical data that's why it also has history_index.
So both commits (initial and the latest one) are on a heads/master branch.
commit 7475743dac0e03d5bd17be4eca299d2989dc1143 (HEAD -> master, origin/master, origin/HEAD)
Date: Thu Oct 5 19:03:16 2017 +0200
update README
README.md | 33 ++++++++++++++++++++++++++++++++-
1 file changed, 32 insertions(+), 1 deletion(-)
commit 025a2ab434160065a0b0b2506dbbc9e23c0f79ac
Date: Thu Oct 5 18:57:18 2017 +0200
Initial commit
.gitignore | 101 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
README.md | 1 +
2 files changed, 102 insertions(+)
These two files (README and gitignore) were added on initial commit, so both commits contain them. In other words if you query refs for each commit, you'll always get somehow duplicated, because file lives in a commit till it will be deleted.
I think the main problem is that the example returns an unexpected result.
@smacker @ncordon First query is 10 top repos by file count at HEAD not Top 10 repositories by commit count from HEAD
the result is correct, the number of files, not the number of commits.
this query:
SELECT repository_id,commit_count
FROM (
SELECT
repository_id,
COUNT(*) AS commit_count
FROM ref_commits
WHERE ref_name = 'HEAD'
GROUP BY repository_id
) AS q
ORDER BY commit_count DESC
LIMIT 10;
Is giving the number of commits, and it is working properly.
But yes, you're right, we should change ref_commits to refs
@smacker @ncordon First query is 10 top repos by file count at HEAD not Top 10 repositories by commit count from HEAD
the result is correct, the number of files, not the number of commits.
this query:
SELECT repository_id,commit_count
FROM (
SELECT
repository_id,
COUNT(*) AS commit_count
FROM ref_commits
WHERE ref_name = 'HEAD'
GROUP BY repository_id
) AS q
ORDER BY commit_count DESC
LIMIT 10;
Is giving the number of commits, and it is working properly.
Yeah, sorry, my bad, I meant the file_count one. I will change the title and description in my comment
@ajnavarro should we migrate this issue to sourced-ce?
@erizocosmico right, I thought that was gitbase documentation.
BTW @carlosms @smacker ; should we consolidate documentation and have all the examples on the same place? that can cause a lot of doc inconsistencies in the near future.
wah. I also thought that was gitbase documentation. π sorry
The query can be improved a little bit more:
Changing ref_commits to refs for correct results.
Removing join with files, it just makes the query slower.
Removing superfluous aliases.
The following one should do the same, and works both with gitbase and Spark SQL:
SELECT repository_id, num_files FROM (
SELECT COUNT(*) num_files, repository_id
FROM refs
NATURAL JOIN commit_files
WHERE ref_name = 'HEAD'
GROUP BY repository_id
) AS t
ORDER BY num_files DESC
LIMIT 10;
|
2025-04-01T04:35:35.014812
| 2017-02-14T17:22:58
|
207581251
|
{
"authors": [
"dpordomingo",
"erizocosmico"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11026",
"repo": "src-d/go-kallax",
"url": "https://github.com/src-d/go-kallax/pull/97"
}
|
gharchive/pull-request
|
Find error in structs with AliasOfSlice properties
Described by https://github.com/src-d/go-kallax/issues/96
Leave this PR open but don't merge it because tests will start failing
I'm merging this on my branch, since I will be working on this and will need the tests.
|
2025-04-01T04:35:35.017254
| 2018-10-16T13:14:51
|
370606316
|
{
"authors": [
"erizocosmico",
"jfontan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11027",
"repo": "src-d/go-siva",
"url": "https://github.com/src-d/go-siva/pull/38"
}
|
gharchive/pull-request
|
Improve siva index generation in ReadWriter
When using a ReadWriter the index in regenerated each time Index is called. These are the steps to generate one usable index:
Create new index merging index from file and current changes
Remove duplicates
Sort index by position
Also, to find a file in the index it is walked until a match is found. This needs to be done each time a file has to be opened.
For small number of files this is OK but when a repo has a lot or references the time spent here can be a lot.
Now there's a new index type called OrderedIndex that stores the IndexEntries in lexicographic order. This allows to do binary searches for faster file location and also makes possible update the index
instead of regenerating it each time.
Do you have some number about how this might affect speed on go-git?
|
2025-04-01T04:35:35.038435
| 2020-08-25T18:19:16
|
685692819
|
{
"authors": [
"cmatomic",
"srevinsaju"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11028",
"repo": "srevinsaju/firefox-appimage",
"url": "https://github.com/srevinsaju/firefox-appimage/issues/2"
}
|
gharchive/issue
|
Firefox-AppImage stands out .
I changed the name of my repository so you can use the name Firefox-AppImage, your repository is better than mine so I recommend changing to Firefox-AppImage.
Thanks for your excellent job you had with the repository .
@cmatomic Wow Thanks! How did you come to know of this repository?
@srevinsaju
I was trying to find a solution to make an appimage of one of my repositories updateable related to "gh-releases-zsync"
Renamed, and updated the updateinformation to point to new repository. Thanks :tada:
|
2025-04-01T04:35:35.107094
| 2024-10-01T13:08:20
|
2559179944
|
{
"authors": [
"Devanshukoli",
"sristy17"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11029",
"repo": "sristy17/MindDrive",
"url": "https://github.com/sristy17/MindDrive/issues/14"
}
|
gharchive/issue
|
[FEATURE] Add commitizen in the Project
Is this a unique feature?
[X] I have checked "open" AND "closed" issues and this is not a duplicate
Is your feature request related to a problem/unavailable functionality? Please describe.
Yes, My Feature request is related to an unavailable functionality in the project. And I would like to add it.
Proposed Solution
So basically Commitizen is GitHub cli tool that helps when commit. And It's very easy to use by just typing git cz in terminal and then can accessible.
Here are it's official doc
Screenshots
Demo Link
Do you want to work on this issue?
Yes
If "yes" to above, please explain how you would technically implement this
I have add a video demo link on how this works. And I think it is very easy to use and beginner friendly as well. Which is daily developer routine as well.
Sounds cool! Can give a try. @Devanshukoli
|
2025-04-01T04:35:35.112126
| 2023-11-30T08:26:06
|
2018124246
|
{
"authors": [
"hellt"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11030",
"repo": "srl-labs/containerlab",
"url": "https://github.com/srl-labs/containerlab/issues/1758"
}
|
gharchive/issue
|
Parallelize smoke tests
Smoke test collection grew out of proportion and takes quite some time to finish.
It is better to start compartmentalize the cicd.yml file by moving out jobs to their own files and parallelize the execution.
This was done for vxlan tests -- https://github.com/srl-labs/containerlab/blob/46bd91f34b3d60e701c64ee02f078bf6811beb51/.github/workflows/cicd.yml#L412
Once moved to its own file the tests can be further parallelized by implementing a matrix strategy and calling each test suite in parallel -- https://github.com/srl-labs/containerlab/blob/46bd91f34b3d60e701c64ee02f078bf6811beb51/.github/workflows/vxlan-tests.yml#L9-L16
done in #1800
|
2025-04-01T04:35:35.154203
| 2018-10-07T08:36:07
|
367533405
|
{
"authors": [
"codecov-io",
"izumin5210"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11032",
"repo": "srvc/wraperr",
"url": "https://github.com/srvc/wraperr/pull/4"
}
|
gharchive/pull-request
|
Fix reviewdog token
WHY
WHAT
REF
Codecov Report
Merging #4 into master will not change coverage.
The diff coverage is n/a.
@@ Coverage Diff @@
## master #4 +/- ##
=======================================
Coverage 76.56% 76.56%
=======================================
Files 4 4
Lines 192 192
=======================================
Hits 147 147
Misses 38 38
Partials 7 7
Continue to review full report at Codecov.
Legend - Click here to learn more
Ξ = absolute <relative> (impact), ΓΈ = not affected, ? = missing data
Powered by Codecov. Last update e442151...56c8a7a. Read the comment docs.
|
2025-04-01T04:35:35.158403
| 2021-05-08T18:34:29
|
881155923
|
{
"authors": [
"arj03",
"staltz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11033",
"repo": "ssb-js/muxrpc",
"url": "https://github.com/ssb-js/muxrpc/pull/70"
}
|
gharchive/pull-request
|
use standard for linting and code style
Similar to https://github.com/ssb-js/secret-stack/pull/53, this just introduces standardx which applies a conventional code style on all JS files, and warns against weird code, like multiple statements on the same line (muxrpc had lots of this). I didn't do much else than npm run lint and fix linting warnings, just to keep this PR simple to review. I have other code quality PRs coming up for this module.
I recommend taking a look at package.json changes first, and then changes to the *.js files.
Thanks for taking the time to clean up this code.
|
2025-04-01T04:35:35.172285
| 2017-02-04T00:35:36
|
205306531
|
{
"authors": [
"sschmid"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11034",
"repo": "sschmid/Entitas-CSharp",
"url": "https://github.com/sschmid/Entitas-CSharp/issues/271"
}
|
gharchive/issue
|
Integrate EntityLink into Entitas
Ship Entitas.Unity.VisualDebugging plugin including EntityLink for linking entities to GameObjects.
See EntityLink https://github.com/sschmid/Entitas-CSharp/wiki/Tools-and-Extensions
Done.
|
2025-04-01T04:35:35.207655
| 2024-08-18T08:00:53
|
2471846474
|
{
"authors": [
"g0odman",
"thdxr"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11035",
"repo": "sst/ion",
"url": "https://github.com/sst/ion/issues/880"
}
|
gharchive/issue
|
Install script redownloads SST on every execution
Since this commit the install script doesn't cache the version correctly since the output of the sst version command has changed.
fixed in v3.0.64
|
2025-04-01T04:35:35.214410
| 2024-04-17T08:11:01
|
2247664873
|
{
"authors": [
"WangTianYi0102",
"sstary"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11036",
"repo": "sstary/SSRS",
"url": "https://github.com/sstary/SSRS/issues/17"
}
|
gharchive/issue
|
How to obtain the pretrain model βvmamba_tiny_e292.pthβ
Very great workοΌ Can you tell me the detailed method to obtain this pre-trained file? I am not understand how to get
it from your linkγ Thank youοΌ
You can get from 'https://github.com/zifuwan/Sigma' and rename.
|
2025-04-01T04:35:35.215962
| 2014-05-16T14:24:27
|
33678872
|
{
"authors": [
"jwestbrook"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11037",
"repo": "sstephenson/prototype",
"url": "https://github.com/sstephenson/prototype/issues/210"
}
|
gharchive/issue
|
SCRIPT5007: Unable to get value of the property '1' fix
Pull Request
https://github.com/sstephenson/prototype/pull/108
closed by de0beadd5347727f06fae4c373c8f99679a9e4ea
|
2025-04-01T04:35:35.248086
| 2024-01-20T18:16:28
|
2092190437
|
{
"authors": [
"EmilMassey",
"staabm"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11038",
"repo": "staabm/phpstan-todo-by",
"url": "https://github.com/staabm/phpstan-todo-by/issues/82"
}
|
gharchive/issue
|
DX: Define error identifiers
In phpstan 1.11 error identifiers will be available for use to ignore specific errors. We can already use PHPStan\Rules\RuleErrorBuilder::identifier() to define custom identifiers for errors reported by this extension. They will be useful when 1.11 is released.
I think it would be a nice feature to have. If we want to have it, we should discuss what the identifiers should look like, eg. todo_by.date, todo_by.ticket, etc.
I think we can start with one error per rule. phpstan seems to use camel case for them.
implemented in https://github.com/staabm/phpstan-todo-by/pull/83
|
2025-04-01T04:35:35.249969
| 2021-06-23T00:49:50
|
927753655
|
{
"authors": [
"kylebarron",
"lossyrob"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11039",
"repo": "stac-utils/pgstac",
"url": "https://github.com/stac-utils/pgstac/issues/6"
}
|
gharchive/issue
|
CI for publishing pypgstac
I noticed that the latest version pypgstac on pypi is 0.2.4. The main branch has a version of 0.2.3, and there's not tags for all released versions.
We should set up GitHub Actions to do the publishing, much like we are doing in pystac. That way releases happen on each GitHub release, and the python package is published by CI against tagged versions. This will require the stacutils user to be made a maintainer on the pypgstac pypi repository.
:+1: Was just going to make an issue that there aren't Git tags for every issue. Would be nice to have them in sync with what gets pushed to pypi.
|
2025-04-01T04:35:35.258517
| 2019-02-22T02:39:52
|
413216487
|
{
"authors": [
"gabrielebndn",
"jcarpent"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11040",
"repo": "stack-of-tasks/pinocchio",
"url": "https://github.com/stack-of-tasks/pinocchio/pull/698"
}
|
gharchive/pull-request
|
Align Python Jacobian to C++, fix #545
Mostly four things:
Add a benchmark called timings-jacobian to compare between getJointJacobian, computeJointJacobians and related algorithms.
Use the corresponding C++ function for the Python bindings of jointJacobian. Now, the signature and side-effects of jointJacobian in Python are completely aligned to C++. The old signature is now deprecated. This fixes #545
Implemented missing frameJacobian function in C++, and use the newly-created function for the Python bindings of frameJacobian. The old signature is now deprecated. This also fixes #545
Change the C++ return policy getFrameVelocity and getFrameAcceleration. Now, the value is returned instead of being written in a reference
By the way, when you commit, can you add before the commit message which topic you are then considering (like algo, spatial, ...). It is much more readable and this is also the main way of committing on Pinocchio.
By the way, when you commit, can you add before the commit message which topic you are then considering (like algo, spatial, ...). It is much more readable and this is also the main way of committing on Pinocchio.
You are totally right, I did it for the first 3 commits but then I forgot for the rest. I can easily edit the commit messages
@gabrielebndn It will be fine for now concerning the commit messages. I just need you to remove the explicit const_cast.
@jcarpent I have really no means to work on it right now (traveling). I will be home this late afternoon, I'll try to fix it then
@gabrielebndn No problem. It was just to summarize the remaining tasks to perform before merging.
@gabrielebndn It would be nice if we can add this feature to the new upcoming release of Pinocchio. Will you have time to deal with it today?
I am working on it. I am not on my own computer, so I need to install pretty much everything
I have removed the explicit const_cast and I have fixed the commit messages. Also, if you want to remove the Jacobian benchmark, you can just discard the last commit
@gabrielebndn Thank you very much for this contribution.
|
2025-04-01T04:35:35.260545
| 2021-11-03T08:25:32
|
1043200318
|
{
"authors": [
"hrp2-14",
"nim65s"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11041",
"repo": "stack-of-tasks/sot-tools",
"url": "https://github.com/stack-of-tasks/sot-tools/pull/26"
}
|
gharchive/pull-request
|
fix submodule url
ref. https://github.com/humanoid-path-planner/hpp-core/issues/256.
If someone prefer using another url scheme, git can globally be configured with eg.:
git config --global url."git@github.com:".insteadOf https://github.com/
Hi ! This project doesn't usually accept pull requests on master. If this wasn't intentionnal, you
can change the base branch of this pull request to devel (No need to close it for that). Best, a bot.
|
2025-04-01T04:35:35.264528
| 2024-11-13T00:50:09
|
2653770844
|
{
"authors": [
"endocytosis",
"robert-hopla"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11042",
"repo": "stackblitz/bolt.new",
"url": "https://github.com/stackblitz/bolt.new/issues/2148"
}
|
gharchive/issue
|
Use the prediction outputs feature to only make changes, not rewrite ALL code. (we are only interested in the diff)
Is your feature request related to a problem? Please describe:
updates take longer than needed
Describe the solution you'd like:
using the prediction outputs, we can supply code and speedup repsonses
Describe alternatives you've considered:
i have tried waiting, but it doesn't work for me
Additional context:
can we connect bolt with hosting platforms to push to dev/live? for example cloudflare API
Hi @robert-hopla,
Appreciate the feedback! We are previewing diff-based file editing. Users with subscriptions can try here: bolt.fyi/diff-mode. Note: This preview URL connects to a staging database so your production projects will not be present there. Use this link for experimental testing purposes only.
We would appreciate any feedback you have while testing this feature!
Instructions:
Log in to Bolt at the above link.
Mouseover the left-hand side of the page to bring up the options menu. Click Settings.
Click Feature Previews and then toggle Diffs to the active position.
Click Save and Reload.
Thanks again for your support as we continue to improve Bolt!
|
2025-04-01T04:35:35.268663
| 2024-11-19T08:15:14
|
2671222378
|
{
"authors": [
"everspader",
"kc0tlh"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11043",
"repo": "stackblitz/bolt.new",
"url": "https://github.com/stackblitz/bolt.new/issues/2733"
}
|
gharchive/issue
|
npm ERR! code ERR_INVALID_PROTOCOL on Safari browser only
Describe the bug
When running npm run dev, there is an error with the dev server that is only observer on Safari browser:
npm ERR! code ERR_INVALID_PROTOCOL
npm ERR! Protocol "https:" not supported. Expected "http:"
npm ERR! A complete log of this run can be found in: /home/.npm/_logs/2024-11-19T07_40_58_889Z-debug-0.log
Link to the Bolt URL that caused the error
N/A
Steps to reproduce
Using Safari:
Start a NextJS project
Try to start the dev server with nom run dev
Expected behavior
The development server should start after installing the package dependencies.
Screen Recording / Screenshot
No response
Platform
OS: macOS
Browser: Safari
Version: 18.2
Additional context
This issue has only been observer with the Safari browser.
@everspader Bolt.new does not currently support Safari. We are adding a banner to indicate this soon, but in the meantime, please switch to Google Chrome or another Chromium based browser.
Tagging #176
@kc0tlh thank you, I appreciate the explanation!
|
2025-04-01T04:35:35.272644
| 2024-11-27T15:37:49
|
2698961650
|
{
"authors": [
"endocytosis",
"mmonty23"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11044",
"repo": "stackblitz/bolt.new",
"url": "https://github.com/stackblitz/bolt.new/issues/3279"
}
|
gharchive/issue
|
Upgraded and lost my project
Describe the bug
I was working on a postcard project on the free plan and when I upgraded to a paid plan, I can't find it.
Link to the Bolt URL that caused the error
https://bolt.new/~/sb1-kvrisw
Steps to reproduce
No project found
Expected behavior
When I upgraded my plan, my expectation was that I would be brought back to my project I was in the middle of.
Screen Recording / Screenshot
No response
Platform
Browser name = Chrome
Full version = <IP_ADDRESS>
Major version = 130
navigator.appName = Netscape
navigator.userAgent = Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/<IP_ADDRESS> Safari/537.36
performance.memory = {
"totalJSHeapSize": 40112201,
"usedJSHeapSize": 37109837,
"jsHeapSizeLimit":<PHONE_NUMBER>
}
Username = mmonty23
Chat ID = 60f7752aa42f
Additional context
No response
Sorry to hear this! Are you able to find your project by going to stackblitz.com, logging in (same credentials as Bolt), clicking collections at the left-hand side of the page and the opening Bolt Collection?
If the issue remains, we are aware of this and are tracking it here.
|
2025-04-01T04:35:35.277474
| 2023-12-31T17:30:40
|
2061126272
|
{
"authors": [
"SamVerschueren",
"djdev"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11045",
"repo": "stackblitz/webcontainer-core",
"url": "https://github.com/stackblitz/webcontainer-core/issues/1283"
}
|
gharchive/issue
|
Slow starting project. Dependencies installation taking forever.
Describe the bug
Github project does not open in browser because dependencies installation does not complete.
Link to the blitz that caused the error
https://stackblitz.com/~/github.com/djdev/dj-badge-component-starter
Steps to reproduce
Start selected Github project.
Installing dependencies takes forever. Not sure if it is by design or an actual bug
Expected behavior
Dependencies should be installed quickly based on package.json file and index page should load in the embedded browser.
Parity with Local
[X] I have run the project in my local machine and I could not reproduce the issue.
Screenshots
Platform
Version = 1.84.0
Hash = a5b1ce2088a8c48a3bfda4a131ef20e00b36d9eb
WebContainer = a2aabdd9
Browser name = Chrome
Full version = <IP_ADDRESS>
Major version = 120
navigator.appName = Netscape
navigator.userAgent = Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/<IP_ADDRESS> Safari/537.36
performance.memory = {
"totalJSHeapSize": 65878110,
"usedJSHeapSize": 56409562,
"jsHeapSizeLimit":<PHONE_NUMBER>
}
Additional context
No response
Hey @djdev π !
I looked into it and it appears that react-scripts is quite big and has quite some dependencies themselves. If I install this on my local machine with a clean cache (as I never installed these packages before), it takes around 47 seconds.
If you want fast installs, it's better to commit your package-lock.json to that repository. That will drastically speed up the install process on StackBlitz. The install with a package-lock.json took 8s for me.
|
2025-04-01T04:35:35.301728
| 2024-04-19T08:28:38
|
2252403831
|
{
"authors": [
"coveralls",
"jhrozek"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11046",
"repo": "stacklok/minder",
"url": "https://github.com/stacklok/minder/pull/3137"
}
|
gharchive/pull-request
|
Address review comments for labels filtering
Summary
In the interest of time we merged labels filtering with some comments
not being addressed. This patch fixes that:
exclude_labels now excludes any profiles containing any of the exclude
labels. Add a test for this new behaviour.
fix confusing comment in LabelsFromFilter
special-case wildcard matching to only return the wildcard
sort the slice in test to avoid issues where the slices contain the
right elements, but in wrong order
Change Type
Mark the type of change your PR introduces:
[ ] Bug fix (resolves an issue without affecting existing features)
[ ] Feature (adds new functionality without breaking changes)
[ ] Breaking change (may impact existing functionalities or require documentation updates)
[ ] Documentation (updates or additions to documentation)
[x] Refactoring or test improvements (no bug fixes or new functionality)
Testing
make test
Review Checklist:
[x] Reviewed my own code for quality and clarity.
[x] Added comments to complex or tricky code sections.
[ ] Updated any affected documentation.
[x] Included tests that validate the fix or feature.
[ ] Checked that related changes are merged.
coverage: 48.143%. remained the same
when pulling 45f9d605d9bc4797cd14115b22ce40868b5286e1 on jhrozek:labels_comments
into 9a1a4ab820b2dcdb77fd0e209cadf4374468c155 on stacklok:main.
|
2025-04-01T04:35:35.306876
| 2024-07-17T11:07:51
|
2413316433
|
{
"authors": [
"blkt",
"coveralls"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11047",
"repo": "stacklok/minder",
"url": "https://github.com/stacklok/minder/pull/3917"
}
|
gharchive/pull-request
|
Removed option from minder history list.
Summary
Option --rule-name was meant to be used as filter on rule names, but we decided not to implement it for now as its usefulness is still under discussion.
Change Type
Mark the type of change your PR introduces:
[ ] Bug fix (resolves an issue without affecting existing features)
[ ] Feature (adds new functionality without breaking changes)
[ ] Breaking change (may impact existing functionalities or require documentation updates)
[ ] Documentation (updates or additions to documentation)
[X] Refactoring or test improvements (no bug fixes or new functionality)
Testing
The option was unused.
Review Checklist:
[X] Reviewed my own code for quality and clarity.
[ ] Added comments to complex or tricky code sections.
[ ] Updated any affected documentation.
[ ] Included tests that validate the fix or feature.
[X] Checked that related changes are merged.
coverage: 54.311%. remained the same
when pulling 3b9a8b1addd0411eed5396916f4d1bd1fdec015e on fix/remove-unused-history-list-option
into fb8bd92f86e89f02416ed66a263e92b3b2f56c41 on main.
|
2025-04-01T04:35:35.312501
| 2024-10-10T14:39:06
|
2579065254
|
{
"authors": [
"blkt",
"coveralls"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11048",
"repo": "stacklok/minder",
"url": "https://github.com/stacklok/minder/pull/4721"
}
|
gharchive/pull-request
|
Fix race in github webhook handler tests.
Summary
Recent changes introduced a couple additional cases in which we forward events from the webhook handler to Minder core, causing helper utility PassthroughQueue to get stuck sending on a channel that was not consumed anymore.
This change should fix the issue by
consuming the new messages (and running assertions over them)
ensuring no additional messages are there to be consumed
Additionally, in case tests fail to consume all produced messages, the test now panics if ran with -race.
Change Type
[X] Bug fix (resolves an issue without affecting existing features)
[ ] Feature (adds new functionality without breaking changes)
[ ] Breaking change (may impact existing functionalities or require documentation updates)
[ ] Documentation (updates or additions to documentation)
[ ] Refactoring or test improvements (no bug fixes or new functionality)
Testing
Unit tests.
Review Checklist:
[X] Reviewed my own code for quality and clarity.
[ ] Added comments to complex or tricky code sections.
[ ] Updated any affected documentation.
[ ] Included tests that validate the fix or feature.
[X] Checked that related changes are merged.
coverage: 53.355% (+0.02%) from 53.34%
when pulling 2e875851a7f17d658b323eb29a19678f5d41ef6a on fix/fix-github-webhook-handler-tests-race
into a4d5c7fd7599dbee64ed4e193c04fe5c9de1bf95 on main.
|
2025-04-01T04:35:35.318929
| 2023-11-02T22:42:38
|
1975152864
|
{
"authors": [
"RTann"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11049",
"repo": "stackrox/scanner",
"url": "https://github.com/stackrox/scanner/pull/1318"
}
|
gharchive/pull-request
|
chore(nvd): use NVD API v2
NVD plans to retire the 1.1 JSON feeds on December 15, 2023. This PR swaps our usage of those feeds with the v2 API.
For backwards compatibility purposes with previous Scanner versions, which may still be deployed somewhere and may still be officially supported, we convert the v2 API JSON responses to the 1.1 JSON formatting. This allows us to continue serving NVD data to live Scanners without any downtime nor required updates.
Before, there was a JSON file per year (2002 - Present year). Now, there will be a JSON file for every 20,000 vulnerabilities. The maximum file size, uncompressed, is about 60MB, and the genesis-dump.zip is about 164MB. The genesis-dump.zip associated with the 4.3.0 release is about 166MB, so these changes result in negligible differences in the genesis-dump.
The change in file names will not affect older Scanners (at least no Scanners which are still supported), as each Scanner looks inside of a directory called nvd/ (which has not changed) for all .json files. As long as the file extension is .json, the file name does not matter. See https://github.com/stackrox/scanner/blob/2.28.7/cpe/nvdtoolscache/load.go, which (as of writing this), has not been touched since well before 3.74 (Scanner 2.28) was released.
See https://github.com/stackrox/nvdtools/pull/9 for the related changes to the nvdtools library.
NOTE: Once this is merged, all live Scanners will be getting this data, so this can be a potentially dangerous change.
/retest slim-e2e-tests
/test slim-e2e-tests
/retest
Getting 403s, which is concerning. Perhaps we should look into getting an API key...
/retest
/retest
Failing e2e test is unrelated: https://github.com/stackrox/scanner/pull/1322
will merge tomorrow morning
Verified via these tests, checking for specific known ones, and ensuring StackRox 3.74.7 can handle the vulns without panicking and return scan results. Merging
|
2025-04-01T04:35:35.323573
| 2024-06-05T11:49:43
|
2335693134
|
{
"authors": [
"lvalerom"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11050",
"repo": "stackrox/stackrox",
"url": "https://github.com/stackrox/stackrox/pull/11394"
}
|
gharchive/pull-request
|
ROX-24527: Add deprecation notice for the summary API
Description
Deprecation notice for the v1/summary/counts API
The code will be fully delete here: #11392
Checklist
[x] Investigated and inspected CI test results
~[ ] Unit test and regression tests added~
~[ ] Evaluated and added CHANGELOG entry if required~
~[ ] Determined and documented upgrade steps~
~[ ] Documented user facing changes (create PR based on openshift/openshift-docs and merge into rhacs-docs)~
If any of these don't apply, please comment below.
Testing Performed
Here I tell how I validated my change
[x] CI
Reminder for reviewers
In addition to reviewing code here, reviewers must also review testing and request further testing in case the
performed one does not seem sufficient. As a reviewer, you must not approve the change until you understand the
performed testing and you are satisfied with it.
/retest
|
2025-04-01T04:35:35.391177
| 2022-02-06T11:46:57
|
1125181069
|
{
"authors": [
"stamateas"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11051",
"repo": "stamateas/upptime",
"url": "https://github.com/stamateas/upptime/issues/3463"
}
|
gharchive/issue
|
β οΈ GitLab Server has degraded performance
In 8908039, GitLab Server (https://gitlab01.its-telekom.eu) experienced degraded performance:
HTTP code: 200
Response time: 6917 ms
Resolved: GitLab Server performance has improved in 110fe90.
|
2025-04-01T04:35:35.441926
| 2021-12-29T02:02:27
|
1090220201
|
{
"authors": [
"Diggsey",
"Godzil",
"NeunEinser",
"SimonRovder",
"cnkeats",
"d-albrecht",
"santiagodg"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11052",
"repo": "standupmaths/xmastree2021",
"url": "https://github.com/standupmaths/xmastree2021/pull/25"
}
|
gharchive/pull-request
|
Adding 2020 examples as CSVs
2020 Examples
These examples were taken from the standupmaths/xmastree2020 repository.
The python scripts from 2020 were processed using santiagodg/xmastree-2020py-to-2021csv to generate 2021's animation CSV files and sometimes manually modifying them to make them work.
These CSVs have not been verified by the original animators.
Credits
Animation: Original Author
3d_octahedron: @danielkwalsh
3dfire: @Godzil
3dplasma: @Godzil
colour-waves: @Tirine
fire: @Diggsey
lightning: @Tirine
rainbow_rings: @tealnoise
rehoboam: @LoganPress
rgb-spheres: @NeunEinser
salesman: @SimonRovder
xmaslights-contagion: @christopherpaul
xmaslights-tetrahedron: @cnkeats
I didn't make "lightning"
I didn't make "lightning"
@Diggsey Thank you! I am now attributing 'lightning' to the correct author upon inspecting commit histories.
The RPi uses the color-coding GRB, so you have to interpret them accordingly.
Thanks, @d-albrecht! Nice catch, green fire felt fine to me for some reason. Will be fixing.
[ ] Fix output so that it outputs RGB instead of current GRB.
Hey thanks for running my script and crediting me! I'm not going to bother verifying the output since it's so long and I plan on doing something different this year, but I hope it runs!
It should be fine since the one thing I forgot last year was to tell the lights to turn on...
The RPi uses the color-coding GRB, so you have to interpret them accordingly.
Thanks, @d-albrecht! Nice catch, green fire felt fine to me for some reason. Will be fixing.
* [ ] Fix output so that it outputs RGB instead of current GRB.
Alternatively write a small script that swaps the colors around. That might be easier than integrating this into your replay analysis. You know, single responsibility principle! And as long as your rust code invokes this script automatically, no-one has to know that your replay analysis gets the colors wrong. ;)
I need to check how it looks with the pre-rendered method :)
I was unable to test my code on the real hardware before submitting it and it ended to be pretty slow (there are probably way of optimising it like not using Python XD ) but an actual 3D frame buffer and finding matching LEDs and doing average for each pixel is not a simple light task.
About rgb-spheres, it is supposed to repeat after 4800 frames, but because of a rounding error and the initial random starting radius not being based on a whole frame number, this is not the case with the file present in the 2020 repository.
I fixed those bugs now, to make a repeating animation for this format.
Here is a loopable csv file: https://www.dropbox.com/s/rzmb6xkn42rfzjd/rgb-spheres.csv?dl=0
And for reference the fixed code. Changed lines are 108 and 126.
def xmaslight():
# This is the code from my
#NOTE THE LEDS ARE GRB COLOUR (NOT RGB)
# Here are the libraries I am currently using:
import time
#from sim import board
#from sim import neopixel
import board
import neopixel
import re
import math
import random
# You are welcome to add any of these:
# import numpy
# import scipy
# import sys
# If you want to have user changable values, they need to be entered from the command line
# so import sys sys and use sys.argv[0] etc
# some_value = int(sys.argv[0])
# IMPORT THE COORDINATES (please don't break this bit)
#coordfilename = "./coords.txt"
coordfilename = "Python/coords.txt"
fin = open(coordfilename,'r')
coords_raw = fin.readlines()
coords_bits = [i.split(",") for i in coords_raw]
coords = []
for slab in coords_bits:
new_coord = []
for i in slab:
new_coord.append(int(re.sub(r'[^-\d]','', i)))
coords.append(new_coord)
#set up the pixels (AKA 'LEDs')
PIXEL_COUNT = len(coords) # this should be 500
pixels = neopixel.NeoPixel(board.D18, PIXEL_COUNT, auto_write=False)
# YOU CAN EDIT FROM HERE DOWN
# Calculates the distance of 2 vectors
def vdist(v1: list, v2: list):
if len(v1) != len(v2):
return -1
result = 0
for i in range(len(v1)):
result += (v1[i] - v2[i]) ** 2
return math.sqrt(result)
# Find coordinate that maximizes the distance for a given set of other coords
def find_furthest(points: list):
max_dist = 0
cur_pnt = points[0]
for coord in coords:
dist = math.inf
for p in points:
p_dist = vdist(p, coord)
if p_dist < dist:
dist = p_dist
if (dist > max_dist):
max_dist = dist
cur_pnt = coord
return cur_pnt
# init sphere origins.
# First sphere's origin is furthest from the coordinate system's origin
# Second sphere's origin is the LED with the greatest distance from the first sphere's origin
# Third sphere's origin is the LED where the distance for both other spheres is maximized.
sphere_origins = []
sphere_origins.append(find_furthest([[0, 0, 0]]))
sphere_origins.append(find_furthest(sphere_origins))
sphere_origins.append(find_furthest(sphere_origins))
# calculate maximum distance of any LED for each sphere's origin.
# Used to determine the max radius each sphere will ever receive
max_dists = [0, 0, 0]
for coord in coords:
for i in range(3):
dist = vdist(coord, sphere_origins[i])
if max_dists[i] < dist:
max_dists[i] = dist
# The rate in which each sphere enlargens. When negative, the sphere is currently shrinking.
increment_rates = [0, 0, 0]
# The radius of each sphere. Initial value is randomized
radii = [0, 0, 0]
# set initial increment rates and radii
for i in range(3):
# Frames per cycle for current sphere
frames = i * 40 + 120
increment_rates[i] = max_dists[i] / frames
# Random start radius
radii[i] = int(random.random()) * frames * increment_rates[i]
# infinitly many frames. Wohoo.
while True:
for i in range(PIXEL_COUNT):
# calculate color for current pixel. Each rgb (grb) color value is 255 * dist / max_dist
color = [0, 0, 0]
for s in range(3):
dist = abs(vdist(sphere_origins[s], coords[i]) - radii[s])
color[s] = int(255 * (1 - dist / max_dists[s]) ** 3)
pixels[i] = color
pixels.show()
# calculate radii for next iteration.
for s in range(3):
# Switch from enlarging to shrinking and vice versa, as needed
new_radius = radii[s] + increment_rates[s]
if new_radius > max_dists[s] + 1e-5 or new_radius < -1e-5:
increment_rates[s] = -increment_rates[s]
radii[s] += increment_rates[s]
return 'DONE'
# yes, I just put this at the bottom so it auto runs
xmaslight()
Hello, thanks for taking the time to do this! At a glance I would say salesman.py was converted correctly, there are 502 lines and they seem to gradually light more LEDs, so it certainly passes the basic sanity check.
But... Has the exact same tree been recreated? This is important because if the bulbs are in different locations then obviously the traveling salesman solution will not work anymore. The solution was precomputed and hardcoded because computing it took hours. π Just NP problems...
@SimonRovder Hi! Thank you for making the animation in the first place, it's a great addition and I really liked it!
Has the exact same tree been recreated?
Yes, I took your exact same submitted code from last year and made the new CSV with it. The produced animation does not perfectly match 2021's coordinates.
If the bulbs are in different locations then obviously the traveling salesman solution will not work anymore. The solution was precomputed and hardcoded because computing it took hours. π Just NP problems...
If you would like to update the salesman solution, that would be great! I would be glad to generate the CSV again, it would be no problem and better to have the animation be correct. For the most part, I think the current animation looks good and gets the effect across, though you sometimes do notice the traversal jumping incorrectly. You can review your animation using a simulator such as MPTree (mine) or other ones found linked in this repo or PRs.
@Godzil I re-rendered the 3dplasma animation and changed dimLight to 0.5 as suggested. Looks good, thank you!
The 3dfire animation was also updated during the fix of RGB instead of GRB issue. Hopefully it looks better for everyone now. :)
@NeunEinser Hi, I updated the file to be the one you shared for looping rgb-spheres. Thanks!
@SimonRovder salesman.csv is now removed. Thank you for being attentive to this PR anyway!
|
2025-04-01T04:35:35.448730
| 2021-05-24T18:41:58
|
899913957
|
{
"authors": [
"coveralls",
"gcampax"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11053",
"repo": "stanford-oval/almond-cloud",
"url": "https://github.com/stanford-oval/almond-cloud/pull/981"
}
|
gharchive/pull-request
|
update-dataset: validate commandpedia against approved subset of devices
So we don't show commands that are not available in Commandpedia
Pull Request Test Coverage Report for Build 3258
9 of 10 (90.0%) changed or added relevant lines in 2 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage increased (+0.01%) to 66.277%
Changes Missing Coverage
Covered Lines
Changed/Added Lines
%
util/admin-thingpedia-client.js
3
4
75.0%
Totals
Change from base Build 3256:
0.01%
Covered Lines:
6494
Relevant Lines:
9222
π - Coveralls
|
2025-04-01T04:35:35.450366
| 2019-11-05T23:01:29
|
518108559
|
{
"authors": [
"gcampax",
"rayslxu"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11054",
"repo": "stanford-oval/thingtalk",
"url": "https://github.com/stanford-oval/thingtalk/issues/166"
}
|
gharchive/issue
|
Make use of display value in confirms and responses
If the user says Seattle, response with "Weather for Seattle ..." , not "Weather for Seattle, King County, Washington, United States of America ..."
The specific location problem was fixed in https://github.com/stanford-oval/almond-cloud/pull/684 and the rest I don't think it's a bug. Closing.
|
2025-04-01T04:35:35.480729
| 2018-10-11T07:10:40
|
368981847
|
{
"authors": [
"keavil",
"kidzik"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11055",
"repo": "stanfordnmbl/osim-rl",
"url": "https://github.com/stanfordnmbl/osim-rl/issues/177"
}
|
gharchive/issue
|
Question about target velocity
After reading codes to generate target velocity here https://github.com/stanfordnmbl/osim-rl/blob/3ceadccc2f9104c9012281a482cfff5203f703bd/osim/env/osim.py#L499 , I got a question:
In description in #164 , you mentioned '(changing the heading of the velocity vector at the rate 20 degrees per second)'. But in the code, the value of poisson_lambda is 300 and it means the velocity changes about every 300 steps (3 seconds). Moreover, the code generate 10 different times to change the velocity and heading, but 7 of them are after 1000. It seems like a bug that poisson_lambda should be 100? Or it is the desired behavior?
Another related question is that currently the bonus for not falling is 10. This value is so high that even just standing there could receive more than 8000 rewards. Is this desired?
Thanks for your reply!
Yes, it's 3 seconds, we realized that a change every second might be too quick.
Regarding the high reward for not falling, the basic interpretation of the task is:
"Don't fall. If you manage to do that, follow the velocity vector" :)
All the solutions that don't fall will have exactly the same number of points for not falling, so the winning solution must also optimize for the velocity objective.
|
2025-04-01T04:35:35.483169
| 2023-10-13T22:42:16
|
1942734052
|
{
"authors": [
"boivado2",
"stanleyugwu"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11056",
"repo": "stanleyugwu/react-native-bottom-sheet",
"url": "https://github.com/stanleyugwu/react-native-bottom-sheet/issues/1"
}
|
gharchive/issue
|
Unable to resolve "./types"
Hello @boivado2 , thanks for opening this issue.
Can you please provide more context to this issue? What are you trying to do exactly?
Did you try to use the package or build it?
Hello again @boivado2 this issue has been resolved by PR #2 and a new version has been released 0.1.2
Thanks again for pointing this issue.
Thanks @stanleyugwu, pardon me for not giving a detailed explanation about the error. I was trying to use the package.
|
2025-04-01T04:35:35.496691
| 2019-01-29T19:20:00
|
404447078
|
{
"authors": [
"codepretty",
"layershifter"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11057",
"repo": "stardust-ui/react",
"url": "https://github.com/stardust-ui/react/issues/793"
}
|
gharchive/issue
|
Icon variable outline should be a prop
Feature Request
Problem description
Setting outline variable on an Icon is common enough that I think it would make more sense and be a much simpler design to have outline be a prop instead of variable.
<Icon name="call-video" variables={{ 'outline': true }} />
Proposed solution
A much simpler way to describe an Icon that shows its filled state would be...
<Icon name="call-video" outline />
Library
Link
State
Evergreen
https://evergreen.segment.com/components/icons/
Missing
MS Fabric
https://developer.microsoft.com/en-us/fabric#/components/icon
Missing
Material UI
https://material-ui.com/style/icons/
Has pattern, but the each Icon is a separate component (!)
Grommet
https://icons.grommet.io/
Missing
BlueprintJS
https://blueprintjs.com/docs/#core/components/icon
Missing
AtlasKit
https://atlaskit.atlassian.com/packages/core/icon
Missing
ReactIcons
https://www.npmjs.com/package/react-icons
Missing
AntDesign
https://ant.design/components/icon/
Has, as theme prop (!)
FontAwesome
https://fontawesome.com/icons/eye?style=regular https://fontawesome.com/icons/eye?style=solid
Has solid, regular and light. Icons in separate packages
I have mixed feelings about this prop π€
After the internal discussion we decided to implement it. Reason: may be it's not to common concept in Icon components, but it's a common design word in icons (for example, https://nucleoapp.com/premium-icons).
|
2025-04-01T04:35:35.499355
| 2020-10-02T16:26:03
|
713747888
|
{
"authors": [
"jeffreyscarpenter",
"olim7t"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11058",
"repo": "stargate/stargate",
"url": "https://github.com/stargate/stargate/issues/188"
}
|
gharchive/issue
|
Type check in ExternalStorage#supportsParameter is too permissive
Not an issue currently, but I ran into this on a temporary local branch.
@Override
public boolean supportsParameter(
ParameterContext parameterContext, ExtensionContext extensionContext)
throws ParameterResolutionException {
return parameterContext.getParameter().getType().isAssignableFrom(ClusterConnectionInfo.class);
}
According to the API contract, this will match if the parameter type "is either the same as, or a superclass or interface of" ClusterConnectionInfo. In other words the connection info will be injected into any Object parameter.
This is too permissive: if you have parameterized tests, it's possible that one of them could use an Object parameter. Either explicitly, or via erasure if the test method is parameterized (that's how I ran into it, I can provide an example if this is not clear).
If Junit finds two parameter resolvers that handle the same type, it throws an exception.
I think exact equality would be sufficient in this case, Cluster is private so we know it can't be referenced from a test class:
return parameterContext.getParameter().getType() == ClusterConnectionInfo.class;
can't repro
|
2025-04-01T04:35:35.502338
| 2022-10-14T13:10:21
|
1409332186
|
{
"authors": [
"ivansenic"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11059",
"repo": "stargate/stargate",
"url": "https://github.com/stargate/stargate/pull/2157"
}
|
gharchive/pull-request
|
closes #2131: fixes swagger issues in the docs api v2
What this PR does:
Fixes all Swagger issues in the docs API:
fields and where passed as string (JSON), unfortunately layout in Swagger is not perfect
document-path param was sent wrongly, again fallback to string
fixed example for schema attaching
minor documentation improvements
Which issue(s) this PR fixes:
Fixes #2131
@tatu-at-datastax Not sure if this is relevant for the restapi, but you should have a look..
|
2025-04-01T04:35:35.510409
| 2018-10-19T00:33:26
|
371763545
|
{
"authors": [
"jhunt",
"krutten"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11060",
"repo": "starkandwayne/shield",
"url": "https://github.com/starkandwayne/shield/pull/467"
}
|
gharchive/pull-request
|
Make S3 region configurable
This makes life easier on some S3 work-alikes, which can require that
S3 operations be given region names that do not correspond to those
used by AWS itself.
Fixes #465
Changes look good. Defaults remain the same if not configure.
|
2025-04-01T04:35:35.522417
| 2022-08-22T17:11:03
|
1346739908
|
{
"authors": [
"Goldziher",
"cofin"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11062",
"repo": "starlite-api/starlite",
"url": "https://github.com/starlite-api/starlite/issues/396"
}
|
gharchive/issue
|
Bug: OpenAPI Schema Should Generate FileUpload Form in Swagger Docs.
Describe the bug
The bug may be broader than this, but it was first noticed when testing multi-part file uploads with the integrated swagger docs.
To Reproduce
Given a route that looks like this:
import logging
from starlite import Body, RequestEncodingType, UploadFile, post
logger = logging.getLogger(__name__)
@post(path="/upload", cache=False, tags=["Collection"])
async def handle_collection_upload(
data: UploadFile = Body(media_type=RequestEncodingType.MULTI_PART),
) -> dict[str, str]:
"""Upload a file"""
logger.info("Processing Uploaded File")
return {
"status": "file uploaded",
}
The OpenAPI specs are not generating the schema section of the multipart-form-data section. Here is the generated OpenAPI doc:
"/upload": {
"post": {
"tags": [
"Collection"
],
"operationId": "Handle Collection Upload",
"requestBody": {
"content": {
"multipart/form-data": {
"schema": {}
}
},
"required": false
},
"responses": {
"201": {
"description": "Document created, URL follows",
"headers": {},
"content": {
"application/json": {
"schema": {
"type": "object"
}
}
}
},
"400": {
"description": "Bad request syntax or unsupported method",
"content": {
"application/json": {
"schema": {
"properties": {
"status_code": {
"type": "integer"
},
"detail": {
"type": "string"
},
"extra": {
"additionalProperties": {},
"type": [
"null",
"object",
"array"
]
}
},
"type": "object",
"required": [
"detail",
"status_code"
],
"description": "Validation Exception",
"examples": [
{
"status_code": 400,
"detail": "Bad Request",
"extra": {}
}
]
}
}
}
}
},
"deprecated": false
}
}
Additional context
Here is what is generated from a similar endpoint in FastAPI:
"/api/workspaces/{workspace_id}/upload": {
"put": {
"tags": ["Workspace"],
"summary": "Upload Advisor",
"description": "Upload file to advisor",
"operationId": "uploadAdvisor",
"parameters": [
{
"required": true,
"schema": { "title": "Workspace Id", "type": "integer" },
"name": "workspace_id",
"in": "path"
}
],
"requestBody": {
"content": {
"multipart/form-data": {
"schema": { "$ref": "#/components/schemas/Body_uploadAdvisor" }
}
},
"required": true
},
"responses": {
"200": {
"description": "Successful Response",
"content": {
"application/json": {
"schema": { "$ref": "#/components/schemas/Workspace" }
}
}
},
"422": {
"description": "Validation Error",
"content": {
"application/json": {
"schema": { "$ref": "#/components/schemas/HTTPValidationError" }
}
}
}
},
"security": [{ "OAuth2PasswordBearerCookie": [] }]
}
},
I've also attached a screenshot for what the Swagger docs generate:
Should I close this or wait until the next release is made?
lets close it
|
2025-04-01T04:35:35.525838
| 2020-03-16T09:54:10
|
582151524
|
{
"authors": [
"Atsidir",
"tfoldi"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11063",
"repo": "starschema/COVID-19-data",
"url": "https://github.com/starschema/COVID-19-data/pull/19"
}
|
gharchive/pull-request
|
airflow etl dag generator
implemented, as it was described in https:/β¦/github.com/starschema/COVID-19-data/wiki/dynamic-dag
I happy to write the test case for this (at least I understand what it does :D)
|
2025-04-01T04:35:35.545982
| 2019-08-05T19:48:20
|
477027743
|
{
"authors": [
"prein2"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11064",
"repo": "stas-demydiuk/domoticz-zigbee2mqtt-plugin",
"url": "https://github.com/stas-demydiuk/domoticz-zigbee2mqtt-plugin/issues/135"
}
|
gharchive/issue
|
ikea control outlet E1603/E1702
Device description
IKEA control outlet, bought today doesn't work with domoticz. Using latest dev version zigbee2mqtt.
message from domoticz log:
2019-08-05 21:44:59.689 (Zigbee2MQTT) MQTT message: zigbee2mqtt/plug tuinberegening {'state': 'ON', 'linkquality': 21}
2019-08-05 21:44:59.690 (Zigbee2MQTT) Unsupported zigbee device type with model "E1603/E1702"
MQTT message with device information
8/5/2019, 9:24:59 PM - info: New device 'TRADFRI control outlet' with address 0x000d6ffffeca916d connected!
8/5/2019, 9:24:59 PM - info: MQTT publish: topic 'zigbee2mqtt/bridge/log', payload '{"type":"device_connected","message":"0x000d6ffffeca916d","meta":{"modelID":"TRADFRI control outlet"}}'
MQTT messages from device
8/5/2019, 9:35:00 PM - info: MQTT publish: topic 'zigbee2mqtt/plug tuinberegening', payload '{"state":"ON","linkquality":23}'
Fixed by adding folling line to adapters/init.py:
'E1603/E1702': OnOffSwitchAdapter, # IKEA TRADFRI control outlet
don't know how to create a pull request though..
|
2025-04-01T04:35:35.640409
| 2022-08-28T13:40:55
|
1353357091
|
{
"authors": [
"dstpierre",
"rostikts"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11065",
"repo": "staticbackendhq/core",
"url": "https://github.com/staticbackendhq/core/pull/49"
}
|
gharchive/pull-request
|
WIP: add local cache sub/pub events
Add a local observer for the Pub/sub cache events
Observer provides a possibility to subscribe to a specific event and publish a message for all subscribers of the channel.
The memObserver struct contains a map with a slice of subscribers where the key of the map is a specific channel.
The subscriber is basically a wrapper for the channel that is listening for the changes
@dstpierre Hi, the task is still in progress, I'll update the logic according to the Redis cache, but the core of the dev SubPub is done, could you take a look at the Observer implementation?
Also, I want to clarify a few things:
IMO it's a little bit confusing that the subscription logic and the event listening are implemented in the single method Subscribe(send chan Command, token, channel string, close chan bool). And the current implementation breaks the SRP principle.
I'm confused about these two conditions:
if msg.Type == internal.MsgTypeChanIn {
msg.Type = internal.MsgTypeChanOut
} else if msg.IsSystemEvent {
Should there be some business logic that I'm not aware of?
Yep, you're deep in SB event / messaging aspect.
This is to prevent an infinite loop for the following scenario:
The user would subscribe to a topic and have a server-side function run on new message.
If they were to push to that very own topic from the function, it would get re-trigger without the "SystemEvent" type.
Imagine a live chat for example, where the intent is to call a server-side function for each messages in a topic.
This entire part of StaticBackend (event / messages) is still in its early design phase.
@rostikts this look good to me so far.
One small thing, I'd rename the dev_observer package either just observer or devobserver if we need to keep the "dev", but we could lose it IMO. Same for the dev_cache_observer.go, i'd just name it observer.go in the observer package.
Underscore in package name is to be avoided as much as possible.
Am I understanding correctly that the remaining implementation is for the PublishDocument function and this would be ready for more test?
Thanks for this PR by the way :wave:
@dstpierre Yep, it's still in progress, you're right, the PublishDocumentfunction and more tests are planned in the next iteration, it should be ready today, maybe Saturday. Also, I should review the func (o *memObserver) Publish(channel string, msg interface{}) error function, I guess I should add here timer with some timeout after which the existing message in channel will be discarded to avoid the blocking
|
2025-04-01T04:35:35.729875
| 2022-09-14T13:38:17
|
1373005627
|
{
"authors": [
"yyoncho",
"zah"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11066",
"repo": "status-im/nim-workspace",
"url": "https://github.com/status-im/nim-workspace/pull/3"
}
|
gharchive/pull-request
|
Bootstrap nimble
PR split into 2 commits to make the reviewing easier.
Here it is the code that I used for generating the nimble bootstrap files
#!/usr/bin/env bash
rm ~/.cache/nim/nimble_d/ -rf
nim cc --genScript src/nimble.nim
sed -i 's#-I.*nimble/src##' ~/.cache/nim/nimble_d/compile_nimble.sh
cp ~/Sources/nim/Nim/lib/nimbase.h ~/.cache/nim/nimble_d/
mkdir ~/.cache/nim/nim-workspace/ -p
rm -rf ~/Sources/nim/nim-workspace/scripts/nimble-linux/
cp -r ~/.cache/nim/nimble_d/ ~/Sources/nim/nim-workspace/scripts/nimble-linux/
# windows
rm ~/.cache/nim/nimble_d/ -rf
nim cc --os:windows --genScript src/nimble.nim
sed -i 's#-I.*nimble/src##' ~/.cache/nim/nimble_d/compile_nimble.bat
cp ~/Sources/nim/Nim/lib/nimbase.h ~/.cache/nim/nimble_d/
mkdir ~/.cache/nim/nim-workspace/ -p
rm -rf ~/Sources/nim/nim-workspace/scripts/nimble-windows/
cp -r ~/.cache/nim/nimble_d/ ~/Sources/nim/nim-workspace/scripts/nimble-windows/
The Nimble C sources should be kept in a separate repo. This would allow us to easily prune the history of this repo to control the initial git clone size.
|
2025-04-01T04:35:35.965647
| 2017-07-13T16:13:28
|
242754953
|
{
"authors": [
"kgryte",
"pcowgill"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11067",
"repo": "stdlib-js/stdlib",
"url": "https://github.com/stdlib-js/stdlib/issues/93"
}
|
gharchive/issue
|
"numerical computing" is more common than "numeric computing"
Checklist
Please ensure the following tasks are completed before filing an issue.
[x] Read and understood the Code of Conduct.
[x] Searched for existing issues and pull requests.
[x] If this is a general question, searched the FAQ for an existing answer.
[x] If this is a feature request, the issue name begins with RFC: .
Description
Description of the issue (or feature request).
The README should say "numerical computing" rather than "numeric computing".
Related Issues
Does this issue (or feature request) have any related issues?
No.
Questions
Any questions for reviewers?
No.
Other
Any other information relevant to this issue (or feature request)? This may include screenshots, references, stack traces, sample output, and/or implementation notes.
https://trends.google.com/trends/explore?date=all&q="numeric computing","numerical computing"
@pcowgill Thanks for the link. I think, however, we will pass on changing our usage of "numeric computing" to "numerical computing". We find references to both within the field, and our intent is to indicate "computing with numbers", while avoiding confusion with numerical analysis.
If this becomes a stumbling block in the future, we can revisit, but, for the time being, we will leave as is.
|
2025-04-01T04:35:35.969820
| 2024-07-21T21:17:45
|
2421604818
|
{
"authors": [
"headlessNode"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11068",
"repo": "stdlib-js/stdlib",
"url": "https://github.com/stdlib-js/stdlib/pull/2645"
}
|
gharchive/pull-request
|
test: add tests to ndarray/base/nullary
Progresses #2229.
Description
What is the purpose of this pull request?
This pull request:
add tests to @stdlib/ndarray/base/nullary/test/test.7d.js for 100% test coverage
Related Issues
Does this pull request have any related issues?
This pull request:
progresses #2229
Questions
Any questions for reviewers of this pull request?
No.
Other
Any other information relevant to this pull request? This may include screenshots, references, and/or implementation notes.
No.
Checklist
Please ensure the following tasks are completed before submitting this pull request.
[x] Read, understood, and followed the contributing guidelines.
@stdlib-js/reviewers
@kgryte Please review, thanks.
|
2025-04-01T04:35:35.985603
| 2018-06-15T07:19:33
|
332670216
|
{
"authors": [
"relativityboy",
"roboza"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11069",
"repo": "steemit/devportal-tutorials-js",
"url": "https://github.com/steemit/devportal-tutorials-js/pull/88"
}
|
gharchive/pull-request
|
28 vote on content
Closes #28
@roboza - What part of this commit gives the user feedback on whether or not the vote succeeded or failed? A link to the broadcast operation doesn't do anything to address that PR requirement.
You need to craft a user interface element that runs in the tutorial and gives the user visual feedback as to the success or failure of the operation.
After further consideration, it also looks like the testnet gets so few posts that it can be difficult to find something to vote on. This tutorial will also need to generate a piece of content on which to vote.
@roboza - I can see you put a lot of work into it. But in the UI you've 'fixed' the username in place as well as the posting key. When the testnet resets (and it does; it's a rule of the testnet), this tutorial will break again.
|
2025-04-01T04:35:35.991653
| 2018-04-19T12:34:55
|
315857352
|
{
"authors": [
"bonustrack",
"gregory-latinier"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11070",
"repo": "steemit/steemconnect",
"url": "https://github.com/steemit/steemconnect/issues/251"
}
|
gharchive/issue
|
Setup dedicated node for SteemConnect for server side usage
SteemConnect is using public node api.steemit.com to connect to Steemd API. When this node is not working, 3p apps relying on SC canot broadcast transaction anymore. SteemConnect could have his own dedicated node. Steemd API url can remain secret since transaction for SC posting API happen on server side.
Depends on #250
How do you want to proceed with this ?
Should I setup a node first to test on a VM for instance and then we rent a server on AWS ?
|
2025-04-01T04:35:36.033784
| 2021-11-07T09:28:43
|
1046700093
|
{
"authors": [
"gergoerdi",
"stefan-hoeck"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11071",
"repo": "stefan-hoeck/idris2-webidl",
"url": "https://github.com/stefan-hoeck/idris2-webidl/issues/7"
}
|
gharchive/issue
|
How do I actually run it?
Because idris2-dom is missing IndexedDB support, I thought I'd just plop in the WebIDL files from e.g. https://searchfox.org/mozilla-central/source/dom/webidl and it should Just Work (tm), but alas, I can't even figure out how to run it. It seems to look for an .idr output file in some directory somewhere (that I can kind-of override mostly with the -o flag, if only I pass the -o flag before the input file name?!), but I thought the whole point is to create the .idr file.
Also, one of the outputs seems to be the Web.Internal.Types module, but that module is already part of idris2-dom so I will have a bad time if my IndexedDB FFI package also has that module. Or is the idea that it would have to be included with idris2-dom, together with all other W3C specs past, present and future?
OK I kind of figured out how it works: basically, it needs $OUT/Web/Raw and $OUT/Web/Internal directories to already exist (if run with -o $OUT), but otherwise doesn't seem to do any readback.
Which turns out to be a problem on its own, because it means I can't re-run it on idris2-dom, without knowing the exact set of input files. Otherwise, it removes all the existing definitions from Web.Interlal.Types.
Yeah sorry, running this is very basic, but it didn't bother me so far... Here's how to run this after building it with make lib:
$ build/exec/generateDomBindings -o outdir idl/*.webidl
where outdir is the source dir of the idris2-dom project (for instance: ../idris2-dom/src). And yes, I need to make this more accessible, possibly by adding a proper make target which automatically generates the necessary directories in outdir.
Please note also that it makes no sense to run this with a single .webidl file, since many interfaces are spread across several files.
And yes, the Web.Internal.Types is hardcoded so far. Another thing that needs changing, if this should be used by other projects than idris2-dom.
Yeah sorry, running this is very basic, but it didn't bother me so far... Here's how to run this after building it with make lib:
$ build/exec/generateDomBindings -o outdir idl/*.webidl
Ah, I missed that /idl directory in idris2-webidl because I was looking for it in idris2-dom (since in my mind, idris2-webidl was a generic tool and idris2-dom was just one use of it, with its own set of IDLs to compile). So in that case, I should be able to re-create the whole of idris2-dom with the addition of the IndexedDB interfaces. Let me try just that.
|
2025-04-01T04:35:36.038523
| 2023-06-09T14:30:17
|
1750009690
|
{
"authors": [
"Timmmy-nlb",
"gr2m"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11072",
"repo": "stefanbuck/github-issue-parser",
"url": "https://github.com/stefanbuck/github-issue-parser/issues/58"
}
|
gharchive/issue
|
Info error at run
Hello,
I have always a error message, but it works as expected:
Error: Error: ENOENT: no such file or directory, open '.github/ISSUE_TEMPLATE/Bestellung-einer-Benutzeraenderung.yml'
- uses: stefanbuck/github-issue-parser@v3
id: issue-parser
with:
template-path: .github/ISSUE_TEMPLATE/Bestellung-einer-Benutzeraenderung.yml # optional but recommended
But the template is in the same repo as the workflow and the issue.
Whats wrong?
try running actions/checkout beforehand
- uses: actions/checkout@v3
That was the solution, thank you.
|
2025-04-01T04:35:36.064710
| 2016-12-15T13:42:58
|
195808479
|
{
"authors": [
"kevinvanmierlo",
"stefanhaustein"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11073",
"repo": "stefanhaustein/HtmlView2",
"url": "https://github.com/stefanhaustein/HtmlView2/issues/2"
}
|
gharchive/issue
|
Help/Wiki + release on Android Arsenal
This is the lightweight webview I've been looking for a long time. I have had a few apps where I load local html (nothing fancy, only bold, italic, ul, ol). But there wasn't a library that could do it all except WebView which is unbelievably slow. So first, thank you!
I've just been using the library, but there is not much explanation. For example I'm trying to set the base text size. But no clue how to do this. Also I find the indentation for ol and ul lists too big. Is there a way to make that smaller?
Like I said, I've been looking for this a long time. So maybe put it on Android Arsenal or other websites where people can find it and be as excited as me.
Will consider when I have addressed the other issue. What is your relation to AndroidArsenal?
@stefanhaustein I will file bug reports as other issues.
I'm not related to AndroidArsenal, it's just the website where I go to look for specific libraries. So if this library is ready for release, I would put it on multiple library sites. AndroidArsenal was just the first that popped into my mind.
Wouldn't it make more sense to publish to jcenter first for gradle / maven integration using gradle? I guess part of the infrastructure could then be re-used for AndroidArsenal?
@stefanhaustein Yeah that definitely makes more sense.
Android Arsenal is not a website to push your library to. It's just a collection of libraries with a link to the webpage where you can get the library (like github.com). So it's just a website to discover new libraries.
Ok, that makes publishing to AndroidArsenal easier -- as it doesn't seem to require any additional changes then... :)
@stefanhaustein Haha yeah! That's why I made this issue. I always look for libraries there
Ok, I have created a release that should be directly accessible in gradle via jitpack. Would you mind double-checking the instructions at the end of the the main readme.md befor I advertise this on Android Arsenal?
@stefanhaustein Sorry for the late response, I'm a bit busy at the moment. From what I can see it is fine. Perhaps in the future some advanced options in the wiki, but for now with the demo it is fine.
|
2025-04-01T04:35:36.099379
| 2015-12-08T03:46:33
|
120923827
|
{
"authors": [
"ApesTalk",
"Freeway1979",
"hustlzp",
"steipete",
"wangqianjun",
"xiaochaoren"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11075",
"repo": "steipete/Aspects",
"url": "https://github.com/steipete/Aspects/issues/73"
}
|
gharchive/issue
|
why can't hook viewDidLoad (UIViewController) ??
Hi Steipete, why can't hook viewDidLoad (UIViewController) ??
Error:
Aspects: Blog signature <NSMethodSignature: 0x7fbc8bd35740> doesn't match <NSMethodSignature: 0x7fbc8bf0cc40>.
Please help~
Thank u.
Hi, if you post some code I can help.
Steipete, I create a Category for UIViewcontroller, and have a Class method viewDidloadLogging:
(void)viewDidloadLogging
{
[UIViewController aspect_hookSelector:@selector(viewDidLoad) withOptions:AspectPositionAfter usingBlock:^(idaspectInfo, BOOL animated){
NSLog(@"View Controller %@ will appear animated: %tu", [aspectInfo.instance class], animated);
NSLog(@"arguments:%@",[aspectInfo arguments]);
NSLog(@"originalInvocation:%@",[aspectInfo originalInvocation]);
} error:NULL];
}
In my demo AppDelegate,
(BOOL)application:(UIApplication *)application didFinishLaunchingWithOptions:(NSDictionary *)launchOptions {
// Override point for customization after application launch.
[UIViewController viewDidloadLogging];
return YES;
}
There are no problems usingοΌ
+1
+1's don't help. Somebody who digs in and figures out what fails would help though. Thanks!
So that question answer is whatοΌSteipeteοΌ^.^
This is the problem of execution order of [UIViewController viewDidLoad] and [UIViewController viewDidloadLogging]. Please make sure [UIViewController viewDidloadLogging] is invoked before [UIViewController viewDidLoad].
I got the same issue while hooking viewDidLoad selector of UIViewController class.
I can't call the block I provided but "dealloc" is working.
Can you figure out the solution?
Thanks a lot.
The code as following:
(BOOL)application:(UIApplication *)application didFinishLaunchingWithOptions:(NSDictionary *)launchOptions {
// Ignore hooks when we are testing.
if (!NSClassFromString(@"XCTestCase")) {
[[LogManager sharedInstance] logViewControllerLifeCycle];
}
//...other code
}
(void)logViewControllerLifeCycle
{
[UIViewController aspect_hookSelector:@selector(viewDidLoad)
withOptions:AspectPositionBefore
usingBlock:^(id info, BOOL animated) {
UIViewController *controller = [info instance];
NSString *key = NSStringFromClass([controller class]);
NSInteger refCount = [[self.dicViewControllerRefCount objectForKey:key] integerValue];
refCount++;
[self.dicViewControllerRefCount setObject:[NSString smc_stringFromInteger:refCount] forKey:key];
NSLog(@"%s %@ viewDidLoad",FUNCTION,key);
NSLog(@"%s alloc %@,%ld",FUNCTION,key,refCount);
} error:NULL];
//...other code
}
Hi everyone,
I made it.
The root cause is that "aspect_isCompatibleBlockSignature" verifies that argument count between blockSignature and methodSignature,and if not matched ,nil returned.
You can see these code :
if (blockSignature.numberOfArguments > methodSignature.numberOfArguments) {
signaturesMatch = NO; //You will go here if wrong block parameters provide as above.
}else {
Now ,we can correct it by changing
[UIViewController aspect_hookSelector:@selector(viewDidLoad)
withOptions:AspectPositionBefore
usingBlock:^(id info, BOOL animated) {
to
[UIViewController aspect_hookSelector:@selector(viewDidLoad)
withOptions:AspectPositionBefore
usingBlock:^(id info) {
It works now for "viewDidLoad".
Hi Steipete,
I suggest to throw an exception here for developers to correct it to right things.
Thanks.
How can i use AspectPositionInstead option replace viewDidLoad method? Any help would be appreciated.
[MySubViewController aspect_hookSelector:@selector(viewDidLoad) withOptions:AspectPositionInstead usingBlock:^(id info){
//How can i use super here ???
// [super viewDidLoad];
WTSubViewController *vc = (WTSubViewController *)info.instance;
vc.view.backgroundColor = [UIColor blueColor];
} error:NULL];
|
2025-04-01T04:35:36.113296
| 2022-08-05T23:26:01
|
1330570627
|
{
"authors": [
"Shaptic"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11076",
"repo": "stellar/go",
"url": "https://github.com/stellar/go/pull/4513"
}
|
gharchive/pull-request
|
exp/lighthorizon: Correctly set Content-Type, plus JSONify errors
What
This sets the content-type at the right time and also makes error messages proper JSON.
Why
The headers can't be set after Write() is called on w.
@sreuland yess absolutely, the tech debt is building up :sob:
|
2025-04-01T04:35:36.120839
| 2024-06-07T20:19:20
|
2341104898
|
{
"authors": [
"chowbao",
"tamirms"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11077",
"repo": "stellar/go",
"url": "https://github.com/stellar/go/pull/5337"
}
|
gharchive/pull-request
|
services/horizon: Update guide for developers
PR Checklist
PR Structure
[ ] This PR has reasonably narrow scope (if not, break it down into smaller PRs).
[ ] This PR avoids mixing refactoring changes with feature changes (split into two PRs
otherwise).
[ ] This PR's title starts with name of package that is most changed in the PR, ex.
services/friendbot, or all or doc if the changes are broad or impact many
packages.
Thoroughness
[ ] This PR adds tests for the most critical parts of the new functionality or fixes.
[ ] I've updated any docs (developer docs, .md
files, etc... affected by this change). Take a look in the docs folder for a given service,
like this one.
Release planning
[ ] I've updated the relevant CHANGELOG (here for Horizon) if
needed with deprecations, added features, breaking changes, and DB schema changes.
[ ] I've decided if this PR requires a new major/minor version according to
semver, or if it's mainly a patch change. The PR is targeted at the next
release branch if it's not a patch change.
What
Improved services/horizon/docker/start.sh script so that it could be executed from any directory. Previously, the script only worked if you executed it in the services/horizon/docker directory
Updated the configuration in the developer docs to use the protocol 21 stellar core version
Fixed the docker compose commands in the "Database Setup" so that they reference the docker compose files from the repository root.
Added " --remove-orphans -v" flags to the docker-compose down command so that the db volumes are wiped out
Updated the environment variables in the IDE configuration
@chowbao can you take another look at the PR? I think I have addressed all your code review feedback
@chowbao can you take another look at the PR? I think I have addressed all your code review feedback
I think all the content looks good. I still wonder if the docs are easy enough for a new user to use though. But I don't think that should block the PR
nit: it would be nice if there was a table of contents.
|
2025-04-01T04:35:36.123647
| 2017-05-08T07:18:15
|
226959133
|
{
"authors": [
"PikiPoki",
"bartekn"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11078",
"repo": "stellar/horizon",
"url": "https://github.com/stellar/horizon/issues/362"
}
|
gharchive/issue
|
Unable to send Lumens
{
"type": "https://stellar.org/horizon-errors/server_error",
"title": "Internal Server Error",
"status": 500,
"detail": "An error occurred while processing this request. This is usually due to a bug within the server software. Trying this request again may succeed if the bug is transient, otherwise please report this issue to the issue tracker at: https://github.com/stellar/horizon/issues. Please include this response in your issue.",
"instance": "horizon-live-002/WbPeIf66nB-25841297"
}
Seems to be working again.
|
2025-04-01T04:35:36.124769
| 2022-04-27T23:24:19
|
1217978102
|
{
"authors": [
"marcelosalloum"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11079",
"repo": "stellar/java-stellar-anchor-sdk",
"url": "https://github.com/stellar/java-stellar-anchor-sdk/issues/220"
}
|
gharchive/issue
|
Update SEP-31 to be able to use Circle to generate the transaction memos
A complementary part of Circle Payment Detection
GanttStart: 2022-04-27
GanttDue: 2022-04-29
Cc @reecexlm
|
2025-04-01T04:35:36.127895
| 2022-08-01T18:22:06
|
1324800295
|
{
"authors": [
"jonjove",
"leighmcculloch"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11080",
"repo": "stellar/rs-soroban-sdk",
"url": "https://github.com/stellar/rs-soroban-sdk/pull/386"
}
|
gharchive/pull-request
|
Fix contractmetav0 custom section not emitted
What
Move contractmetav0 custom section link section static variable into function that is exported instead of being a crate root static, and reference that exported function in the crates of contracts.
Why
The custom section is not always in contracts when contracts are built with codegen-units greater than 1. Because the custom section is not in the contract crate itself, it could end up in an object file for the SDK that doesn't end up getting linked. When codegen-units is 1 this isn't an issue because there should only be one.
We recommend that codegen-units should always be 1 to improve the optimization of the contract builds, but this is an annoying footgun for anyone who misses that.
By moving the link section into an exported function that gets referenced in the contract crate, the object file should always be included in linking. We reference it in every contract function because there is no way to just do it once for all functions since macros have no contract state.
I tested this change with the examples to make sure that it didn't increase the size of the binary. No binaries changed.
Close https://github.com/stellar/rs-soroban-sdk/issues/383
@leighmcculloch as far as I can tell, this doesn't actually cause the section to be generated on the token contract (when codegen-units > 1). Am I doing something wrong?
|
2025-04-01T04:35:36.131222
| 2015-09-15T14:24:51
|
106567392
|
{
"authors": [
"nullstyle",
"sacarlson"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11081",
"repo": "stellar/ruby-stellar-base",
"url": "https://github.com/stellar/ruby-stellar-base/issues/16"
}
|
gharchive/issue
|
Can add account signers but can't delete them
Seems I can add account signers with no problem but just can't delete them when I set weight to zero they are still listed on the stellar-core db with a weight of 1. the function I run that works to add but fails to delete with weight set to 0 is:
def add_signer_public_key(account, key, weight)
set_options account, signer: Stellar::Signer.new({
pub_key: key,
weight: weight
})
end
but I also tried to delete a signer from the developer web interface https://www.stellar.org/developers/tools/client/#/accountmanager and find I can't delete it from there also but have no problem adding them there. so maybe the problem is at the stellar-core level or something else?
account I was using:
"address": "GDBGEVX7CD5AS3ECPPJMDUGXK25DFUM5JK5DHDSMNYND46SPC2OMDJZY",
"secret": "SBPXGA7EEX3Y6LSNNZW3OBXI5UBV2PLH65PTBNNOWCIVDJVSHUSBNJGS"
}
Closing this in favor of https://github.com/stellar/stellar-core/issues/755
|
2025-04-01T04:35:36.150022
| 2024-01-06T01:19:17
|
2068271671
|
{
"authors": [
"marcelosalloum",
"stellar-jenkins"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11082",
"repo": "stellar/stellar-disbursement-platform-backend",
"url": "https://github.com/stellar/stellar-disbursement-platform-backend/pull/142"
}
|
gharchive/pull-request
|
[SDP-1006] Add the tenant_id column to the TSS submitter_transactions table
What
Add the tenant_id column to the TSS submitter_transactions table.
Why
So we can associate every stellar_transaction to one tenant, using this to execute per-tenant transactions using each Tenant's separate balance.
Closes https://stellarorg.atlassian.net/browse/SDP-1006.
Known limitations
[TODO or N/A]
Checklist
PR Structure
[ ] This PR has a reasonably narrow scope (if not, break it down into smaller PRs).
[ ] This PR title and description are clear enough for anyone to review it.
[ ] This PR does not mix refactoring changes with feature changes (split into two PRs otherwise).
Thoroughness
[ ] This PR adds tests for the new functionality or fixes.
[ ] This PR contains the link to the Jira ticket it addresses.
Configs and Secrets
[x] No new CONFIG variables are required -OR- the new required ones were added to the helmchart's values.yaml file.
[x] No new CONFIG variables are required -OR- the new required ones were added to the deployments (pr-preview, dev, demo, prd).
[x] No new SECRETS variables are required -OR- the new required ones were mentioned in the helmchart's values.yaml file.
[x] No new SECRETS variables are required -OR- the new required ones were added to the deployments (pr-preview secrets, dev secrets, demo secrets, prd secrets).
Release
[ ] This is not a breaking change.
[ ] This is ready for production.. If your PR is not ready for production, please consider opening additional complementary PRs using this one as the base. Only merge this into develop or main after it's ready for production!
Deployment
[ ] Does the deployment work after merging?
stellar-disbursement-platform-backend-preview is available here:SDP: https://sdp-backend-pr142.previews.kube001.services.stellar-ops.com/healthAP: https://sdp-ap-pr142.previews.kube001.services.stellar-ops.com/healthFrontend: https://sdp-backend-dashboard-pr142.previews.kube001.services.stellar-ops.com
stellar-disbursement-platform-backend-preview is available here:SDP: https://sdp-backend-pr142.previews.kube001.services.stellar-ops.com/healthAP: https://sdp-ap-pr142.previews.kube001.services.stellar-ops.com/healthFrontend: https://sdp-backend-dashboard-pr142.previews.kube001.services.stellar-ops.com
stellar-disbursement-platform-backend-preview is available here:SDP: https://sdp-backend-pr142.previews.kube001.services.stellar-ops.com/healthAP: https://sdp-ap-pr142.previews.kube001.services.stellar-ops.com/healthFrontend: https://sdp-backend-dashboard-pr142.previews.kube001.services.stellar-ops.com
stellar-disbursement-platform-backend-preview is available here:SDP: https://sdp-backend-pr142.previews.kube001.services.stellar-ops.com/healthAP: https://sdp-ap-pr142.previews.kube001.services.stellar-ops.com/healthFrontend: https://sdp-backend-dashboard-pr142.previews.kube001.services.stellar-ops.com
stellar-disbursement-platform-backend-preview is available here:SDP: https://sdp-backend-pr142.previews.kube001.services.stellar-ops.com/healthAP: https://sdp-ap-pr142.previews.kube001.services.stellar-ops.com/healthFrontend: https://sdp-backend-dashboard-pr142.previews.kube001.services.stellar-ops.com
stellar-disbursement-platform-backend-preview is available here:SDP: https://sdp-backend-pr142.previews.kube001.services.stellar-ops.com/healthAP: https://sdp-ap-pr142.previews.kube001.services.stellar-ops.com/healthFrontend: https://sdp-backend-dashboard-pr142.previews.kube001.services.stellar-ops.com
|
2025-04-01T04:35:36.163774
| 2020-10-18T15:23:33
|
724030863
|
{
"authors": [
"IoanaAlexandru",
"campovski"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11083",
"repo": "stelynx/network_image_mock",
"url": "https://github.com/stelynx/network_image_mock/issues/2"
}
|
gharchive/issue
|
Package breaks with Flutter 1.23
Tests using mockNetworkImagesFor now fail after upgrading from 1.22.0-12.4.pre to 1.23.0-18.1.pre, on the beta channel. This is the error:
Stack trace
βββ‘ EXCEPTION CAUGHT BY IMAGE RESOURCE SERVICE βββββββββββββββββββββββββββββββββββββββββββββββββββββ
The following ReachabilityError was thrown resolving an image codec:
`null` encountered as case in a switch expression with a non-nullable enum type.
I will check it out in greater detail tomorrow, however this package should not have any problems with null-safety being added.
In meantime, please try upgrading this package to v1.0.2, it was released yesterday. It should not fix your problems as it seems, but it is worth a try.
@IoanaAlexandru After quick research, #1 should fix your problem, therefore just bump this package to network_image_mock: ^1.0.2 and it should work fine.
I just did and it worked, thank you!
|
2025-04-01T04:35:36.206570
| 2022-02-23T02:13:21
|
1147557483
|
{
"authors": [
"varunsh-coder"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11084",
"repo": "step-security/agent",
"url": "https://github.com/step-security/agent/issues/203"
}
|
gharchive/issue
|
Add annotation if agent fails
Agent can fail is someone adds a wrong domain. In this case developer may not get to know there was a problem. Add annotation to inform developer that the agent had failed.
@h0x0er please take this up. Here is an example: https://github.com/harden-runner-canary/mruby/runs/8139733824?check_suite_focus=true#step:16:22
When one of the domains cannot be resolved in block mode, the agent reverts. We need to write this to the annotation log, so it shows up as an error. Otherwise users do not notice this.
|
2025-04-01T04:35:36.210520
| 2019-09-11T18:10:12
|
492388718
|
{
"authors": [
"azzageee",
"dw-ll",
"mksonkar"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11085",
"repo": "stephane-monnot/react-vertical-timeline",
"url": "https://github.com/stephane-monnot/react-vertical-timeline/issues/36"
}
|
gharchive/issue
|
Overriding border style?
Has anyone figured out how to customize the CSS for elements such as the underlining border & arrow pointing to the timeline? I'd like to blend them in to a different background color that isn't white. Here's an example:
@dw-ll Hey Daniel,
Did you figure out how to modify the border?
You need to edit the boxShadow property in iconStyle. Change the #FFF white color to any color of your choice.
<VerticalTimelineElement
iconStyle={{
boxShadow: "0 0 0 4px #FFF, inset 0 2px 0 rgba(0, 0, 0, .08), 0 3px 0 4px rgba(0, 0, 0, .05)"
}}
</VerticalTimelineElement>
|
2025-04-01T04:35:36.225330
| 2021-01-11T09:58:16
|
783227957
|
{
"authors": [
"ketanpandey8",
"leshchenko1979"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11086",
"repo": "stephenou/fruitionsite",
"url": "https://github.com/stephenou/fruitionsite/issues/72"
}
|
gharchive/issue
|
Issue with tawk.to chat widget
I added tawk.to widget code in CUSTOM_SCRIPT but I cannot see it anywhere on my notion site.
Below is the script code:
<!--Start of Tawk.to Script-->
<script type="text/javascript">
var Tawk_API=Tawk_API||{}, Tawk_LoadStart=new Date();
(function(){
var s1=document.createElement("script"),s0=document.getElementsByTagName("script")[0];
s1.async=true;
s1.src='https://embed.tawk.to/5ffc08d7c31c9117cb6d5a94/1ero852s6';
s1.charset='UTF-8';
s1.setAttribute('crossorigin','*');
s0.parentNode.insertBefore(s1,s0);
})();
</script>
<!--End of Tawk.to Script-->
Thanks in advance.
I have the same issue.
To clarify -- the script is present inside the page, but it doesn't work as it should.
|
2025-04-01T04:35:36.252564
| 2024-11-27T05:51:26
|
2697223345
|
{
"authors": [
"TimerChen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11087",
"repo": "stereolabs/zed-python-api",
"url": "https://github.com/stereolabs/zed-python-api/issues/242"
}
|
gharchive/issue
|
no async_image_retrieval in sl.InitParameters
Preliminary Checks
[X] This issue is not a duplicate. Before opening a new issue, please search existing issues.
[X] This issue is not a question, feature request, or anything other than a bug report directly related to this project.
Description
I found a paramter async_image_retrieval for zed python api in the document, but it does not exist in this repo.
Steps to Reproduce
sl.InitParameters(async_image_retrieval=True)
Expected Result
...
Actual Result
File "pyzed/sl.pyx", line 6606, in pyzed.sl.InitParameters.__cinit__
TypeError: __cinit__() got an unexpected keyword argument 'async_image_retrieval'
ZED Camera model
ZED Mini
Environment
OS: Ubuntu
ZED SDK: 4.2
Anything else?
No response
Update sdk from 4.2 to 4.1 solve this problem
Update sdk from 4.2 to 4.1 solve this problem
|
2025-04-01T04:35:36.258046
| 2022-08-10T23:30:39
|
1335309940
|
{
"authors": [
"JamesCarnley",
"sapien99",
"stefa2k"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11088",
"repo": "stereum-dev/ethereum-node",
"url": "https://github.com/stereum-dev/ethereum-node/issues/646"
}
|
gharchive/issue
|
appImage does not launch due to missing dependency in Ubuntu 22.04
I'm trying to install Stereum 2.0 on a fresh Ubuntu 22.04 LTS instance. I downloaded the Linux install file from stereum.net, marked it as executable, and tried running it:
$ ./Stereum-Launcher-2.0.0-stack-rc4.AppImage
/tmp/.mount_StereuEhlQS4/stereum-launcher: error while loading shared libraries: libnss3.so: cannot open shared object file: No such file or directory
Are there dependencies that need to be installed first? I couldn't find any notes on this.
Will setup a vm with a ubuntu 22-04 desktop and try myself. I heard about electron not providing libnss some years ago, but lets find out.
On my ubuntu 22.04 desktop release libnss is present.
As our AppImage requires libfuse2 and ubuntu dropped it and switched to fuse2 i had to install libfuse2 manually:
apt-add-repository universe
apt-get install libfuse2
Did you maybe use a ubuntu server for the graphical installer, not ubuntu desktop?
Hmm, you're right, sorry. This is a server version. I was assuming I could install Stereum and then connect to it via a browser.
Is it possible to install Stereum via command line?
In principle it should be possible (basically the stereum launcher does excactly that - make you choose your options and then install stereum on a server), but do we have an explicit commandline installer @stefa2k ?
Is it possible to install Stereum via command line?
Theoretically yes, practically no. With Stereum version 2 there are way more commands ran directly by the Launcher. Electron is basically a small web browser and a web app. The launcher connects via SSH and runs all the commands for you, also provides you with SSH tunnels to access Grafana, etc.
Hmm. It seems like the guts of Stereum running on a headless machine and the Web UI running on a Windows desktop or phone are nearly supported by the architecture and perhaps also a use case Stereum would like to support as it would be a good UX for users and make future enterprise support easier to add. Functionality-wise everything seems like it would work fine.
Is this something that might be worked on? I can close this issue if not or leave it open to track it over time.
That's already the case, the web UI runs in Electron.
I fundamentally misunderstood how Stereum works, sorry.
|
2025-04-01T04:35:36.265293
| 2022-03-30T01:40:33
|
1185705555
|
{
"authors": [
"lpk229"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11089",
"repo": "sternie-devops-squad/wishlists",
"url": "https://github.com/sternie-devops-squad/wishlists/issues/26"
}
|
gharchive/issue
|
Add an action route
As a developer
I need to be able to action query results
So that that it can be maintained and edited
Details & Assumptions:
wishlists are already created
results are queried first
Acceptance Criteria:
Given queried results
When a user queries their wishlist
Then those items can also be actioned on
already done in sprint 1
|
2025-04-01T04:35:36.271861
| 2023-11-12T16:44:13
|
1989483128
|
{
"authors": [
"arutonee1",
"christianjuth",
"collindutter",
"stevearc"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11090",
"repo": "stevearc/oil.nvim",
"url": "https://github.com/stevearc/oil.nvim/issues/221"
}
|
gharchive/issue
|
feature request: :wq
Did you check existing requests?
[X] I have searched the existing issues
Describe the feature
:wq currently behaves like :w. If the command was successful (the user presses [O]k), :wq should close Neovim as well.
Provide background
I genuinely might see myself using oil.nvim as a file manager replacement, and this would save a lot of time for quick operations.
Additional details
No response
I added support for this. It's a bit of a hack, but it seems to work well enough. I'm a little concerned that there may be some unintended edge cases, so there's a chance I'll have to tweak this or roll it back.
I'm running into an edge case of this change that causes Neovim to close unexpectedly.
I have the following keymaps:
vim.keymap.set('n', '<leader>w', '<CMD>w<CR>', {})
vim.keymap.set('n', '-', '<CMD>Oil<CR>', {})
To reproduce:
Open Neovim.
Close Neovim with :x.
Open Neovim.
Open Oil with - keymap.
Create a file.
Save the buffer with <leader>w keymap.
Neovim closes.
This happens because commands run through keymaps are not inserted into the history table so the last command is still :x.
Thanks for all the work you do!
I'm using LunarVim and the default w doesn't play nice with oil. However, I found overwriting the default w fixes it.
vim.api.nvim_del_keymap('n', '<leader>w')
vim.api.nvim_set_keymap('n', '<leader>w', ':w<CR>', { noremap = true, silent = true })
|
2025-04-01T04:35:36.298138
| 2019-04-02T00:44:31
|
427972118
|
{
"authors": [
"Wind010",
"filip7sr",
"stevejgordon"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11091",
"repo": "stevejgordon/CorrelationId",
"url": "https://github.com/stevejgordon/CorrelationId/issues/38"
}
|
gharchive/issue
|
Ability to mock the CorrelationContext.
Would like to mock the ICorrelationContextAccessor to provide back a CorrelationContext, so I can access the CorrelationId property. Unfortunately there CorrelationContext does not inherit from an interface and the constructor is internal.
Try this in stead:
var httpClient = new HttpClient(handlerMock.Object);
httpClient.BaseAddress = new Uri("https://localhost:44315/");
var accessor = new CorrelationContextAccessor();
var factory = new CorrelationContextFactory(accessor);
var context = factory.Create(Guid.NewGuid().ToString(), "X-Correlation-ID");
_yourHttpClient = new EmployeePlanningClient(httpClient, accessor);
Added support for this by making the ctor public and adding a sample test project
|
2025-04-01T04:35:36.299621
| 2019-01-15T20:47:32
|
399528423
|
{
"authors": [
"steven-murray"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11092",
"repo": "steven-murray/halomod",
"url": "https://github.com/steven-murray/halomod/issues/9"
}
|
gharchive/issue
|
API documentation
Most API documentation is written, but needs to be completed, and Sphinx/RTD support added
Sphinx support was added in #20, and an example of a "full" module of documentation as well (for the bias.py module).
@zhaotingchen it would be great if you could have a go at extending the range of documentation to other modules.
You can build the docs to have a look at them by
cd docs
make clean
make html
xdg-open _build/html/index.html
|
2025-04-01T04:35:36.302288
| 2023-04-03T19:10:42
|
1652619472
|
{
"authors": [
"imaxisXD",
"ptr-h"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11093",
"repo": "steven-tey/precedent",
"url": "https://github.com/steven-tey/precedent/issues/33"
}
|
gharchive/issue
|
Backdrop blur for sign-in modal on window resize
only an issue when resizing from small to large, the extra portion of screen will not have blur.
could not find a fix myself, will update if I do.
(big thank you for the template)
Also get this effect when moving from mobile sign-in modal to desktop
Also get this bug when on desktop and moving from mobile width to desktop width, with the modal open.
I am not able to replicate this, it will be much more helpful if you can give a video on how to replicate this , Thanks
|
2025-04-01T04:35:36.312549
| 2021-05-05T21:30:37
|
876863444
|
{
"authors": [
"stevens68"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11094",
"repo": "stevens68/twixtbot-ui",
"url": "https://github.com/stevens68/twixtbot-ui/issues/76"
}
|
gharchive/issue
|
installer for windows to simplify installation
some users asked for a simpler installation - e.g. via an installer.
I started to play around with pynsist.
Implemented in PR#78
|
2025-04-01T04:35:36.395232
| 2018-04-08T03:05:35
|
312266668
|
{
"authors": [
"lizhineng1018",
"sthalles"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11095",
"repo": "sthalles/deeplab_v3",
"url": "https://github.com/sthalles/deeplab_v3/issues/13"
}
|
gharchive/issue
|
predict my own picture?
How to predict my own pictures?Looking forward to your guidance
Just run the test.py file, passing your picture. You might need to tweak it.
I have updated the code to use the newest TF version. Also, updated the CreateTfRecords.ipynb, please refer to the README.
|
2025-04-01T04:35:36.404164
| 2017-07-19T20:58:30
|
244172068
|
{
"authors": [
"HaolinWang8",
"bessfernandez",
"catamphetamine",
"iwangulenko",
"stipsan"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11096",
"repo": "stipsan/scroll-into-view-if-needed",
"url": "https://github.com/stipsan/scroll-into-view-if-needed/issues/126"
}
|
gharchive/issue
|
Can i set a offset?
I have some thing floating on the screen like header. I want to set a offset to scroll more and show the content behind the floating header out.
Maybe someone can implement this and take some code from here: https://github.com/yiminghe/dom-scroll-into-view/blob/master/src/dom-scroll-into-view.js
Also wondering this same thing!
Maybe implement similar to:
https://github.com/yiminghe/dom-scroll-into-view/blob/master/src/dom-scroll-into-view.js#L46
Thanks to @iwangulenko and #127 this is now possible π
Hi all.
I've read the readme and it says:
offset
This was always a buggy feature and warned against using in v1 as it might get dropped. It's much safer to use CSS wrapper elements for this kind of thing.
https://github.com/stipsan/scroll-into-view-if-needed#offset
Seems that this feature has been removed in v2.
Shouldn't this issue be reopened then?
There seems to be no way to use this library on websites with position: fixed "floating" headers.
Actually, seems this like not an issue: center should be used (not top) and then it doesn't matter whether there's a floating header or not.
There're issues when the input field itself is higher than the screen height (for example, a big JSON editor) but that can be worked around by setting the input max height less than screen height.
So ignore the above comment.
|
2025-04-01T04:35:36.421193
| 2021-06-10T01:21:33
|
916796137
|
{
"authors": [
"stnolting",
"umarcor"
],
"license": "BSD-3-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11097",
"repo": "stnolting/neorv32",
"url": "https://github.com/stnolting/neorv32/pull/61"
}
|
gharchive/pull-request
|
Reorganise setups
In this PR, Arty and Nexys example setups/scripts/READMEs are updated. Now, when create_project.tcl is sourced, the bitstream is generated straightaway. By the way, the duplicated scripts for Nexys boards are merged into a single one.
Then, subdirs setups/vivado, setups/quartus and setups/radiant are created. The "board" subdirs are reorganised/moved accordingly. The structure we have corresponds to the tools, not necessarily to the boards. For instance, UPduino is supported both in setups/radiant/UPduino_v3 and setups/osflow/UPduino_v3.
A setup/vivado/README.md is added for reducing the duplication in Arty and Nexys READMEs. Both batch and GUI modes are explained.
Very nice! Thank you for cleaning up π
Nice! Note that the Fomu PR is "broken" now. Do not worry. I'll fix (rebase) that :wink:
Nice! Note that the Fomu PR is "broken" now.
Right. But I thought it does not matter since FOMU is still under construction.
I'm sorry if this has caused you extra work. :worried:
Right. But I thought it does not matter since FOMU is still under construction.
I'm sorry if this has caused you extra work. π
No worries! That was expected :wink:
|
2025-04-01T04:35:36.427021
| 2019-08-19T09:28:38
|
482202577
|
{
"authors": [
"valamidev",
"victorarthur"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11099",
"repo": "stockmlbot/TraderCore",
"url": "https://github.com/stockmlbot/TraderCore/issues/1"
}
|
gharchive/issue
|
Incorrect candle data & SQL record duplication on insert
Despite the errors, trading advice is being generated.
Error messages
error: Strategy error Cannot read property 'low' of undefined {"stack":"TypeError: Cannot read property 'low' of undefined\n at Strategy.update (/root/stockml/TraderCore/src/strategies/bb_pure/index.js:41:20)\n at Emulator.update (/root/stockml/TraderCore/src/emulator/emulator.js:72:29)\n at process._tickCallback (internal/process/next_tick.js:68:7)"}
error: SQL errorDuplicate entry '142-BCH/USDT-1566206580000' for key 'strategy_guid' {"code":"ER_DUP_ENTRY","errno":1062,"sqlState":"23000","sqlMessage":"Duplicate entry '142-BCH/USDT-1566206580000' for key 'strategy_guid'","stack":"Error: Duplicate entry '142-BCH/USDT-1566206580000' for key 'strategy_guid'\n at PromisePool.query (/root/stockml/TraderCore/node_modules/mysql2/promise.js:322:22)\n at LiveEmulator.save_advice (/root/stockml/TraderCore/src/emulator/live_emulator.js:168:18)\n at LiveEmulator.single_update (/root/stockml/TraderCore/src/emulator/live_emulator.js:128:16)\n at process._tickCallback (internal/process/next_tick.js:68:7)"} error: Strategy error Cannot read property 'low' of undefined {"stack":"TypeError: Cannot read property 'low' of undefined\n at Strategy.update (/root/stockml/TraderCore/src/strategies/bb_pure/index.js:41:20)\n at Emulator.update (/root/stockml/TraderCore/src/emulator/emulator.js:72:29)\n at LiveEmulator.single_update (/root/stockml/TraderCore/src/emulator/live_emulator.js:114:63)\n at process._tickCallback (internal/process/next_tick.js:68:7)"}
Strat settings
Should be fine from now.
|
2025-04-01T04:35:36.433208
| 2022-05-12T15:55:52
|
1234189727
|
{
"authors": [
"KevinFCormier",
"eemurphy"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11100",
"repo": "stolostron/console-chart",
"url": "https://github.com/stolostron/console-chart/pull/187"
}
|
gharchive/pull-request
|
Update for release 2.6
Signed-off-by: Erin Murphy<EMAIL_ADDRESS>Issue: https://github.com/stolostron/backlog/issues/22278
/lgtm
|
2025-04-01T04:35:36.435885
| 2024-07-29T08:34:51
|
2434800478
|
{
"authors": [
"elgnay",
"haoqing0110"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11101",
"repo": "stolostron/managed-serviceaccount-e2e",
"url": "https://github.com/stolostron/managed-serviceaccount-e2e/pull/21"
}
|
gharchive/pull-request
|
update the csae to enable ManagedServiceAccount feature in MCE
Fix: https://github.com/stolostron/backlog/issues/27649
An alternative fix is to remove this case.
Need to cherry-pick this PR to backplane-2.6, backplane-2.5, and backplane-2.4 branches as well.
/assign @elgnay @zhujian7
/approve
/lgtm
/cherry-pick backplane-2.4
/cherry-pick backplane-2.6
/cherry-pick backplane-2.7
/cherry-pick backplane-2.6
/cherry-pick backplane-2.4
|
2025-04-01T04:35:36.439587
| 2023-05-10T08:13:08
|
1703358171
|
{
"authors": [
"clyang82",
"jlpadilla",
"morvencao"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11102",
"repo": "stolostron/must-gather",
"url": "https://github.com/stolostron/must-gather/pull/146"
}
|
gharchive/pull-request
|
add multicluster global hub resources.
Related Issue: https://issues.redhat.com/browse/ACM-4574
Description of Changes:
Add Multicluster Global Hub resources for must gather scripts.
What resource is being added:
MulticlusterGlobalHub
Kafka
KafkaTopic
PostgresCluster
Also
added namespaces of kafka-operator and postgres-operator in Hub cluster.
added namespace of open-cluster-management-global-hub-system namespace in managed cluster.
Is this a Hub or Managed cluster change?:
Hub Cluster and Managed Cluster
Notes:
/assign @bjoydeep Please help to review. Thanks.
/assign @bjoydeep
/lgtm
|
2025-04-01T04:35:36.442103
| 2022-12-13T06:24:46
|
1493485065
|
{
"authors": [
"SherinV"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11103",
"repo": "stolostron/search-api",
"url": "https://github.com/stolostron/search-api/pull/209"
}
|
gharchive/pull-request
|
update qs package
Signed-off-by: Sherin Varughese<EMAIL_ADDRESS>Related Issue: https://issues.redhat.com/browse/ACM-2281
Description of changes
Added...
/cherrypick release-2.7
|
2025-04-01T04:35:36.446939
| 2024-01-05T22:14:33
|
2068110292
|
{
"authors": [
"chris-shaw-2011",
"jsahoka",
"rzemykers",
"stompro"
],
"license": "BSD-2-Clause",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11104",
"repo": "stompro/pfsense-import-certificate",
"url": "https://github.com/stompro/pfsense-import-certificate/issues/6"
}
|
gharchive/issue
|
Could this script also update the haproxy certificate
Hey,
Thanks for the script, I got it working.
I use the haproxy package in pfsense. Could the script change the haproxy settings so that is also uses the imported certificate?
Thanks!
+1
Probably, I'll try to add that when I get a chance. It normally just takes setting a config value to the cert fingerprint.
@stompro I have a PR to do this here: https://github.com/stompro/pfsense-import-certificate/pull/8
|
2025-04-01T04:35:36.463235
| 2022-07-12T15:16:24
|
1302209887
|
{
"authors": [
"mhmxs",
"nolancon"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11105",
"repo": "storageos/kubectl-storageos",
"url": "https://github.com/storageos/kubectl-storageos/pull/193"
}
|
gharchive/pull-request
|
Auto-generated: Update kuttl e2e test files
Automated changes by create-pull-request GitHub action
@nolancon Kind included in kuttl doesn't support 1.24. What we did in Trousseau are set startKind: false and create kind cluster manually.
@nolancon Kind included in kuttl doesn't support 1.24. What we did in Trousseau are set startKind: false and create kind cluster manually.
Ack.. that's annoying, thanks for the pointer.
|
2025-04-01T04:35:36.484969
| 2024-09-23T09:05:37
|
2542008498
|
{
"authors": [
"Dawntraoz",
"eunjae-lee"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11107",
"repo": "storyblok/field-plugin",
"url": "https://github.com/storyblok/field-plugin/pull/404"
}
|
gharchive/pull-request
|
feat(lib): add event listener for ESC key handler
What?
Whenever a modal is opened inside a custom field plugin, it is impossible to close the modal on escape. This keyboard listener fixes that issue inside and outside the iframe by changing the isModalOpen state and notifying the container.
Why?
JIRA: SHAPE-2151
This is important for people who use a keyboard only to navigate.
How to test? (optional)
Run any template locally
Go to the field-plugin editor at Partner Portal (Enable local dev mode)
Open the modal, put the focus on something inside the iframe
Press ESC, see the modal closing
Working on the unit tests right now, but would love to have some review of the approach also @BibiSebi @eunjae-lee @demetriusfeijoo π
I wonder if this will affect the existing plugins (once they upgrade the field plugin library version)
I wonder if this will affect the existing plugins (once they upgrade the field plugin library version)
The idea was to update one of ours to try out after this PR is merged, but in my mind, that's actually what I tried to do putting it in the lib and not in the templates, I did right then?
just realized that Alba is not included as a contributor :)
@all-contributors please add @Dawntraoz for code.
just realized that Alba is not included as a contributor :)
@all-contributors please add @Dawntraoz for code.
How cool is that bot π€© I didn't know it exists, you just made my day @eunjae-lee
|
2025-04-01T04:35:36.516319
| 2017-06-05T13:38:26
|
233590172
|
{
"authors": [
"shilman",
"xavcz"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11108",
"repo": "storybooks/storybook",
"url": "https://github.com/storybooks/storybook/issues/1196"
}
|
gharchive/issue
|
README images are broken on npm
why
Relatively-linked images broken on https://www.npmjs.com/
what
Update github repository URL to https://github.com/... for all packages. See discussion in https://github.com/SamVerschueren/listr/pull/61
Once published, it should be OK! π
|
2025-04-01T04:35:36.518591
| 2017-10-11T22:59:27
|
264762340
|
{
"authors": [
"elisechant",
"ndelangen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11109",
"repo": "storybooks/storybook",
"url": "https://github.com/storybooks/storybook/issues/2022"
}
|
gharchive/issue
|
Question: what is "module" the second argument of storiesOf?
For this code, what is module? and what can you do with it?
storiesOf('Button', module)
.add('with text', () => (
<Button onClick={action('clicked')}>Hello Button</Button>
))
A colleague at my workplace has suggested that module relates to the ES5 export syntax module.exports.
that's indeed it @elisechant !
Storybook needs a reference to the file/module where your story-code is int to enable hpt-module-replacement.
If you do not supply it, you'd need to refresh your browser for every change you make to your component & story code.
I hope this helps!
|
2025-04-01T04:35:36.520873
| 2017-07-30T23:46:43
|
246617208
|
{
"authors": [
"Hypnosphi",
"ndelangen"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11110",
"repo": "storybooks/storybook",
"url": "https://github.com/storybooks/storybook/pull/1548"
}
|
gharchive/pull-request
|
Use ReactModal for search box
Issue: #871
What I did
Wrapped search box in ReactModal
How to test
Open search with β β§ P. Close it with Esc key or a click outside of the box
Thanks, this looks really good! I want to test it locally before I'll merge.
NOT TO SELF: I should really setup automatic example deployment
|
2025-04-01T04:35:36.528305
| 2012-01-29T13:23:19
|
3011190
|
{
"authors": [
"fwouts",
"stouset"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11111",
"repo": "stouset/twitter_bootstrap_form_for",
"url": "https://github.com/stouset/twitter_bootstrap_form_for/issues/45"
}
|
gharchive/issue
|
does not accept html options
Correct me if I'm wrong, but in standard Rails forms, you can specify HTML options such as a class for a <select> element with the following syntax:
f.select :field, choices, {}, { :class => 'my_css_class' }
With twitter_bootstrap_form_for, I was expecting the syntax to be:
f.select :field, label, choices, {}, { :class => 'my_css_class' }
However that solution does not seem to work. Is that a known issue?
Turns out my server was just ignoring the changes to my file, and rebooting it fixed it. Sorry about this!
No problem. :) Glad to hear it was resolved.
|
2025-04-01T04:35:36.567835
| 2018-09-27T11:25:02
|
364427832
|
{
"authors": [
"njsandilya",
"vroyer"
],
"license": "apache-2.0",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11112",
"repo": "strapdata/elassandra",
"url": "https://github.com/strapdata/elassandra/issues/227"
}
|
gharchive/issue
|
How to migrate entire Data from DataStax Cassandra to Elassandra in aks-cluster
Describe the feature:
Elasticsearch version:
Plugins installed: []
JVM version (java -version):
OS version (uname -a if on a Unix-like system):
Description of the problem including expected versus actual behavior:
Steps to reproduce:
Please include a minimal but complete recreation of the problem, including
(e.g.) index creation, mappings, settings, query etc. The easier you make for
us to reproduce it, the more likely that somebody will take the time to look at it.
Provide logs (if relevant):
Hi,
We are trying to move our data total of 41TB from DataStax cassandra (DSE version 5.1.11) to elassandra.
Brief: We have a existing cassandra setup on DSE where we have 12 streaming nodes with 8 TB each node. Currently we are trying to migrate entire 41TB of data from cassandra to elassandra which is there is aks-cluster in Azure.
We are unable to find any tool or other mechanism to do this migration.
Kindly advice here.
Hi,
Basically, you can migrate from cassandra 3.11.x to elassandra 5.5+ or 6.2+ with your current CQL schema and SSTables (like a backup restore) or replace C* binaries by elassandra ones and restart. You can contact strapdata here to get more details.
Thanks.
|
2025-04-01T04:35:36.594372
| 2017-12-11T09:26:15
|
280946626
|
{
"authors": [
"t-bast"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11113",
"repo": "stratumn/go-indigocore",
"url": "https://github.com/stratumn/go-indigocore/issues/283"
}
|
gharchive/issue
|
tmpop: include Tendermint signatures in Proof and verify them
See https://github.com/tendermint/tendermint/issues/921
See details about the lite package here: https://godoc.org/github.com/tendermint/tendermint/lite
Or we can investigate using Block.Evidence or Block.LastCommit directly ourselves
|
2025-04-01T04:35:36.603954
| 2023-01-13T16:20:39
|
1532574438
|
{
"authors": [
"botberry",
"sairevanth0007"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11114",
"repo": "strawberry-graphql/strawberry",
"url": "https://github.com/strawberry-graphql/strawberry/pull/2456"
}
|
gharchive/pull-request
|
updated .pre-commit-config.yaml
Description
I have added the blacken-docs hook to .pre-commit-config.yaml file as it was posted as an good first issue.
Types of Changes
[ ] Core
[ ] Bugfix
[ ] New feature
[x] Enhancement/optimization
[ ] Documentation
Issues Fixed or Closed by This PR
Checklist
[x] My code follows the code style of this project.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
[x] I have read the CONTRIBUTING document.
[ ] I have added tests to cover my changes.
[ ] I have tested the changes and verified that they work and don't break anything (as well as I can manage).
@patrick91 I am doing the correct changes? Do I need to change anything?
Thanks for contributing to Strawberry! π You've been invited to join
the Strawberry GraphQL organisation π
You can also request a free sticker by filling this form: https://forms.gle/dmnfQUPoY5gZbVT67
And don't forget to join our discord server: https://strawberry.rocks/discord π₯
I've made some small updates to get CI green, but other than that it looks great! thanks!
Thank you.
|
2025-04-01T04:35:36.610227
| 2024-02-06T09:34:36
|
2120338674
|
{
"authors": [
"XChikuX",
"botberry",
"patrick91"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11115",
"repo": "strawberry-graphql/strawberry",
"url": "https://github.com/strawberry-graphql/strawberry/pull/3374"
}
|
gharchive/pull-request
|
Upgrade python-multipart to >=0.0.7
There is a Re-DoS bug in older versions of multipart uploads, that was recently fixed
Description
Types of Changes
[x] Core
[x] Bugfix
[ ] New feature
[ ] Enhancement/optimization
[ ] Documentation
Issues Fixed or Closed by This PR
Checklist
[x] My code follows the code style of this project.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
[ ] I have read the CONTRIBUTING document.
[ ] I have added tests to cover my changes.
[ ] I have tested the changes and verified that they work and don't break anything (as well as I can manage).
Hi, thanks for contributing to Strawberry π!
We noticed that this PR is missing a RELEASE.md file. We use that to automatically do releases here on GitHub and, most importantly, to PyPI!
So as soon as this PR is merged, a release will be made π.
Here's an example of RELEASE.md:
Release type: patch
Description of the changes, ideally with some examples, if adding a new feature.
Release type can be one of patch, minor or major. We use semver, so make sure to pick the appropriate type. If in doubt feel free to ask :)
Here's the tweet text:
π Release (next) is out! Thanks to @Seetku for the PR π
Get it here π https://beta.strawberry.rocks/release/(next)
perfect! let's also add a release note for this π
Made a new PR. Closing this one.
|
2025-04-01T04:35:36.639165
| 2020-01-07T17:06:45
|
546400758
|
{
"authors": [
"adesurirey",
"streamich"
],
"license": "Unlicense",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11116",
"repo": "streamich/react-use",
"url": "https://github.com/streamich/react-use/issues/863"
}
|
gharchive/issue
|
useUpdateEffect not calling returned function on unmount
What is the current behavior?
Does not call the returned function on unmount
Steps to reproduce it and if possible a minimal demo of the problem
https://codesandbox.io/s/gifted-hopper-tb9gn
What is the expected behavior?
Should call returned function when unmounting
A little about versions:
OS: macOS 10.15.2
Browser (vendor and version): Google Chrome 79.0.3945.88
React: 16.12.0
react-use: 13.14.0
Did this worked in the previous package version? idk
:tada: This issue has been resolved in version 13.14.1 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
|
2025-04-01T04:35:36.674179
| 2022-08-19T02:22:52
|
1343848202
|
{
"authors": [
"KyleZhang0536",
"wolfboys"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11118",
"repo": "streamxhub/streamx",
"url": "https://github.com/streamxhub/streamx/pull/1410"
}
|
gharchive/pull-request
|
[Bug] should use fromYamlFile function to read conf file in flink 1.15 shims
https://github.com/streamxhub/streamx/issues/1409
Problem Summary:
What is changed and how it works?
Proposal: xxx
Feel free to ping committers for the review!
Contribution Checklist
Make sure that the pull request corresponds to a GITHUB issue.
Name the pull request in the form "[Feature] [component] Title of the pull request", where Feature can be replaced by Hotfix, Bug, etc.
Minor fixes should be named following this pattern: [hotfix] [docs] Fix typo in README.md doc.
Purpose of this pull request
Thank you for your contribution. I'll review it later
@wolfboys PTAL
LGTM
Thanks for your contribution, we are looking forward to your next contribution.ππππ
|
2025-04-01T04:35:36.677950
| 2023-05-18T20:46:03
|
1716239755
|
{
"authors": [
"VaclavElias"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11119",
"repo": "stride3d/stride-website",
"url": "https://github.com/stride3d/stride-website/issues/95"
}
|
gharchive/issue
|
Search - When accessed directly the spinner is visible
We don't link search page directly, but would be good to not show the spinner if we access the page directly.
Done
|
2025-04-01T04:35:36.681707
| 2024-10-09T10:11:34
|
2575457632
|
{
"authors": [
"Eideren",
"kopffarben"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11120",
"repo": "stride3d/stride",
"url": "https://github.com/stride3d/stride/pull/2482"
}
|
gharchive/pull-request
|
Small fix in CommandList.Direct3D.cs
PR Details
Small fix in CommandList.Direct3D.cs
CommandList.DrawInstanced(Buffer argumentsBuffer, int alignedByteOffsetForArgs = 0) call DrawIndexedInstancedIndirect instead of DrawInstancedIndirect
public void DrawInstanced(Buffer argumentsBuffer, int alignedByteOffsetForArgs = 0)
{
if (argumentsBuffer == null) throw new ArgumentNullException("argumentsBuffer");
PrepareDraw();
// was DrawIndexedInstancedIndirect( ... ) but should be DrawInstancedIndirect( ... )
NativeDeviceContext.DrawInstancedIndirect(argumentsBuffer.NativeBuffer, alignedByteOffsetForArgs);
GraphicsDevice.FrameDrawCalls++;
}
Types of changes
[ ] Docs change / refactoring / dependency upgrade
[x] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
Checklist
[ ] My change requires a change to the documentation.
[ ] I have added tests to cover my changes.
[x] All new and existing tests passed.
[x] I have built and run the editor to try this change out.
@dotnet-policy-service agree
Makes sense, thanks for the contribution @kopffarben !
|
2025-04-01T04:35:36.746999
| 2021-06-10T09:32:59
|
917129655
|
{
"authors": [
"paulbryan",
"thorsten-stripe"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11121",
"repo": "stripe/stripe-react-native",
"url": "https://github.com/stripe/stripe-react-native/issues/325"
}
|
gharchive/issue
|
Cannot call confirmSetupIntent with a setup intent created elsewhere
Describe the bug
Cannot confirm a setupIntent created outside the app
To Reproduce
Steps to reproduce the behavior:
Create a setupIntent with payment_method and usage:off_session outside the app
use the client_secret to call confirmSetupIntent
error returned = {"code":"Failed","message":"You must provide card details"}
Expected behavior
The setup intent to be confirmed
Smartphone:
Device: iphone 12 Pro Max (emulated)
OS: iOS14.4
SDK Verion: 0.1.4
Additional context
sample setup intent
seti_1J0jtQCOnQnnH0At0rVt4Wjg_secret_Je1xyh6q3Wtxpignlr2LltzDf1HvDiC
Invariant Violation: No callback found with cbID 17727 and callID 8863 for StripeSdk.confirmSetupIntent - most likely the callback was already invoked. Args: '[{"code":"Unknown","message":"Unhandled error occured","domain":"ABI41_0_0RCTErrorDomain","userInfo":null,"nativeStackIOS":["0 Exponent 0x0000000107919632 ABI41_0_0RCTJSErrorFromCodeMessageAndNSError + 79","1 Exponent 0x00000001078de01a _50-[ABI41_0_0RCTModuleMethod processMet...(truncated)...","2 Exponent 0x00000001079e420e $sSo8NSStringCSgACSo7NSErrorCSgIeyByyy_SS...(truncated)...","3 Exponent 0x00000001079ecf18 $s28ABI41_0_0stripe_react_native9StripeSd...(truncated)...","4 Exponent 0x00000001079e6749 $s28ABI41_0_0stripe_react_native9StripeSd...(truncated)...","5 CoreFoundation 0x00000001178d62fc invoking + 140","6 CoreFoundation 0x00000001178d37b6 -[NSInvocation invoke] + 303","7 CoreFoundation 0x00000001178d3a47 -[NSInvocation invokeWithTarget:] + 70","8 Exponent 0x00000001078de824 -[ABI41_0_0RCTModuleMethod invokeWithBrid...(truncated)...","9 Exponent 0x00000001078e0c72 _ZN17ABI41_0_0facebook14ABI41_0_0ReactL11...(truncated)...","10 Exponent 0x00000001078e09ed ___ZN17ABI41_0_0facebook14ABI41_0_0React2...(truncated)...","11 libdispatch.dylib 0x000000011d4f332f _dispatch_call_block_and_release + 12","12 libdispatch.dylib 0x000000011d4f4508 _dispatch_client_callout + 8","13 libdispatch.dylib 0x000000011d4fa3f9 _dispatch_lane_serial_drain + 715","14 libdispatch.dylib 0x000000011d4faf74 _dispatch_lane_invoke + 403","15 libdispatch.dylib 0x000000011d505577 _dispatch_workloop_worker_thread + 782","16 libsystem_pthread.dylib 0x000000011d9b945d _pthread_wqthread + 314","17 libsystem_pthread.dylib 0x000000011d9b842f start_wqthread + 15"]}]'
Can you explain your use case, please? Why would you create a SetupIntent if you already have a payment method for the customer? The confirmSetupIntent method has to be used in conjunction with the CardField component: https://stripe.com/docs/payments/save-and-reuse?platform=react-native
We want to move to paymentIntents but have some legacy code that is using the old charges api to charge the customers default source (we are upgrading this legacy code but it will take time).
We tried adding the card info using the stripe cardfield but the legacy code did not charge because it is looking for the default_source and adding the card as a paymentMethod the default is stored in invoice_settings.
What we tried was send the card info to our api, which creates a cardToken, add it to a cardSource, set the customers default_source, then create a setupIntent attach the cardSource and return the setupIntent to the front end if the setupIntent needed action. What we need is a way to trigger the 3dsecure check on the front end if the setupIntent status is requires_action.
I'm afraid that isn't something we're planning to support.
Can your CardField element create a card token rather than a payment method for the setupIntent?
|
2025-04-01T04:35:36.772872
| 2023-10-03T16:32:32
|
1924521659
|
{
"authors": [
"UrbanChrisy",
"amansoomro062",
"nazli-stripe"
],
"license": "MIT",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11122",
"repo": "stripe/stripe-terminal-react-native",
"url": "https://github.com/stripe/stripe-terminal-react-native/issues/549"
}
|
gharchive/issue
|
SDK ^0.0.1-beta.13: collectPaymentMethod Errors on Optional Parameters after iOS Tap to Pay Update
Describe the bug
After we've updated SDK to ^0.0.1-beta.13 to support IOS Tap to Pay, we're having issue with collectPaymentMethod.
Previously we had code like this
let paymentCollect = await collectPaymentMethod({
paymentIntentId: paymentIntent.id
});
AND
let paymentCollect = await collectPaymentMethod({
paymentIntentId: paymentIntent.id,
skipTipping: true,
tipEligibleAmount: 1500,
updatePaymentIntent: false,
});
We are aware that using tipEligibleAmount: 1500 is returning error: Error collecting payment: {"error": {"code": "FeatureNotAvailableWithConnectedReader", "message": "This feature is currently not available for the selected reader."}, "paymentIntent": undefined} But removing the tipEligibleAmount: 1500 throws error:
Error collecting payment: {"error": {"code": "InvalidRequiredParameter", "message": "A required parameter was invalid or missing."}, "paymentIntent": undefined}
Even though it is optional as defined below:
export declare type CollectPaymentMethodParams = {
paymentIntentId: string;
skipTipping?: boolean;
tipEligibleAmount?: number;
updatePaymentIntent?: boolean;
};
To Reproduce
Steps to reproduce the behavior:
Connect Tap 2 Pay reader on ios using
const { reader, error } = await connectLocalMobileReader({
reader: selectedReader,
locationId: locationId,
});
Create payment intent setup
let i = await createPaymentIntentFunction(token, body);
const { paymentIntent, error } = await retrievePaymentIntent(clientSecret);
if (error) {
// Placeholder for handling exception
console.log("ERROR", error);
return;
}
const paymentIntentIdDict = { paymentIntentId: paymentIntent.id };
let paymentCollect = await collectPaymentMethod({
paymentIntentId: paymentIntent.id,
skipTipping: true,
tipEligibleAmount: 1500,
updatePaymentIntent: false,
});
if (paymentCollect.error) {
console.log("Error collecting payment:", paymentCollect);
}
You should be able to see the error
If you remove other parameters, It'll still throw error.
Expected behavior
The collect method should ignore if we are not passing other parameters which are not required and the payment should progress.
Stripe Terminal React Native SDK version
^0.0.1-beta.13
Smartphone (please complete the following information):
Device: iPhone 12 pro and iPhone 15 pro
OS: 16.6.1 and 17.0.2
I am getting the same issue, did you manage to find a work around.
Seems like tap to pay iOS is not ready for production use yet.
Hey @amansoomro062 we got in contact with Stripe and they where able to debug the issue for us which might be the same with you.
Because we had created a payment intent on our backend for a stripe connect account you need to supply the account id of said account in the terminal connection using the onBehalfOf field.
const {reader, error} = await connectLocalMobileReader({
reader: selectedReader,
locationId: 'tml_...',
onBehalfOf: 'acct_...',
});
hey all, @UrbanChrisy is right on the cause of the issue. @amansoomro062 if that doesn't fix it for you let us know. The error message is unfortunately not helpful in this case, but it will be fixed in a future release.
Hi @nazli-stripe
I've tried above, still the same:
const { reader, error } = await connectLocalMobileReader({
reader: selectedReader,
locationId: userData?.terminalLocationId,
onBehalfOf: userData?.id
});
Error collecting payment: {"error": {"code": "FeatureNotAvailableWithConnectedReader", "message": "This feature is currently not available for the selected reader."},}
Thanks @UrbanChrisy and @nazli-stripe
I was able to sort the issue by removing
skipTipping: true,
tipEligibleAmount: 1500,
updatePaymentIntent: false,
and adding
onBehalfOf: 'acct_...',
Appreciate the support π
|
2025-04-01T04:35:36.774678
| 2018-04-06T23:27:15
|
312147817
|
{
"authors": [
"aubrey-stripe",
"sjung-stripe"
],
"license": "Apache-2.0",
"license_source": "github-api",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11123",
"repo": "stripe/terraform-provider-signalform",
"url": "https://github.com/stripe/terraform-provider-signalform/pull/2"
}
|
gharchive/pull-request
|
Adding support for TF structure for arbitrary many watermarks
Necessary for converting chart markers from datadog to signalfx.
Please review and/or comment.
oh you want to merge into the branch called stripe/master, not master. i wanted to leave our master to track yelp's master
Okay I moved the base of this PR to stripe/master.
|
2025-04-01T04:35:36.783426
| 2016-08-03T21:27:19
|
169238366
|
{
"authors": [
"AGausmann",
"plato2000",
"rhauch"
],
"license": "mit",
"license_source": "bigquery",
"license_type": "permissive",
"provenance": "gharchive-dolma-0004.json.gz:11124",
"repo": "strongback/strongback-java",
"url": "https://github.com/strongback/strongback-java/issues/80"
}
|
gharchive/issue
|
NavX support for Strongback
Our team uses the NavX MXP device. It consists of a suite of sensors including a 3 axis gyroscope, a 3 axis accelerometer, a magnetometer, and an altimeter. It has its own library that makes getting this data from the sensor easy. I think the library also allows for some fancy calculations for distance traveled and some other stuff.
I would like to add support for this sensor to Strongback. However, I'm not sure of the best way to do this. I think the NavX is popular enough that it should be supported by Strongback (complete with mock sensor for unit tests), but not popular enough to force all Strongback users to download the library for it. Also, I don't know whether the best idea is to consider the NavX as separate sensors for the accelerometer and the gyroscope or to consider it as its own class with fields for the individual sensors. The link above has some examples of how the library is used if that helps in making the decision.
I would be glad to submit a pull request with support for the NavX if someone would tell me the best method of adding it to fit with Strongback's system.
Our team uses the NavX MXP device. It consists of a suite of sensors including a 3 axis gyroscope, a 3 axis accelerometer, a magnetometer, and an altimeter.
I have no experience with this device, so the big question you have to answer is: what should the API look like for this device? Is it as simple as a composition of a 3-axis gyroscope, 3-axis accelerometer, compass, and distance sensor (altimeter)? If so, then we probably just want to create an interface that has getter methods for the other existing components. Not sure what that might be named, or where it should live, but I'm fine with proceeding (at least for now) with it in the org.strongback.components package.
Then, an implementation of this new interface needs to be added to the hardware package, and it could be named HardwareNavX and should implement the new interface. Then, a new factory method needs to be added to the Hardware class.
It has its own library that makes getting this data from the sensor easy.
That's okay. The library appears to be licensed under MIT open source license, so that means we can distribute it. It would be easier for Strongback to include the library so that it works if people want to use it, but if they don't they can always remove it (if they really care about the small space).
For now, the JARs and source JARs should be included in the libs directory of our codebase. We may want to eventually change the build to be able to download the libraries from the official location, but we can worry about that later.
I think the library also allows for some fancy calculations for distance traveled and some other stuff.
Hopefully the interface for this device remains pretty simple, and any other functionality (like distance travelled) should be exposed through existing components (e.g., DistanceSensor).
This should be a pretty exciting addition, but it may take multiple iterations to settle on something that is generic and that reuses much of the existing API. Either way, how about trying to fit it in as I described above, submitting a pull request, and then we'll review it and see where it needs to go?
@plato2000 any progress on a contribution to add support for the NavX?
I would be interested in implementing this in Strongback if no progress has been made. @plato2000 ?
@AGausmann Unfortunately, I have not had time with school and everything to make progress on this. I probably won't have time before the Christmas break just before the season. I have not made any progress on this yet. Feel free to implement it yourself if you can start by then.
Yep, our team has one. I'll get started then
I have a pull request (targeted for the Strongback 2017.1.0 version) that isn't quite ready that will add 3rd party libraries for CTRE TalonSRX library and the navX-MXP library. The latter version that is compatible with the 2017 WPILib for Java will be officially released on 1/6/2016, so the initial version of Strongback 2017.1.0 will add both libraries and the new Strongback CLI will properly install them into the WPILib's user library directory. This work should build upon that work.
The master branch now includes the navX-MXP library (see #88), the build already includes it in the artifacts, and the Strongback CLI will automatically install the library as a WPILib user library. All that remains is to add the APIs to Strongback to make use of the navX-MXP library.
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.