id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
2225079238 | chore(main): release 0.7.0
:robot: I have created a release beep boop
0.7.0 (2024-04-04)
Features
对影响整体布局紧凑度的内外边距值进行 CSS 变量提取,并增加 UnoCSS 支持 (8e1aad7)
Bug Fixes
修复 AppStore 中没有正确验证配置文件与没有 resolve 的问题 (82b0523)
修复 UserStore 没有正确地抛出异常导致 Token 过期后系统加载异常的问题 (e9db47d)
修复登录表单有默认值时 label 状态异常的问题,增加账户 input 的自动聚焦 (05f2ea8)
修复系统根节点的颜色错误问题,调整 rem 基值 (e1be34f)
补充路由守卫中遗漏的错误处理 (4d0d5ef)
This PR was generated with Release Please. See documentation.
:robot: Release is at https://github.com/thelastcookies/WEB-BASE/releases/tag/v0.7.0 :sunflower:
| gharchive/pull-request | 2024-04-04T10:18:49 | 2025-04-01T06:45:59.823358 | {
"authors": [
"thelastcookies"
],
"repo": "thelastcookies/WEB-BASE",
"url": "https://github.com/thelastcookies/WEB-BASE/pull/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2278598480 | Board Request - SparkleIoT ESP32-C3F
As someone pointed out in the Tasmota repo (https://github.com/arendst/Tasmota/discussions/21326), I'm adding a request here to add SparkleIoT ESP32-C3F (and should be the same with other ESP32-C3F boards). Board and its information is available on their Amazon page (https://www.amazon.com/gp/product/B0BNWTCPV8/).
Picture of Front
Pinout from front
The board is now available
Thanks
| gharchive/issue | 2024-05-04T00:08:19 | 2025-04-01T06:45:59.826203 | {
"authors": [
"lightmaster",
"thelastoutpostworkshop"
],
"repo": "thelastoutpostworkshop/gpio_viewer",
"url": "https://github.com/thelastoutpostworkshop/gpio_viewer/issues/115",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2708955 | Missing deps on test sections?
Hi,
starting from a fresh install, I fail to install oasis:
philippe@robby:~$ echo $OCAMLPATH
/home/philippe/.odb/lib
philippe@robby:~$ rm -rf ./.odb
philippe@robby:~$ ocaml odb.ml oasis
Installing oasis
Installing odn
Installing type-conv
Deps for type-conv satisfied
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 50638 100 50638 0 0 114k 0 --:--:-- --:--:-- --:--:-- 114k
type-conv-3.0.4/.hg_archival.txt
[SNIP]
Installed /home/philippe/.odb/lib/type-conv/META
Successfully installed type-conv
Deps for odn satisfied
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 57339 100 57339 0 0 64935 0 --:--:-- --:--:-- --:--:-- 98k
ocaml-data-notation-0.0.7/
ocaml-data-notation-0.0.7/tests/
ocaml-data-notation-0.0.7/tests/TestConf.ml.ab
ocaml-data-notation-0.0.7/tests/test.ml
ocaml-data-notation-0.0.7/tests/data/
ocaml-data-notation-0.0.7/tests/data/variant.ml
ocaml-data-notation-0.0.7/tests/data/oasis-examples/
ocaml-data-notation-0.0.7/tests/data/oasis-examples/OASISTypes.ml
ocaml-data-notation-0.0.7/tests/data/oasis-examples/main.ml
ocaml-data-notation-0.0.7/tests/data/oasis-examples/PropList.ml
ocaml-data-notation-0.0.7/tests/data/tuples.ml
ocaml-data-notation-0.0.7/tests/data/polyvariants.ml
ocaml-data-notation-0.0.7/tests/TestCommon.ml
ocaml-data-notation-0.0.7/src/
ocaml-data-notation-0.0.7/src/pa_noodn.ml
ocaml-data-notation-0.0.7/src/odn.odocl
ocaml-data-notation-0.0.7/src/pa_odn.mllib
ocaml-data-notation-0.0.7/src/pa_noodn.mllib
ocaml-data-notation-0.0.7/src/odn.mllib
ocaml-data-notation-0.0.7/src/META
ocaml-data-notation-0.0.7/src/pa_odn.ml
ocaml-data-notation-0.0.7/src/ODN.ml
ocaml-data-notation-0.0.7/CHANGES.txt
ocaml-data-notation-0.0.7/AUTHORS.txt
ocaml-data-notation-0.0.7/setup.ml
ocaml-data-notation-0.0.7/Makefile
ocaml-data-notation-0.0.7/README.txt
ocaml-data-notation-0.0.7/_header
ocaml-data-notation-0.0.7/_headache.config
ocaml-data-notation-0.0.7/INSTALL.txt
ocaml-data-notation-0.0.7/COPYING.txt
ocaml-data-notation-0.0.7/configure
ocaml-data-notation-0.0.7/_tags
ocaml-data-notation-0.0.7/.boring
ocaml-data-notation-0.0.7/myocamlbuild.ml
ocaml-data-notation-0.0.7/_oasis
I: Running command '/home/philippe/usr/godi/bin/ocamlc.opt -config > '/tmp/oasis-f83b31.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %v findlib > '/tmp/oasis-d4a71c.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d type-conv > '/tmp/oasis-7810b4.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d camlp4.lib > '/tmp/oasis-76b18b.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d camlp4.quotations.o > '/tmp/oasis-876ddc.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d type-conv > '/tmp/oasis-d11b4b.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d type-conv > '/tmp/oasis-dca1ad.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %v type-conv > '/tmp/oasis-a8b1f6.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d camlp4.lib > '/tmp/oasis-c3858f.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d camlp4.lib > '/tmp/oasis-74d3b2.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d camlp4.quotations.o > '/tmp/oasis-ba71b8.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d camlp4.quotations.o > '/tmp/oasis-99f95f.txt''
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d oUnit > '/tmp/oasis-f3fae8.txt''
ocamlfind: Package `oUnit' not found
W: Field 'pkg_ounit' is not set: Command '/home/philippe/usr/godi/bin/ocamlfind query -format %d oUnit > '/tmp/oasis-f3fae8.txt'' terminated with error code 2
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d fileutils > '/tmp/oasis-737a7d.txt''
ocamlfind: Package `fileutils' not found
W: Field 'pkg_fileutils' is not set: Command '/home/philippe/usr/godi/bin/ocamlfind query -format %d fileutils > '/tmp/oasis-737a7d.txt'' terminated with error code 2
I: Running command '/home/philippe/usr/godi/bin/ocamlfind query -format %d str > '/tmp/oasis-785e2b.txt''
E: Cannot find findlib package fileutils (>= 0.4.0)
E: Cannot find findlib package oUnit (>= 1.1.0)
E: Failure("2 configuration errors")
Exception: Failure "Could not configure odn".
It seems that odb do not add deps from the test section (fileutils and oUnit), while the configure script of odn requires them. This issue is related to #25, in that there is a need to access test dependencies with oasis-db.
Easy workaround of course, first install the missing dependencies, by invoking ocaml odb.ml fileutils directly. I'm posting it here for the record (hope I'm not duplicating a previous bug report ...)
Sadly, this is a known and undocumented problem. The problem is just in the ocaml-data-notation package. Gildor patched 0.0.6 quickly, and must have used his development version of oasis to generate its setup.ml file. This file tries to configure dependencies needed for testing, and fails because it can't find them. Oasis-db only lists as dependencies those packages needed to build and install, not to test. There's two solutions, the short-term fix-odn solution and the long term fix-oasis-db solution. I'm trying for the long term solution, but need gildor's help.
OK, I totally mismatch my sentence. "The package had NOT been released in setup-dev mode". The rest of the comment stands. There is a bug in dependency calculation with flags. Please fill a bug against oasis. The quick solution is to have "Default: false" for flag Test in odn.
Bug reported, thanks!
| gharchive/issue | 2012-01-03T10:19:15 | 2025-04-01T06:45:59.832921 | {
"authors": [
"gildor478",
"pveber",
"thelema"
],
"repo": "thelema/odb",
"url": "https://github.com/thelema/odb/issues/27",
"license": "WTFPL",
"license_type": "permissive",
"license_source": "github-api"
} |
181568077 | add birthdate to user object, try to grab it from current user reques…
…t, add sample birthdate to test (confirmed format with Spotifys actual API response)
Hi @thelinmichael, this PR contains a simple addition of the 'birthdate' field to the user object, which can be returned on the "Get Current User" request under certain conditions.
References issue #67 I wrote up.
Thanks for your project! Made integration a breeze.
| gharchive/pull-request | 2016-10-07T01:47:50 | 2025-04-01T06:45:59.838490 | {
"authors": [
"samiruppaluru"
],
"repo": "thelinmichael/spotify-web-api-java",
"url": "https://github.com/thelinmichael/spotify-web-api-java/pull/68",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
831454173 | Cannot get certain playlist's tracks
I'm doing the following:
Getting my userId with getMe()
Getting my playlists with getUserPlaylists(userId)
For each playlist I'm running getPlaylistTracks(userId, plId) where plId is the id of each playlist
It's at step 3 where some playlists throw this error:
WebapiRegularError: An error occurred while communicating with Spotify's Web API.
Details: Invalid playlist Id.
I've been using the above flow with the older version of spotify-web-api-node but now with the latest version, it fails. What gives?
I was wondering if the market parameter could be a problem but that appears to have been closed: https://github.com/plamere/spotipy/pull/526
I'm also having a similar issue but trying to get a show's episodes. Have you found a fix yet?
I'm also having a similar issue but trying to get a show's episodes. Have you found a fix yet?
You try using version 4.0.0 or older of this library.
I use 4.0.0 and this functionality works for me. The API works a bit differently though so watch out for that.
I'm also having a similar issue but trying to get a show's episodes. Have you found a fix yet?
You try using version 4.0.0 or older of this library.
I use 4.0.0 and this functionality works for me. The API works a bit differently though so watch out for that.
I found a different API to use which works better for me
| gharchive/issue | 2021-03-15T06:07:35 | 2025-04-01T06:45:59.843612 | {
"authors": [
"Razor5284",
"jasjeev4"
],
"repo": "thelinmichael/spotify-web-api-node",
"url": "https://github.com/thelinmichael/spotify-web-api-node/issues/377",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1433366079 | 🛑 Matters (matters.news) is down
In e5e76b2, Matters (matters.news) (https://matters.news/) was down:
HTTP code: 403
Response time: 116 ms
Resolved: Matters (matters.news) is back up in eeb6b45.
| gharchive/issue | 2022-11-02T15:36:51 | 2025-04-01T06:45:59.856986 | {
"authors": [
"robertu7"
],
"repo": "thematters/uptime",
"url": "https://github.com/thematters/uptime/issues/174",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
39610390 | Updated tmhOAuth.php to support RSA-SHA1 Signing.
I chose this library to handle OAuth authentication for a Xero Integration project, which worked great until using it for 'Private' applications (http://developer.xero.com/documentation/getting-started/private-applications/). Xero uses HMAC-SHA1 for public/partner applications and the RSA-SHA1 signature method for Private applications, which tmhOAuth doesn't support.
As Xero usage is growing rapidly, it'd be nice to incorporate these features into the main library.
Hi Matt, could you please review these changes? Thanks.
i was looking at this. can you also add a test to https://github.com/themattharris/tmhOAuthTest that i can run against this code. once i can run the tests i'll work on integration.
| gharchive/pull-request | 2014-08-06T10:59:17 | 2025-04-01T06:45:59.859595 | {
"authors": [
"joshatdf",
"themattharris"
],
"repo": "themattharris/tmhOAuth",
"url": "https://github.com/themattharris/tmhOAuth/pull/179",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
145788382 | Switch to new recommended syntax
Composer now recommends using two || instead of a single |.
Thank you :)
I use your package https://github.com/vinkla/hashids in almost every project.
Glad to hear that :)
| gharchive/pull-request | 2016-04-04T20:00:18 | 2025-04-01T06:45:59.871739 | {
"authors": [
"themsaid",
"vinkla"
],
"repo": "themsaid/laravel-langman",
"url": "https://github.com/themsaid/laravel-langman/pull/11",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2716726191 | First react frontend
I haven't made it perfect obviously. There are many things to fix. It's just a start.
If you have node and npm installed, you should be able to run it locally with:
npm install
npm run dev
Example:
https://github.com/user-attachments/assets/1c2b350b-6bb7-4e09-a73c-c9968e548047
Cool!!
| gharchive/pull-request | 2024-12-04T06:56:32 | 2025-04-01T06:45:59.875758 | {
"authors": [
"andraghetti",
"theo2021"
],
"repo": "theo2021/The-Good-AI-Lab",
"url": "https://github.com/theo2021/The-Good-AI-Lab/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2280684396 | Fixed spelling in Exceptions
This Fixes #113
Thanks @ATC-4K !
| gharchive/pull-request | 2024-05-06T11:43:03 | 2025-04-01T06:45:59.894196 | {
"authors": [
"ATC-4K",
"MaximeThoonsen"
],
"repo": "theodo-group/LLPhant",
"url": "https://github.com/theodo-group/LLPhant/pull/114",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
869667626 | Adding a lot of aliases
With the new 1.6 functionality of adding alias, we could add new alias to a lot of types, some examples:
const ALotOfNumbers = Vector{<:Number}
const NumberNumber = Tuple{Number, Number}
const EverythingInExistence = Any
const PythonObject = Dict{Any, Any}
const TLDR = Vector{String}
Hi @longemen3000! Adding aliases for type is available even from 1.0, I believe, the difference is that they are now displayed in error messages as well.
However using other sentences for these types is almost too explicit :D, we're more trying to find emoji/kaomoji alternatives!
0️⃣1️⃣2️⃣3️⃣4️⃣5️⃣
🐍Object
📜😳
| gharchive/issue | 2021-04-28T08:26:06 | 2025-04-01T06:45:59.896187 | {
"authors": [
"FPGro",
"longemen3000",
"theogf"
],
"repo": "theogf/WatchJuliaBurn.jl",
"url": "https://github.com/theogf/WatchJuliaBurn.jl/issues/26",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1475650833 | Update form selector on sign in page
The sign in page seems to have been updated and uses a different ID than it had at the time this script was written.
Country I use with: United Arab Emirates
Thanks for the PR - I just checked as well and I see the new form id.
| gharchive/pull-request | 2022-12-05T04:35:48 | 2025-04-01T06:45:59.903227 | {
"authors": [
"aliidrees",
"theoomoregbee"
],
"repo": "theoomoregbee/US-visa-appointment-notifier",
"url": "https://github.com/theoomoregbee/US-visa-appointment-notifier/pull/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
702405377 | Process options in AbstractProvider->getAccessToken().
Fixes #774
The test failures seem unrelated to these changes.
The new test failures are because the mocked method $grant->prepareRequestParameters($params, $options) is receiving new values in $options that the test does not expect.
A few issues with this PR:
There should be a test that shows the problem reported in #774 that this PR fixes
I do not believe this change is related to the problem reported in #774
By calling $this->getAuthorizationParameters() here, you're merging authorization request parameters onto the access token request, and this could potentially leak credentials into the access token request, which we don't want to happen
For these reasons, I'm closing this PR.
Feel free to open a new PR, if you think there is still a change that needs to be made here. If so, please provide a test that shows the problem you're solving, and please don't call getAuthorizationParameters() within getAccessToken(), since there's a security risk in doing so.
Thanks!
| gharchive/pull-request | 2020-09-16T02:50:48 | 2025-04-01T06:45:59.910519 | {
"authors": [
"imclean557",
"ramsey"
],
"repo": "thephpleague/oauth2-client",
"url": "https://github.com/thephpleague/oauth2-client/pull/856",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
453844381 | Signed integer overflow causes undefined behaviour
Signed integer overflow in Rust is supposed to wrap, except in debug builds where it panics to warn developers of possibly unintentional overflow. However, when the following code is compiled into C, the overflow is left as-is, leaving it at the mercy of the C compiler to be optimized in ways that aren't allowed in Rust.
fn main() {
let mut i = 0i32;
while i > -1 {
i += 1;
}
println!("Done");
}
When compiled with the reference implementation, this prints “Done”, in accordance with the defined behaviour for signed wrapping in Rust. However, when compiled into C and then compiled with GCC, it optimizes the entire main function away, and leaves just an infinite loop.
I have some good news. I was able to trick the compiler into treating signed overflow as wrapping using the following C code, without requiring -fwrapv.
typedef union {
signed s;
unsigned u;
} sint;
int count_up()
{
sint i;
i.s = 0;
while(i.s > -1) {
i.u += 1;
}
return i.s;
}
int count_up_ub()
{
int i;
i = 0;
while(i > -1) {
i += 1;
}
return i;
}
I wonder if this technique could be used by the C code generator.
Signed integer overflow in Rust is supposed to wrap
Not exactly. Integer overflow (both signed and unsigned) in Rust is
invalid. That is, if your code causes an integer overflow, then your code
is defective. Whether or not the language catches you doing this is up to
the compiler options. By default, in debug builds, overflow reliably causes
a panic. In release builds, the behavior is that overflow occurs, and you
get whatever you get. This just happens to be wrapping arithmetic, but this
is not a guarantee of the language or compiler. The behavior, even in
release builds, can be changed by enabling overflow checks.
If your code overflows, it is broken. You should never rely on overflow
wrapping. If you want wrapping behavior, then you should use the wrapping
operators:
i = i.overflowing_add(1);
See:
https://github.com/rust-lang/rfcs/issues/359
http://huonw.github.io/blog/2016/04/myths-and-legends-about-integer-overflow-in-rust/
On Sat, Jun 8, 2019 at 4:20 PM Serentty notifications@github.com wrote:
I have some good news. I was able to trick the compiler into treating
signed overflow as wrapping using the following C code, without requiring
-fwrapv.
typedef union {
signed s;
unsigned u;
} sint;
int count_up()
{
sint i;
i.s = 0;
while(i.s > -1) {
i.u += 1;
}
return i.s;
}
int count_up_ub()
{
int i;
i = 0;
while(i > -1) {
i += 1;
}
return i;
}
I wonder if this technique could be used by the C code generator.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
https://github.com/thepowersgang/mrustc/issues/117?email_source=notifications&email_token=ADLILBDBAIJVGO4EVZRZW3LPZQ5CXA5CNFSM4HWIJRC2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXIAN5Q#issuecomment-500172534,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ADLILBCDB2HQSLWFJN7EHELPZQ5CXANCNFSM4HWIJRCQ
.
This just happens to be wrapping arithmetic, but this is not a guarantee of the language or compiler.
I believe it is guaranteed to either panic or wrap around on overflow.
Quoting the reference:
When the programmer has enabled debug_assert! assertions (for example, by enabling a non-optimized build), implementations must insert dynamic checks that panic on overflow. Other kinds of builds may result in panics or silently wrapped values on overflow, at the implementation's discretion.
In the case of implicitly-wrapped overflow, implementations must provide well-defined (even if still considered erroneous) results by using two's complement overflow conventions.
"It may be X or it may be Y" is not something you can rely on, though. If
you're writing a library and publishing it as a crate, you don't have
control over the compiler flags that are used. So you shouldn't rely on one
behavior or the other -- you should express the semantics you want, by
using the overflowing_xxx() operations. Yes, they're more verbose, but they
are precise.
On Mon, Jun 10, 2019 at 11:50 AM bjorn3 notifications@github.com wrote:
This just happens to be wrapping arithmetic, but this is not a
guarantee of the language or compiler.
I believe it is guaranteed to either panic or wrap around on overflow.
Quoting the reference
https://doc.rust-lang.org/reference/behavior-not-considered-unsafe.html:
When the programmer has enabled debug_assert! assertions (for example, by
enabling a non-optimized build), implementations must insert dynamic checks
that panic on overflow. Other kinds of builds may result in panics or
silently wrapped values on overflow, at the implementation's discretion.
In the case of implicitly-wrapped overflow, implementations must provide
well-defined (even if still considered erroneous) results by using two's
complement overflow conventions.
—
You are receiving this because you commented.
Reply to this email directly, view it on GitHub
https://github.com/thepowersgang/mrustc/issues/117?email_source=notifications&email_token=ADLILBHMYGPWWDJ5FAQJL2TPZ2O5TA5CNFSM4HWIJRC2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGODXK3HAA#issuecomment-500544384,
or mute the thread
https://github.com/notifications/unsubscribe-auth/ADLILBCEN2C4V6LCJNT4PVDPZ2O5TANCNFSM4HWIJRCQ
.
"It may be X or it may be Y" is not something you can rely on, though.
You can rely on the fact that it isn't UB.
As far as I understand, overflow in Rust is defined to either panic or wrap. Which it does is not defined, but it is defined that it does one of those two, and which actually happens depends on compiler flags. What is not allowed by the language standard is for the compiler to assume that overflow will not occur, the way that the C language standard allows. This means that optimizations based on the assumption that overflow will not occur are not valid.
We can solve this in two different ways:
For GCC and Clang, we can use -fwrapv in release builds.
On other platforms, we can add runtime checks that unconditionally panic.
While adding -fwrapv to the default gcc arguments would be the easiest approach, there doesn't seem to be an equivalent option for MSVC (well, not an officially supported one).
| gharchive/issue | 2019-06-08T23:09:16 | 2025-04-01T06:45:59.929904 | {
"authors": [
"DemiMarie",
"Serentty",
"arlied-google",
"bjorn3",
"thepowersgang"
],
"repo": "thepowersgang/mrustc",
"url": "https://github.com/thepowersgang/mrustc/issues/117",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1728462959 | back button in sign up page
Is your feature request related to a problem? Please describe.
when I tried opening sign up page and wanted to back there was no option so its convenient if back button is there
Describe the solution you'd like
add back button would solve the problem button in extreme right corner
Describe alternatives you've considered
A clear and concise description of any alternative solutions or features you've considered.
Additional context
Add any other context or screenshots about the feature request here.
Hello, Gaur
Thank you for bringing up the issue and suggesting a solution. It seems that there is currently no back button available on the Sign Up page, which makes it inconvenient for users who want to go back to the previous page.
To address this problem, I suggest adding a back button in the extreme right corner of the Sign Up page. This button would allow users to easily navigate back to the previous page without any hassle.
Alternatively, instead of a back button, you could consider adding a "Cancel" or "Go Back" link at the top or bottom of the page as an alternative solution. This would provide a clear and intuitive way for users to navigate back without cluttering the UI.
Adding a back button or a navigational link will improve the user experience by providing a seamless way to go back and forth between pages.
If you have any additional context or screenshots that could further illustrate the feature request, please provide them. It will help in better understanding the issue and finding an appropriate solution.
Thank you for your contribution and for raising this concern. Let me know if you have any further questions or if there's anything else I can assist you with.
Best regards,
@thestarsahil
Sir I can add "cancel" or "Go back " button.
Please assign me this issue so that I can resolve it under GSSOC-23
@gaur15
Thank you for your contribution to the project! We appreciate your willingness to help out.
At this time, we don't need this fearture to add on , but we will definitely keep it in mind for the future. If we do need your help, we will reach out to you.
| gharchive/issue | 2023-05-27T05:07:46 | 2025-04-01T06:46:00.035111 | {
"authors": [
"gaur15",
"manoranjanonline",
"thestarsahil"
],
"repo": "thestarsahil/Counsellor",
"url": "https://github.com/thestarsahil/Counsellor/issues/69",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
211069326 | No timeout
I'm using both the WebsocketServer and WebsocketClient. When I e.g. unplug ethernet cable with a connection open, the server has a timeout so it will eventually drop the connection, however on the client side there is no timeout causing the client to never drop it's connection. This prevents automatic reconnect that I have inside the close handler of the WebsocketConnection. As a workaround I'm currently doing the following which seems to work fine.
ws_client.on('connect', function(ws_conn) {
// Set a socket timeout to allow reconnections
ws_conn.socket.setTimeout(30*1000);
ws_conn.socket.on('timeout', function() {
ws_conn.socket.destroy();
});
}
But feels slightly hacky. Should I look into this and propose PR. Is this workaround normal? Please advice.
This thing should definitely be configurable as client option with reasonable default value, just faced it in inter-server communication.
| gharchive/issue | 2017-03-01T12:38:50 | 2025-04-01T06:46:00.042662 | {
"authors": [
"nazar-pc",
"vidstige"
],
"repo": "theturtle32/WebSocket-Node",
"url": "https://github.com/theturtle32/WebSocket-Node/issues/275",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
619319923 | can't get Japa to work with new version of Adonis 5
Hi guys,
i just installed japa to write some tests for my repositories/service classes
I'm getting this error while trying to run with node.
With the provided information, all I can do is nothing?
my bad @thetutlage, will provide extra information. Thanks
| gharchive/issue | 2020-05-15T23:13:19 | 2025-04-01T06:46:00.044462 | {
"authors": [
"pikitgb",
"thetutlage"
],
"repo": "thetutlage/japa",
"url": "https://github.com/thetutlage/japa/issues/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
789194553 | Should we accept metadata includes unrecognised fields?
Description of issue or feature request:
The current reference implementation will ignore unrecognised fields in a metadata file when loading it. Furthermore, some TAPs (i.e. snapshot Merkle Tree) are relying on this behaviour to introduce new fields to the specification without making the metadata invalid for older clients which don't recognise the field.
Current behavior:
The current client implementation ignores unrecognised/unexpected fields.
The new metadata API currently does not support unrecognised fields. If a metadata file including an unexpected field is loaded, the load fails.
Expected behavior:
To Be Determined.
We should decide (and capture in an ADR) whether we expect to be able to load metadata which includes additional fields beyond wha the implementation expects.
See also a related issue in the in-toto reference implementation: https://github.com/in-toto/in-toto/issues/186
I don't see the danger as long as the client doesn't change its behavior with respect to unrecognized fields, but I might be missing something.
I don't see the danger as long as the client doesn't change its behavior with respect to unrecognized fields, but I might be missing something.
Allowing the additional fields allows for more flexibility in backwards-compatibility, plus it's implied in the spec. So I'd say we should continue to allow it, and add an ADR to that effect.
Allowing the additional fields allows for more flexibility in backwards-compatibility, plus it's implied in the spec. So I'd say we should continue to allow it, and add an ADR to that effect.
I personally think it is useful to be able to add a custom application level metadata info.
for example it may be a Time server timestamp, apt release file, custom configuration .. etc
I personally think it is useful to be able to add a custom application level metadata info.
for example it may be a Time server timestamp, apt release file, custom configuration .. etc
It feels like we all agree that we should just ignore unrecognized fields.
The next question is what to do with them exactly?
While working on the new metadata Key class I just ignored the keyid_hash_algorithms, but then I had problems in the
tests/test_api.sign_verify().
The problem is that what I am verifying is a metadata object without keyid_hash_algorithms and a signature calculated when it was there.
So, again: what should we do with those unrelated fields? Should we store them in our objects or just ignore them?
It feels like we all agree that we should just ignore unrecognized fields.
The next question is what to do with them exactly?
While working on the new metadata Key class I just ignored the keyid_hash_algorithms, but then I had problems in the
tests/test_api.sign_verify().
The problem is that what I am verifying is a metadata object without keyid_hash_algorithms and a signature calculated when it was there.
So, again: what should we do with those unrelated fields? Should we store them in our objects or just ignore them?
I personally think it is useful to be able to add a custom application level metadata info.
for example it may be a Time server timestamp, apt release file, custom configuration .. etc
You should be able to attach most of this per-package custom metadata using the custom targets metadata.
I personally think it is useful to be able to add a custom application level metadata info.
for example it may be a Time server timestamp, apt release file, custom configuration .. etc
You should be able to attach most of this per-package custom metadata using the custom targets metadata.
The next question is what to do with them exactly? What does it mean to ignore them?
I think we should simply ignore them and not drop anything while reading and writing again.
The next question is what to do with them exactly? What does it mean to ignore them?
I think we should simply ignore them and not drop anything while reading and writing again.
Allowing the additional fields allows for more flexibility in backwards-compatibility, plus it's implied in the spec.
Wow, I'd somehow forgotten this line:
All of the formats described below include the ability to add more attribute-value fields for backwards-compatible format changes. If a backwards incompatible format change is needed, a new filename can be used.
The next question is what to do with them exactly? What does it mean to ignore them?
I think we should simply ignore them and not drop anything while reading and writing again.
Indeed, if we do otherwise we are not supporting the specification behaviour documented above and relied on by some TAPs.
Allowing the additional fields allows for more flexibility in backwards-compatibility, plus it's implied in the spec.
Wow, I'd somehow forgotten this line:
All of the formats described below include the ability to add more attribute-value fields for backwards-compatible format changes. If a backwards incompatible format change is needed, a new filename can be used.
The next question is what to do with them exactly? What does it mean to ignore them?
I think we should simply ignore them and not drop anything while reading and writing again.
Indeed, if we do otherwise we are not supporting the specification behaviour documented above and relied on by some TAPs.
Thinking about this a bit more, and looking at TAPs which reference or rely on this behaviour in the specification (TAPs 6, 10, 14, and 15), I think we (at least python-tuf) have three personas of metadata consumers:
TUF clients (updater), which only read the metadata (they write/download, but don't generate or modify, metadata files)
TUF repositories, which generate metadata – both from scratch, and modifying existing metadata files
developer tools, which generate and modify a subset of metadata
In the case of 1, the metadata API should be able to ignore (not drop, not choke on) unrecognised metadata fields. For 3 also, I think it makes sense to be able to modify metadata (i.e. add new signatures) without necessarily having to recognise and understand all of the fields in the metadata (see i.e. TAP 6).
For 2, I think it would be desirable to be able to (if not default to) only write and sign recognised metadata fields.
So, I think we want our metadata API to be able to read/write metadata without losing any unanticipated fields. But perhaps we should make writing unrecognised fields a boolean option (which defaults to off)?
Thinking about this a bit more, and looking at TAPs which reference or rely on this behaviour in the specification (TAPs 6, 10, 14, and 15), I think we (at least python-tuf) have three personas of metadata consumers:
TUF clients (updater), which only read the metadata (they write/download, but don't generate or modify, metadata files)
TUF repositories, which generate metadata – both from scratch, and modifying existing metadata files
developer tools, which generate and modify a subset of metadata
In the case of 1, the metadata API should be able to ignore (not drop, not choke on) unrecognised metadata fields. For 3 also, I think it makes sense to be able to modify metadata (i.e. add new signatures) without necessarily having to recognise and understand all of the fields in the metadata (see i.e. TAP 6).
For 2, I think it would be desirable to be able to (if not default to) only write and sign recognised metadata fields.
So, I think we want our metadata API to be able to read/write metadata without losing any unanticipated fields. But perhaps we should make writing unrecognised fields a boolean option (which defaults to off)?
So, I think we want our metadata API to be able to read/write metadata without losing any unanticipated fields. But perhaps we should make writing unrecognised fields a boolean option (which defaults to off)?
I agree, adding a signature to something unknown may be a bit risky, but this would allow developers or repositories to add any additional metadata fields in a backwards-compatible manner for the updater.
So, I think we want our metadata API to be able to read/write metadata without losing any unanticipated fields. But perhaps we should make writing unrecognised fields a boolean option (which defaults to off)?
I agree, adding a signature to something unknown may be a bit risky, but this would allow developers or repositories to add any additional metadata fields in a backwards-compatible manner for the updater.
Thinking about this a bit more, and looking at TAPs which reference or rely on this behaviour in the specification (TAPs 6, 10, 14, and 15), I think we (at least python-tuf) have three personas of metadata consumers:
TUF clients (updater), which only read the metadata (they write/download, but don't generate or modify, metadata files)
TUF repositories, which generate metadata – both from scratch, and modifying existing metadata files
developer tools, which generate and modify a subset of metadata
In the case of 1, the metadata API should be able to ignore (not drop, not choke on) unrecognised metadata fields. For 3 also, I think it makes sense to be able to modify metadata (i.e. add new signatures) without necessarily having to recognise and understand all of the fields in the metadata (see i.e. TAP 6).
For 2, I think it would be desirable to be able to (if not default to) only write and sign recognised metadata fields.
So, I think we want our metadata API to be able to read/write metadata without losing any unanticipated fields. But perhaps we should make writing unrecognized fields a boolean option (which defaults to off)?
It seems to me there are multiple good reasons why we would:
not drop or ignore additional unrecognized fields
store those additional fields as an attribute, I would call it additional_kwarg
don't use them in any way other than:
initiating an object (or in the new metadata API terms call from_dict() )
give a dictionary representation of our class (or in the new metadata API terms call to_dict())
I will create an ADR documenting this decision.
not drop or ignore additional unrecognized fields
store those additional fields as an attribute, I would call it additional_kwarg
Something like additional_fields or unrecognised_fields might be a more descriptive name?
don't use them in any way other than:
initiating an object (or in the new metadata API terms call from_dict() )
give a dictionary representation of our class (or in the new metadata API terms call to_dict())
I will create an ADR documenting this decision.
I think that would be useful. No need for it to be a significant document, we just want to ensure we capture the implication in the spec that metadata may contain unrecognised attribute-value fields and that we should handle those in our implementation.
not drop or ignore additional unrecognized fields
store those additional fields as an attribute, I would call it additional_kwarg
Something like additional_fields or unrecognised_fields might be a more descriptive name?
I liked unrecognised_fields. I agree this is a more descriptive name.
don't use them in any way other than:
initiating an object (or in the new metadata API terms call from_dict() )
give a dictionary representation of our class (or in the new metadata API terms call to_dict())
I will create an ADR documenting this decision.
I think that would be useful. No need for it to be a significant document, we just want to ensure we capture the implication in the spec that metadata may contain unrecognised attribute-value fields and that we should handle those in our implementation.
The document will be small this time, I promise. :D :D
| gharchive/issue | 2021-01-19T16:59:42 | 2025-04-01T06:46:00.077695 | {
"authors": [
"MVrachev",
"houdini91",
"joshuagl",
"mnm678",
"trishankatdatadog"
],
"repo": "theupdateframework/tuf",
"url": "https://github.com/theupdateframework/tuf/issues/1266",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
197802527 | Typo?
In chapter: 2.0implementationBasics.md,
We take a sub string of the URL and remove the /delete/ part and we have the ID of the task.
but the code above we strip the /tasks/ actually, is this a typo?
You're right, it should say
remove the /tasks/ part
File a Pull Request?
Yes, it is a typo. It is embarassing. If you want to file a PR I'd be happy to merge it :-)
it's a tiny modification :), you just change it.
okay :-)
| gharchive/issue | 2016-12-28T06:23:37 | 2025-04-01T06:46:00.085547 | {
"authors": [
"rongyi",
"tedyoung",
"thewhitetulip"
],
"repo": "thewhitetulip/web-dev-golang-anti-textbook",
"url": "https://github.com/thewhitetulip/web-dev-golang-anti-textbook/issues/35",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
339409627 | Repeatedly resetting after trying to get NTP time
FYI the problem and solution.
The TTGO ESP32 LoRa kept resetting after boot due to the problem that it cannot found an IP address of NTP server.
The printed serial message was shown:
Setup:: ERROR hostByName NTP
abort() was called at PC 0x400d9c22 on core 1
Backtrace: 0x4008fe80:0x3ffb1ee0 0x40090083:0x3ffb1f00 0x400d9c22:0x3ffb1f20 0x400db652:0x3ffb1f40 0x40137d23:0x3ffb1fa0
Rebooting...
I fix this problem by setting DNS to 8.8.8.8 as the following:
ip_addr_t d;
d.type = IPADDR_TYPE_V4;
d.u_addr.ip4.addr = static_cast<uint32_t>((uint32_t)0x08080808); // Google DNS
dns_setserver(0, &d);
if (!WiFi.hostByName(NTP_TIMESERVER, ntpServer)) // Get IP address of Timeserver
{
die("Setup:: ERROR hostByName NTP");
...
I have the same error with Wemos D1 Mini Pro board.
Thank you all for sharing.
Finally I found that the DNS assignment is not help the gateway in solving a host name.
I guess because of some security issue on the network -- not allowing the DNS query forwarding.
So, in my setup private network, the DHCP server has to assign a DNS server of the office network to the ESP32 LoRa gateway.
The setup network comprises of:
linux (of course),
DHCPd with a configuration line:
...
option domain-name-servers ;
...
HostAPd
After testing so far, nothing recurrently go wrong again.
Finally worked! there is something wrong with my local LAN DHCP. I have put as secondatry option the SSID of my grandmother's house (1 floor down) and worked! Also when I put my mobile as hotspot with credentials of my house wifi works!
I will close this issue, as it seems not related to the gateway but to the network/DNS setup.
| gharchive/issue | 2018-07-09T11:24:32 | 2025-04-01T06:46:00.120863 | {
"authors": [
"dksmar",
"hotzenklotz",
"iPAS",
"platenspeler"
],
"repo": "things4u/ESP-1ch-Gateway-v5.0",
"url": "https://github.com/things4u/ESP-1ch-Gateway-v5.0/issues/37",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
50487275 | Consolidate "count" in TitanVertexQuery
reconcile the different variants. Should "count()" just be for relations? How to count all edges on a vertex then? Where and how much is this currently used?
We will have:
count() => # of relations
edgeCount() => # of edges
propertyCount() => # of properties
| gharchive/issue | 2014-12-01T04:43:11 | 2025-04-01T06:46:00.164201 | {
"authors": [
"mbroecheler"
],
"repo": "thinkaurelius/titan",
"url": "https://github.com/thinkaurelius/titan/issues/859",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1246111315 | Unable to use ThirdwebWeb3Provider in Next Js using Typescript
I'm Using Next Js with typescript. I keep getting the error message Type '{ children: (string | Element)[]; supportedChainIds: number[]; connectors: { injected: {}; }; }' is not assignable to type 'IntrinsicAttributes & ThirdwebWeb3ProviderProps'. Property 'children' does not exist on type 'IntrinsicAttributes & ThirdwebWeb3ProviderProps'. not sure what part is wrong, I did follow the instruction on how to set up @3rdweb/hooks
e
import { ThirdwebWeb3Provider } from '@3rdweb/hooks';
import type { AppProps } from 'next/app';
import 'regenerator-runtime/runtime';
// SOURCE: https://dev.to/byteslash/metamask-authentication-in-nextjs-with-third-web-55ff
// setting up auth with @3rdweb
function MyApp({ Component, pageProps }: AppProps) {
const supportedChainIds = [1, 4];
const connectors = {
injected: {},
};
return (
<ThirdwebWeb3Provider
supportedChainIds={supportedChainIds}
connectors={connectors}
>
<Component {...pageProps} />;
</ThirdwebWeb3Provider>
);
}
export default MyApp;
@NoxinDVictus @3rdweb/hooks is deprecated, please use @thirdweb-dev/react instead.
| gharchive/issue | 2022-05-24T07:33:32 | 2025-04-01T06:46:00.174115 | {
"authors": [
"NoxinDVictus",
"jarrodwatts"
],
"repo": "thirdweb-dev/react",
"url": "https://github.com/thirdweb-dev/react/issues/67",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1327886324 | fix: restore compatibility with experimental built-in fetch in Node v18
As of Node.js v18, Node has a built-in, experimental support for global fetch. Unfortunately, this implementation is not compatible with form-data npm package and when an instance of this package's FormData class is provided as a body the native implementation resorts to serialising the form data by calling .toString() on it, causing the request's body to become the string literal "[object FormData]".
We can work around that compatibility issue by explicitly using the form's headers and body as a buffer as the inputs to fetch.
Amazing, we haven't had much time to look into Node 18 support, so this helps a lot. Does everything else work for you with that fix? If so I can setup our automated tests to run on node 18 and have that running for all changes moving forward
@joaquim-verges well, this was a blocker for me on step one after installing the SDK so I honestly don't know how the rest of the SDK is looking on Node v18, this allows me to actually mint some tokens and that's a great start for me now. 😄 I will let you guys know if I discover any additional issues. 💪
Couldn't get the node 18 tests to run because of a hardhat dependency that doesn't support it. But tested this on node 14/16 and browser and it seems fine. Merging. cc @jnsdls
| gharchive/pull-request | 2022-08-03T23:30:36 | 2025-04-01T06:46:00.177496 | {
"authors": [
"joaquim-verges",
"robertrossmann"
],
"repo": "thirdweb-dev/typescript-sdk",
"url": "https://github.com/thirdweb-dev/typescript-sdk/pull/532",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1469723212 | [deno] 2. Setup API type definition configuration
Background
REST
This should be to setup a place to put API type definitions and ensuring out-of-the-box API type definitions are available.
GraphQL
This should be to setup GraphQL codegen so types are auto-generated.
Acceptance
[ ] Add methods for generating API project types for consumption
The Node kit is handling this through Swagger codegen. We should consider similar approaches.
@vyktoremario, this one is still marked "In Progress". Was there any response from the Discord channel where you asked about the issue with running codegen with Deno?
If you could write a quick status here and let's see how to proceed further.
| gharchive/issue | 2022-11-30T14:26:02 | 2025-04-01T06:46:00.180032 | {
"authors": [
"dariodjuric",
"robocel"
],
"repo": "thisdot/starter.dev",
"url": "https://github.com/thisdot/starter.dev/issues/566",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1782871034 | Update jupyter_embed.py
On Windows, I needed to change the encoding to 'utf-8', since that isn't the default on Windows (and is the default of Linux and mac).
After the simple update I was able to use the embed function in my notebooks.
sorry for the delay here! Will do a release soon.
| gharchive/pull-request | 2023-06-30T17:13:48 | 2025-04-01T06:46:00.201178 | {
"authors": [
"dfhannum",
"thomasballinger"
],
"repo": "thomasballinger/observable-jupyter",
"url": "https://github.com/thomasballinger/observable-jupyter/pull/20",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
} |
732701188 | no Brand icon... in the integrations list
Hardly a first world problem, but in the list of integrations, Fav-icon is failing a brand-icon.
leading to this error in inspector:
Failed to load resource: the server responded with a status of 404 ()
would be cool if this nice little utility would show a (fav)icon.
Just wanted to add this issue too. Seems like an easy fix.
👍
Kind of ironic, though, we now have a custom icon, but the plugin itself does not.
You could just use the icon that we defined, right?
You could just use the icon that we defined, right?
what exactly are you saying here? Are you going to add it in a PR?
It was just a suggestion.
I briefly looked into this (just to satisfy my own curiosity) and it looks like a dynamic icon will not be possible. https://github.com/home-assistant/brands is the repo that holds these icons, if anyone wants to design or propose something that makes sense for this integration.
a good integration that helps on visual !!
could have a fix for this issue?
Idea: the blue favicon and one in another color in front of them, like the copy icon.
What about "updating" the favicons in folder test too?
| gharchive/issue | 2020-10-29T22:17:30 | 2025-04-01T06:46:00.208106 | {
"authors": [
"Mariusthvdb",
"dawiinci",
"hiagocosta",
"mingaldrichgan",
"renarena"
],
"repo": "thomasloven/hass-favicon",
"url": "https://github.com/thomasloven/hass-favicon/issues/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2141047682 | 🛑 Speedtest-v2-NBG1-1 is down
In c188921, Speedtest-v2-NBG1-1 ($URL_SPEEDTEST_V2_NBG1_1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Speedtest-v2-NBG1-1 is back up in 01e4c4e after 5 minutes.
| gharchive/issue | 2024-02-18T15:40:06 | 2025-04-01T06:46:00.232275 | {
"authors": [
"thomasmerz"
],
"repo": "thomasmerz/upptime",
"url": "https://github.com/thomasmerz/upptime/issues/2457",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1523129986 | failed to toggle bookmark
Issue
When an item is bookmarked, it can only be cleared on the main page, clicking the bookmark Item will do nothing
How to reproduce
randomly select a stop (e.g. https://thomassth.github.io/ttc-eta/stops/14684)
bookmarked it by clicking the bookmark button
try clicking the button again to un-bookmark it
expected behaviour
un-bookmarking a stop smoothly
actual behaviour
failed to un-bookmark a stop
Closed with #23
| gharchive/issue | 2023-01-06T20:34:49 | 2025-04-01T06:46:00.243779 | {
"authors": [
"HoiPangCHEUNG",
"thomassth"
],
"repo": "thomassth/ttc-bus-eta",
"url": "https://github.com/thomassth/ttc-bus-eta/issues/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2627236803 | Update dependency Microsoft.Testing.Platform.MSBuild to 1.4.2
This PR contains the following updates:
Package
Type
Update
Change
Microsoft.Testing.Platform.MSBuild
nuget
patch
1.4.1 -> 1.4.2
Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
[ ] If you want to rebase/retry this PR, check this box
This PR has been generated by Renovate Bot.
Edited/Blocked Notification
Renovate will not automatically rebase this PR, because it does not recognize the last commit author and assumes somebody else may have edited the PR.
You can manually request rebase by checking the rebase/retry box above.
⚠️ Warning: custom changes will be lost.
| gharchive/pull-request | 2024-10-31T16:00:39 | 2025-04-01T06:46:00.257333 | {
"authors": [
"thomhurst"
],
"repo": "thomhurst/TUnit",
"url": "https://github.com/thomhurst/TUnit/pull/1129",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2140155533 | Need helper function to change the home load
A helper function is needed (also for Home Assistant integration) to simplify change of the set home load.
Points to clarify and validate before this can be provided:
Can the set_home_load endpoint be used with the same schedule parameter dict as the set_device_parm endpoint?
Does the schedule parameter dict require a full schedule as received with either get_home_load or get_device_parm endpoints?
Is it sufficient to modify only the active schedule slot with the new parameters for home load and discharge setting, or may this delete the other slots in the schedule?
If only a single slot is supported, what happens if the times of the slot are changed? Will the modified timings delete/modify the adjacent slots like in the App?
The help function should provide option switches for:
new home load value
setting of discharge switch
whether the change should be applied to current slot only or all slots
If the whole schedule or parts of it needs to be changed, this should be done via the set_device_parm method providing the full schedule json.
The method also need to ensure to update the internal device dictionarie with the modified schedule for each solarbank device. This is important that changes are immediately reflected in the api data structures and not delay until next device details update cycle.
The schedule slots are also different for single and dual bank setups which must be considered for the helper.
Schedule slot with Single Solarbank in system
{
"id": 0,
"start_time": "20:00",
"end_time": "24:00",
"turn_on": true,
"appliance_loads": [
{
"id": 0,
"name": "Benutzerdefiniert",
"power": 100,
"number": 1
}
],
"charge_priority": 80,
"power_setting_mode": 1,
"device_power_loads": [
{
"device_sn": "25MFU7OHZJTFGFSH",
"power": 50
}
]
}
Schedule slot with Dual Solarbank in system
{
"id": 0,
"start_time": "17:30",
"end_time": "24:00",
"turn_on": false,
"appliance_loads": [
{
"id": 0,
"name": "Benutzerdefiniert",
"power": 170,
"number": 1
}
],
"charge_priority": 80,
"power_setting_mode": 1,
"device_power_loads": [
{
"device_sn": "VQQDB61DTTZEPSVX",
"power": 85
},
{
"device_sn": "DG34YZMOHGOTW6W2",
"power": 85
}
]
}
First tests showed that the set_device_load endpoint accepts the new schedule without error, but it doesn't seem to get applied.
Schedule modifications are only applied via the set_device_parm method.
The parameter must always be the full schedule. Missing slots will automatically extended to complete a full day with the provided slots.
The Power settings must only be applied to the appliance load dict. This will automatically be devided for the individual device power loads. They don't have to be provided for schedule modifications.
Important and modifyable parameters for a slot:
"start_time": "17:30",
"end_time": "24:00",
"turn_on": false,
"appliance_loads": [
{
"name": "Benutzerdefiniert",
"power": 170,
}
],
"charge_priority": 80,
Maybe you can have a look in the npm module ankersolixe1600 - there is setting the homeload (or turning home feeding off) implemented.
The mechanism for appliance load setting is understood and implemented in a helper function.
App version 2.1.0 implemented new capabilities for dual solarbank setup to modify the load sharing. Default so far was 50% of appliance load. The new mechanism is not understood yet and I can't test it. It can be added at a later point when this is clear.
I assume this is why the set_device_load endpoint was implemented. Maybe this works only for dual bank setups, and does nothing for single bank setup...
| gharchive/issue | 2024-02-17T14:46:07 | 2025-04-01T06:46:00.266404 | {
"authors": [
"p4i1ipp",
"thomluther"
],
"repo": "thomluther/anker-solix-api",
"url": "https://github.com/thomluther/anker-solix-api/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1527299255 | Add + to COS probes filename...
... as the kernel release includes one and falco expects it when looking for the filename:
* Filename 'falco_cos_5.4.144+_1.o' is composed of:
@sHesl can and would we delete the old files without the + in the name?
| gharchive/pull-request | 2023-01-10T12:21:19 | 2025-04-01T06:46:00.356589 | {
"authors": [
"markchalloner"
],
"repo": "thought-machine/falco-probes",
"url": "https://github.com/thought-machine/falco-probes/pull/61",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
106241302 | Warnings about unknown URLs still produced when using allow_unknown_urls
I have this configuration:
Capybara::Webkit.configure do |config|
config.allow_unknown_urls
config.ignore_ssl_errors
end
But I still get a warning when running my tests:
Request to unknown URL: https://sub.localdomain.lan:449/api//pastry_kind/cake_color
To block requests to unknown URLs:
Capybara::Webkit.configure do |config|
config.block_unknown_urls
end
To allow just this URL:
Capybara::Webkit.configure do |config|
config.allow_url("https://sub.localdomain.lan:449/api//pastry_kind/cake_color")
end
To allow requests to URLs from this host:
Capybara::Webkit.configure do |config|
config.allow_url("sub.localdomain.lan")
end
Looking at the implementation of allow_unknown_urls, it seems pretty straight-forward, so I must be doing something silly somewhere. Is there any reason that the global allow wouldn't work?
Other URLs that I access (there are multiple), all seem to work fine.
Versions:
Capybara-Webkit 1.7.0
Capybara 2.4.4
Ruby 2.2.3p173
Qt 4.8.6
Have you registered your own driver? It's possible that you're not providing the configuration object to the driver, which would result in the options being ignored. The driver must be registered with Capybara::Configuration.to_hash as the second argument. It's likely that you can eliminate any custom driver you were using as all the options were rolled in to the configuration object. See #806, where we intend to issue a warning for the issue.
Have you registered your own driver?
Nope; I'm using whatever happens by requiring the gem.
It's possible that you're not providing the configuration object to the driver, which would result in the options being ignored.
Unlikely, as I can set the debug option here and it definitely takes effect. I also see certain blocked URL warnings disappear when I use the confit in my original comment. That's why it's such a strange occurrence.
I am also experiencing this.
Interesting. Could you extract a failing example that reproduces the issue? A complete example, including spec_helper.rb would be much appreciated. I haven't been able to reproduce this myself, but it's possible there's something I've missed.
I'm still working on narrowing it down, but this does not reproduce when I run locally on OS X. However, when I run in an Ubuntu Docker image, it does reproduce. I'm seeing about making a smaller example.
I've tracked it down a bit. Putting debug prints in UnknownUrlHandler::reset, I see that it's being called.
I only see this being called from WebPageManager::reset. I do note that if I throw a binding.pry in my step I don't see this occurring.
My current hypothesis is this:
Access a page that performs AJAX calls
Do whatever capybara check we want.
Start to shut down the test, which presumably calls the reset command, which ultimately resets the allowed URLs.
The page still has active JS which triggers an AJAX request at this point.
Any thoughts?
@shepmaster that's possible. Maybe we need to reorder our setup function to delete pending requests before changing other settings?
Just confirming that I'm seeing the same issue; I have explicitly allowed a URL which is requested via JS but I frequently receive a warning about blocking it. When I allow all unknown URLs the behaviour is unchanged, I still see a warning.
I use this workaround:
# spec/spec_helper.rb
Before('@javascript,@webkit') do
page.driver.allow_url('*')
end
This gives me only one deprecation warning instead of many unknown url messages.
I'm experiencing the same issue, seems it's not just showing these warnings, but also resets all custom webkit driver settings for requests on point 4. It caused hangs for my case, because my webkit settings were reseted, so maybe related to #623. Also #816 looks like the same issue.
As a temporary workaround I disabled the reset method on capybara, similar to this http://stackoverflow.com/a/30809650/600638
Same issue here. Capybara 2.4.4, Capybara Webkit 1.7.1.
Same issue here. Capybara 2.4.4, Capybara Webkit 1.7.1.
Any progress on this?
I haven't been able to work on this yet.
@jferris any ETA?
@lucasklaassen I can't provide an ETA, since I work on capybara-webkit during spare time unless I'm encountering a bug I have to fix for my own work. I think this is likely the highest priority issue, so I'll work on this next time I have available for capybara-webkit.
If somebody look at some of the earlier suggestions and put together a pull request, that would take me less time than writing it myself, so it could potentially get merged sooner.
@jferris just pinging you about this again for any updates? Your work on this project is much appreciated :)
I submitted #895 which I believe will resolve the race condition.
@shepmaster @lucasklaassen could one of you test out that branch to see if it resolves your issue? I still haven't been able to reproduce this locally.
@jferris Sorry to disappoint, but I've moved on from the job where I had those tests, so I doubt I'll be of much help in reproducing. The diff appears sane, for what that's worth. :smile_cat:
@jferris I unpacked the gem in a 'workspace' folder in my project and referenced the path it in my gemfile. I made the change from the PR you put up, ran bundle install and then ran bundle exec rspec. The alert messages for "Request to unknown URL" still appear even though I have the following code in my spec_helper.rb
Capybara::Webkit.configure do |config|
config.allow_unknown_urls
end
@lucasklaassen could you include the relevant parts of your configuration? Things like spec_helper.rb, rails_helper.rb, and things from spec/support may be useful. Also, a copy of your debug log (config.debug = true) may be useful.
spec_helper.rb:
require 'capybara/rspec'
require 'pry'
require 'selenium-webdriver'
require 'byebug'
require 'capybara/webkit'
require 'yaml'
require 'helpers/session_helpers'
require 'helpers/web_helpers'
require 'helpers/wait_for_ajax'
require 'faker'
SPEC_ROOT=File.expand_path(File.dirname(__FILE__) + '/../')
Capybara.default_driver = :webkit
Capybara.javascript_driver = :webkit
options = {js_errors: false, debug: false, skip_image_loading: true, ignore_ssl_errors: true, allow_unknown_urls: true}
Capybara.register_driver :webkit do |app|
Capybara::Webkit::Driver.new(app, options)
end
Capybara::Webkit.configure do |config|
config.allow_unknown_urls
end
Capybara.default_max_wait_time = 10
RSpec.configure do |config|
config.after(:all) do
# Destroy fixtures
end
config.around(:each) do |example|
begin
example.run
rescue Exception => ex
save_and_open_page
raise ex
end
end
config.include WaitForAjax, type: :feature
end
I am not using rails for this project so I don't have a rails_helper and I do not have a spec/support folder.
If you register your own Capybara::Webkit::Driver, it won't pick up options from Capybara::Webkit.configure. You can get the configuration Hash from Capybara::Webkit::Configuration.to_hash, which you can merge with other options if you'd like.
Note that this option is being deprecated shortly. You should be able to configure any options using configure block without registering your own driver.
I'm going to close this, as it seems like the race condition leading to this bug has been fixed. Please continue to comment or ask questions if you have any.
| gharchive/issue | 2015-09-13T21:36:42 | 2025-04-01T06:46:00.378881 | {
"authors": [
"Drakula2k",
"TheTeaNerd",
"dg-ratiodata",
"jferris",
"lucasklaassen",
"mhoran",
"patrick99e99",
"shepmaster",
"steakchaser",
"thorn"
],
"repo": "thoughtbot/capybara-webkit",
"url": "https://github.com/thoughtbot/capybara-webkit/issues/837",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1486563242 | It is unclear how to add or contribute to the different types of content on offer
Aside from the absolute basics of getting the site running, and link to the eleventy docs it is pretty unclear how to contribute to this project, the sub-tasks for this I see are:
Offer guidance on how to contribute to this project
Give a better overview of the project, it's history and what we hop it will achieve
List and document the internal resources that would help people to contribute
Document the build & deploy process
document how to add a new schedule
Document how to add a new glossary terms
Document how to add a new FAQ
I might need some help with this one:
Give a better overview of the project, it's history and what we hope it will achieve
I don't want to answer that question by myself.
| gharchive/issue | 2022-12-09T11:09:48 | 2025-04-01T06:46:00.383411 | {
"authors": [
"LkeMitchll"
],
"repo": "thoughtbot/design-sprint-guide",
"url": "https://github.com/thoughtbot/design-sprint-guide/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
55494433 | We must compile the CSS ourselves
Don't gitignore the CSS because GitHub pages won't auto-compile the SCSS to CSS for us. We have to do it ourselves.
Add the compiled CSS file back in.
Weird. I think they should do this now.
It works on my GH-pages site.
Huh. http://griddler.io/ is totally blank right now and has no CSS.
I didn't get any emails about failed builds.
Scroll down. You're right about main.css, though. It's 404ing.
"scroll down" Oh hey it's right there! This CSS thing is silly.
Typekit is also not found...
Figured it out, I think: https://github.com/thoughtbot/griddler/pull/195
| gharchive/pull-request | 2015-01-26T15:22:26 | 2025-04-01T06:46:00.389023 | {
"authors": [
"calebthompson",
"gabebw"
],
"repo": "thoughtbot/griddler",
"url": "https://github.com/thoughtbot/griddler/pull/194",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
49866397 | Running pick using system() in Vim hinders Vim from restoring the terminal on exit.
When I run pick from within Vim using the code below, when exiting Vim, the terminal state is not restored. Instead, the screen Vim drew is still visible.
function! PickCommand(choices_command, pick_args, vim_command)
try
let selection = system(a:choices_command . " | pick " . a:pick_args)
catch /Vim:Interrupt/
redraw!
return
endtry
redraw!
exec a:vim_command . " " . selection
endfunction
nnoremap <leader>p :call PickCommand("find * -type f", "", ":e")<cr>
It would be great if someone else could replicate this. I don't know if this is a bug in pick, Vim or my terminal emulator.
... It's probably in pick.
An idea that @calleerlandsson and I talked about IRL is to explicitly reset the terminal when closing ncurses. Right now we configuring ncurses on boot: https://github.com/thoughtbot/pick/blob/0c03b0d76a8546016f824f7f977593fcbf1dab90/src/ui.c#L46-L65
Maybe doing the same on when we're stopping ncursise might do the trick?
This is actually still happening for me... do you think there are any changes to the vim function needed? Here's what I'm using:
" pick
function! EscapeFilePath(path)
return substitute(a:path, ' ', '\\ ', 'g')
endfunction
function! Pick(choice_command, vim_command)
try
let selection = system(a:choice_command . " | pick")
catch /Vim:Interrupt/
" Swallow the ^C so that the redraw below happens; otherwise there will be
" UI leftovers on the screen
redraw!
return
endtry
redraw!
if strlen(selection) > 0
exec a:vim_command . " " . EscapeFilePath(selection)
endif
endfunction
nnoremap <leader>t :call Pick("find * -type f", ":e")<cr>
@ggilder see https://github.com/thoughtbot/pick/issues/32 you need to use pick -X
| gharchive/issue | 2014-11-24T08:35:48 | 2025-04-01T06:46:00.392689 | {
"authors": [
"Keithbsmiley",
"calleerlandsson",
"ggilder",
"teoljungberg"
],
"repo": "thoughtbot/pick",
"url": "https://github.com/thoughtbot/pick/issues/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
436256983 | Design capacity tracking algorithm that allow to power off node that are not used
In the spirit of trying to be as green as possible, we would like to come up with a scheme that allows a farmer to turn off the node that are not required (the capacity sold from the farm doesn't required these node to be up)
Currently we track the uptime of a node because the node does send heartbeats to the network. This makes things easy. But for this task, we need to come up with something that would allow us to still be sure the node is available while the node is turned off for power saving.
➤ Christophe de Carvalho commented:
My first idea regarding this is something like this:
the node are powered on with a specific flag, e.g: powersave
During the capacity tracking, the node detect the flag and send a special request to the capacity directory saying, I'm alive, here is my capacity,... but I will shut down, tell me when you want to hear from me again.
the directory register that information and return a JWT to the node. This JWT contains identification information and expires in XHours.
the node then receives the JWT, save it, and based on the expiration time, know when it needs to come up again
If the node come up again in time and is able to send the JWT back to the directory, we count the uptime of the node as complete (aka, like it was on all the time)
If the node fails to come up online in time, then we consider the node to be really offline and we don't count the uptime
This schema allows us to still be sure the node are actually available and not just powered off and not providing capacity anymore.
This is still pretty rough ideas thought, needs more thinking
➤ Maxime Daniel commented:
About node power reduction, we could investigate if hardware supports correctly suspend-to-ram, we could suspend and resume hardware quickly with low power impact using this. This could also help to check « does the machine can come back ? » without waiting +5 min hardware boot
➤ Christophe de Carvalho commented:
add a low priority on this one for now.
For now we will just allow farmer to have their node up once a day in order for them to get their farming token.
This give us time to come up with a proper solution on the long run.
| gharchive/issue | 2019-04-23T15:38:18 | 2025-04-01T06:46:00.430281 | {
"authors": [
"rkhamis"
],
"repo": "threefoldtech/digitalmeX",
"url": "https://github.com/threefoldtech/digitalmeX/issues/200",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
583624215 | Issue related to Creating instance [unstable_bcdb_tests]
The first instance is created with a default name regardless of the name I have given, the second instance is created with the right name , this issue on unstable branch
Fixed on unstable_bcdb_tests
verified
| gharchive/issue | 2020-03-18T10:25:47 | 2025-04-01T06:46:00.435837 | {
"authors": [
"Dina-Abd-Elrahman",
"RafyAmgadBenjamin"
],
"repo": "threefoldtech/jumpscaleX_core",
"url": "https://github.com/threefoldtech/jumpscaleX_core/issues/624",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
585694807 | jsx installation enhancment unstable branch
can we enhance below step to be non-interactive while install jsx
This happens because it clones with ssh in interactive mode
Suggested options:
use -s and non-interactive mode
Add github to local know hosts (could be tricky if remote changes)
Disable SSH host key checking For github in installtools
verified
/tmp/jsx container-install -s
| gharchive/issue | 2020-03-22T11:13:54 | 2025-04-01T06:46:00.438571 | {
"authors": [
"Pishoy",
"waleedhammam"
],
"repo": "threefoldtech/jumpscaleX_core",
"url": "https://github.com/threefoldtech/jumpscaleX_core/issues/641",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1490229497 | manual faq - Table of Contents urls leading to 404 error
all Table of Contents urls are leading to 404 not found error
https://library.threefold.me/info/manual/#/manual__faq
The right way to link subsections in this publishing system in a page should be like the following:
[Basic Facts](manual__faq?id=basic-facts)
Instead of
[basic facts](#basic-facts)
which should work fine in vanilla markdown, but not in the current implementation of the markdown in the publishing system.
Also, the publishingtools validation rules the link used to be either a full https link to a resource or just the page name, without any encoding of divs in it. It can't be fixed without a new release of the publishingtools
I'll move to later milestone to be checked again when this gets supported in the publishingtools framework.
Would gladly accept PRs against the new manual
https://github.com/threefoldtech/info_grid
| gharchive/issue | 2022-12-11T17:05:38 | 2025-04-01T06:46:00.442015 | {
"authors": [
"fl0wm0ti0n",
"maayarosama",
"xmonader"
],
"repo": "threefoldtech/test_feedback",
"url": "https://github.com/threefoldtech/test_feedback/issues/356",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
968268420 | Graphql subqueries crash the query node server
example:
{
nodes(limit:1){
publicConfig{
gw4
}
}
}
We are using reverse lookups for the nested objects, https://docs.subsquid.io/v/v3/schema-spec/entity-relationship#reverse-lookups
Fixed when upgrading to subsquid v4!
| gharchive/issue | 2021-08-12T07:40:47 | 2025-04-01T06:46:00.443874 | {
"authors": [
"DylanVerstraete"
],
"repo": "threefoldtech/tfchain",
"url": "https://github.com/threefoldtech/tfchain/issues/27",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2076199888 | can not close error message
Description
env used : main net
version : 2.2.0
error window should be collapsed if i clicked any where on the screen not just the close button
Screencast from 11 ينا, 2024 EET 11:53:03 ص.webm
fixed and verified on devnet #414
https://github.com/threefoldtech/tfgrid-sdk-ts/issues/1325
| gharchive/issue | 2024-01-11T10:01:22 | 2025-04-01T06:46:00.446589 | {
"authors": [
"khaledyoussef24"
],
"repo": "threefoldtech/tfgrid-sdk-ts",
"url": "https://github.com/threefoldtech/tfgrid-sdk-ts/issues/1936",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1476887485 | Add clear error messages when the user is not able to deploy an instance
Description
Add clear error messages when the user is not able to deploy the instance for any reason.
for example:
When there are no nodes in a specific region, there should be a clear message mentioning that this region doesn't have online nodes instead of showing Node ID is required and not being able to select a node.
Same logic should be applied if the resources are not enough in the selected region/farm.
And for every case there should be a clear error message why the user is not able to deploy the instance.
Verified on https://getmastodon.threefold.io/
a new logic was implemented in #79 where the Deploy button is active at all times but if the user clicks on it while an error is present it will focus on the field with a clear error message explaining the error.
and more error messages were added/Improved in:
https://github.com/threefoldtech/www-mastodon/issues/20
https://github.com/threefoldtech/www-mastodon/issues/75
https://github.com/threefoldtech/www-mastodon/issues/90
| gharchive/issue | 2022-12-05T16:04:48 | 2025-04-01T06:46:00.451445 | {
"authors": [
"mohamedamer453"
],
"repo": "threefoldtech/www-mastodon",
"url": "https://github.com/threefoldtech/www-mastodon/issues/76",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2042803129 | Does BPO model support prompt output in Chinese?
Thank you for sharing your model.
I'm interested in learning more details. Does the BPO model support input and output in Chinese language corpora? How can it be configured to produce Chinese-language results?
I've tried to switch input as Chinese or ask to provide CN output, it doesn't work right now
Currently we only support English input and output, we will consider developing models in Chinese scenarios, and a problem lies in the lack of high-quality feedback data in Chinese.
| gharchive/issue | 2023-12-15T02:54:28 | 2025-04-01T06:46:00.481482 | {
"authors": [
"chengjl19",
"huangyongk"
],
"repo": "thu-coai/BPO",
"url": "https://github.com/thu-coai/BPO/issues/13",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2610242743 | Why Running Llama infer in A10 get Wrong answer?
exmaples codes
import torch
from transformers import LlamaTokenizer, LlamaForCausalLM
from sageattention import sageattn
import torch.nn.functional as F
F.scaled_dot_product_attention = sageattn
# 加载预训练的 LLaMA 模型和 tokenizer
model_name = "llama-7b-hf"
tokenizer = LlamaTokenizer.from_pretrained(model_name)
model = LlamaForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16)
# 将模型移动到 GPU(如果可用)
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model.to(device)
# 准备输入文本
input_text = "Once upon a time, there was a little girl"
inputs = tokenizer(input_text, return_tensors="pt").to(device)
# 执行推理
with torch.no_grad():
output = model.generate(**inputs, max_length=50, num_return_sequences=1)
# 解码输出
generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
print(generated_text)
device envirments:
torch 2.5.0
triton 3.1.0
transformers 4.45.2
expects answer
Once upon a time, there was a little girl who loved to play with her friends. One day, she decided to play with her friends in the forest. She was very happy. She played with her friends in the forest. She played with
but got
Once upon a time, there was a little girl whole and the 1882P a.
ficenda2P avalN64YourEm.
ficOnDe Ce the GISP.
gev
We have not test the accuracy by using F.scaled_dot_product_attention = sageattn in Llama.
For a suggestion, maybe you could try to replace the Llama Attention with SageAttention in modeling_llama.py.
How to replace the Llama Attention with SageAttention ? @jt-zhang
Hi, you can set the Llama Attention in the following way:
# Adapted from LlamaAttention.forward
def LlamaSageAttnForward(
self,
hidden_states: torch.Tensor,
attention_mask: Optional[torch.Tensor] = None,
position_ids: Optional[torch.LongTensor] = None,
past_key_value: Optional[Cache] = None,
output_attentions: bool = False,
use_cache: bool = False,
cache_position: Optional[torch.LongTensor] = None,
position_embeddings: Optional[Tuple[torch.Tensor, torch.Tensor]] = None, # will become mandatory in v4.46
**kwargs,
) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
assert not output_attentions, "Output attentions not supported"
assert attention_mask is None, "Attention mask not supported"
assert self.num_key_value_groups == 1, "GQA will be supported in near future"
bsz, q_len, _ = hidden_states.size()
query_states = self.q_proj(hidden_states)
key_states = self.k_proj(hidden_states)
value_states = self.v_proj(hidden_states)
query_states = query_states.view(bsz, q_len, self.num_heads, self.head_dim).transpose(1, 2)
key_states = key_states.view(bsz, q_len, self.num_key_value_heads, self.head_dim).transpose(1, 2)
value_states = value_states.view(bsz, q_len, self.num_key_value_heads, self.head_dim).transpose(1, 2)
if position_embeddings is None:
cos, sin = self.rotary_emb(value_states, position_ids)
else:
cos, sin = position_embeddings
query_states, key_states = apply_rotary_pos_emb(query_states, key_states, cos, sin)
if past_key_value is not None:
# sin and cos are specific to RoPE models; cache_position needed for the static cache
cache_kwargs = {"sin": sin, "cos": cos, "cache_position": cache_position}
key_states, value_states = past_key_value.update(key_states, value_states, self.layer_idx, cache_kwargs)
if q_len == 1:
key_states = repeat_kv(key_states, self.num_key_value_groups)
value_states = repeat_kv(value_states, self.num_key_value_groups)
causal_mask = attention_mask
if attention_mask is not None:
causal_mask = causal_mask[:, :, :, : key_states.shape[-2]]
# SDPA with memory-efficient backend is currently (torch==2.1.2) bugged with non-contiguous inputs with custom attn_mask,
# Reference: https://github.com/pytorch/pytorch/issues/112577.
if query_states.device.type == "cuda" and causal_mask is not None:
query_states = query_states.contiguous()
key_states = key_states.contiguous()
value_states = value_states.contiguous()
# We dispatch to SDPA's Flash Attention or Efficient kernels via this `is_causal` if statement instead of an inline conditional assignment
# in SDPA to support both torch.compile's dynamic shapes and full graph options. An inline conditional prevents dynamic shapes from compiling.
is_causal = True if causal_mask is None and q_len > 1 else False
attn_output = torch.nn.functional.scaled_dot_product_attention(
query_states,
key_states,
value_states,
attn_mask=causal_mask,
dropout_p=self.attention_dropout if self.training else 0.0,
is_causal=is_causal,
)
attn_output = attn_output.transpose(1, 2).contiguous()
attn_output = attn_output.view(bsz, q_len, -1)
attn_output = self.o_proj(attn_output)
else:
# do attention with sage attention here
attn_output = attn_output.transpose(1, 2).contiguous()
attn_output = attn_output.view(bsz, q_len, -1)
attn_output = self.o_proj(attn_output)
return attn_output, None, past_key_value
# load llama model
model = AutoModelForCausalLM.from_pretrained(...)
for layer in model.model.layers:
layer.self_attn.forward = MethodType(LlamaSageAttnForward, layer.self_attn)
Replace the # do attention with sage attention part with the sage attention and this is supposed to work on transformers==4.45.2.
| gharchive/issue | 2024-10-24T02:31:28 | 2025-04-01T06:46:00.487338 | {
"authors": [
"MeJerry215",
"jason-huang03",
"jt-zhang"
],
"repo": "thu-ml/SageAttention",
"url": "https://github.com/thu-ml/SageAttention/issues/21",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2145770826 | Caution: Please be suspicious of this project and code
This problem is very obvious and has been clearly pointed out by other questioners as well.
https://github.com/thuml/Anomaly-Transformer/issues/4
You can easily see the big difference before and after adding the suspicious code called "detection adjustment" in the above link.
The author gives the following incomprehensible and unclear answer to this:
It is a widely accepted practice in this field of academia.
->The two papers you link to as evidence are not published in official journals, and an author is credited with both papers.
If so, please provide review papers published in validated journals that support your claims.
Even if you are right, if tuning to improve a model's low performance to this high is a practice in academia, then that practice should be eliminated.
This adjustment useful in an actual industrial environment because it is possible to receive information and make adjustments in real time(?)
-> Remember the real time industrial data has no label!! Especially, when it comes to the anomaly detection!
Yes, I agree with you. The experiments of this paper is extremely suspicious.
Adding to what you listed above, one thing I found is that with the evaluation process used in the paper, even a (any) random model (randomly initialized model without any training!) yields a good performance.
You can test it simply by commenting out the torch.load for checkpoint in the test function.
This problem is very obvious and has been clearly pointed out by other questioners as well. #4 You can easily see the big difference before and after adding the suspicious code called "detection adjustment" in the above link.
The author gives the following incomprehensible and unclear answer to this:
It is a widely accepted practice in this field of academia.
->The two papers you link to as evidence are not published in official journals, and an author is credited with both papers. If so, please provide review papers published in validated journals that support your claims. Even if you are right, if tuning to improve a model's low performance to this high is a practice in academia, then that practice should be eliminated.
This adjustment useful in an actual industrial environment because it is possible to receive information and make adjustments in real time(?)
-> Remember the real time industrial data has no label!! Especially, when it comes to the anomaly detection!
Yes, I agree with your points.
Adding to what you listed above, one thing I found is that with the evaluation method used in the work, even a (any) random model (randomly initialized model without training!) can yield a good performance as well. You can test it simply by commenting out the torch.load lines for loading checkpoint model in the test function.
Plus, leveraging the anomaly labels from the "testing dataset" when computing anomaly threshold seems to be simply just wrong regardless of whether some previous works did it in the same way or not (it cannot justify the validity of the method..).
Unfortunately, this problem is being propagating in the community as I found a few other works that adapt this evaluation method for anomaly task, where exactly the same issue as this work are found in as well. I believe the authors of the works have noticed the incorrectness of the method.
Be aware of the works (I found two below) that followed the same evaluation method.
"TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis", ICLR 2023.
"One Fits All: Power General Time Series Analysis by Pretrained LM,", NeurIPS, 2023.
I am not specifically being suspicious about the proposal itself in the work, but just the proposal needs to be validated properly.
I agree with your point of view; the arguments in the references mentioned by the author are not sufficient. Given the rigor of scientific research, I believe this method should not be continued.
The views in the original text are as follows:
“”“
In real applications, the human operators generally do not care about the point-wise metrics. It is acceptable for an algorithm to trigger an alert for any point in a contiguous anomaly segment, if the delay is not too long. Some metrics for anomaly detection have been proposed to accommodate this preference, e.g., [22], but most are not widely accepted, likely because they are too complicated. We instead use a simple strategy: if any point in an anomaly segment in the ground truth can be detected by a chosen threshold, we say this segment is detected correctly, and all points in this segment are treated as if they can be detected by this threshold. Meanwhile, the points outside the anomaly segments are treated as usual. The precision, recall, AUC, F-score and best F-score are then computed accordingly. This approach is illustrated in Fig 7。
“””
| gharchive/issue | 2024-02-21T04:00:09 | 2025-04-01T06:46:00.500665 | {
"authors": [
"caocaort",
"cooma04",
"subeom527"
],
"repo": "thuml/Anomaly-Transformer",
"url": "https://github.com/thuml/Anomaly-Transformer/issues/65",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
480029854 | Tutorial and windows build?
Can please somebody record a short video tut (just show what you doing, dosnt necessery to speak/write)
Figure it out
Installed on win10 using anaconda3 following installation instructions provide by authors with changes to runGan.py cmd as referred to in forum(add shell=True)
https://github.com/Sanaxen/tecoGAN_app
| gharchive/issue | 2019-08-13T08:38:15 | 2025-04-01T06:46:00.510747 | {
"authors": [
"EdisonBing",
"aakh1361",
"shadowzoom"
],
"repo": "thunil/TecoGAN",
"url": "https://github.com/thunil/TecoGAN/issues/24",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
341641459 | MUICSS's Radio component for React is not working with Dio.js - should property with undefined values be assigned?
This is a bit complicated to reproduce in SandboxCode because I think it doesn't support a way to define webpack aliases. Here's how it works with React:
https://codesandbox.io/s/qlk44v58x6
If you try that in a local project and add proper alias to react and react-dom to dio.js you will notice that the radios are ignored by Dio.js. All other MUICSS components I've tried so far have worked seamless with Dio.js, but this one seems to be having issues with Dio.js.
Here's the source-code for the component:
https://github.com/muicss/mui/blob/master/src/react/radio.jsx
I've copied that source locally and removed the "form" attribute here to make it work:
https://github.com/muicss/mui/blob/master/src/react/radio.jsx#L41
I suspect React won't touch this attribute if its value is undefined and maybe Dio.js should do the same, what do you think?
Closed this and opened #68 since it's easier to reproduce it there.
| gharchive/issue | 2018-07-16T19:09:48 | 2025-04-01T06:46:00.559184 | {
"authors": [
"rosenfeld"
],
"repo": "thysultan/dio.js",
"url": "https://github.com/thysultan/dio.js/issues/67",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2102675679 | v2.9.4: update flexible savings endpoints
Replaced some deprecated endpoints for updated flexible savings ones. (old endpoints returned "This endpoint has been deprecated, please remove as soon as possible."
Summary
changed end points
Additional Information
Replaced some deprecated endpoints for updated flexible savings ones. (old endpoints returned "This endpoint has been deprecated, please remove as soon as possible."
Done! Package JSON updated
As always thanks for the fast response and review! :)
Np, thanks for helping improve the lib :)
| gharchive/pull-request | 2024-01-26T18:15:39 | 2025-04-01T06:46:00.577762 | {
"authors": [
"MartoMcfly",
"tiagosiebler"
],
"repo": "tiagosiebler/binance",
"url": "https://github.com/tiagosiebler/binance/pull/389",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
607027092 | There were unauthenticated packages and -y was used without --allow-unauthenticated
Hi. I got this error after apt get update ,
fixed with --force-yes but I guess there is an issue in the base docker file ?
WARNING: The following packages cannot be authenticated!
libssl1.0.2 krb5-locales libgmp10 libnettle6 libhogweed4 libidn11 libffi6
libp11-kit0 libtasn1-6 libgnutls30 libkeyutils1 libkrb5support0 libk5crypto3
libkrb5-3 libgssapi-krb5-2 libsasl2-modules-db libsasl2-2 libldap-common
libldap-2.4-2 openssl ca-certificates libunistring0 libidn2-0 libnghttp2-14
libpsl5 librtmp1 libssh2-1 libcurl3 curl libsasl2-modules publicsuffix
E: There were unauthenticated packages and -y was used without --allow-unauthenticated
The command '/bin/sh -c apt-get install -y curl' returned a non-zero code: 100
# Base on "tiangolo/node-frontend:10"
FROM tiangolo/node-frontend:10 as build-stage
WORKDIR /app
COPY package*.json /app/
RUN npm install
COPY ./ /app/
ARG configuration=production
RUN npm run build -- --output-path=./dist/out --configuration $configuration
# Stage 1, based on Nginx, to have only the compiled app, ready for production with Nginx
FROM nginx:1.15
RUN apt-get update
RUN apt-get install -y --force-yes curl
# Copy dist
COPY --from=build-stage /app/entrypoint.sh /
COPY --from=build-stage /app/dist/out/ /usr/share/nginx/html
# Copy default nginx.conf
COPY --from=build-stage /nginx.conf /etc/nginx/conf.d/default.conf
WORKDIR /usr/share/nginx/html
CMD ["/bin/bash", "-c", "/entrypoint.sh && exec nginx -g 'daemon off;'"]
Thanks! And sorry for the long delay. :see_no_evil:
I no longer use nor recommend this Docker image, I think you are much better off building a Docker image yourself, so I updated all the docs to explain how to do it: https://github.com/tiangolo/node-frontend
The only extra step is creating and copying a simple nginx.conf file that you can copy-paste.
Given that, I'll now close this one. But thanks for the interest! :coffee:
| gharchive/issue | 2020-04-26T14:34:05 | 2025-04-01T06:46:00.581057 | {
"authors": [
"johnico",
"tiangolo"
],
"repo": "tiangolo/node-frontend",
"url": "https://github.com/tiangolo/node-frontend/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2075178158 | Implement enzyme constraints with the help of a pseudo reaction and a pseudo metabolite. Remove cached files
See title.
I also removed all files that were stored in the _cache folder. Some of them had special characters and caused problems with git syncing.
There hasn't been any news for a while on this Pull Request. I believe the proposed changes could benefit ECMpy and be appreciated by the community. Please consider testing and accepting.
| gharchive/pull-request | 2024-01-10T20:46:56 | 2025-04-01T06:46:00.640398 | {
"authors": [
"VonAlphaBisZulu"
],
"repo": "tibbdc/ECMpy",
"url": "https://github.com/tibbdc/ECMpy/pull/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1936976329 | 首页filter 刷新问题
插件的展示顺序应该排序固定一下,否则只要tw的tiddler有状态变化, filter就会重新刷新
其实是因为加了shuffle
应该是为了随机推荐,因为现在没有统计下载量
| gharchive/issue | 2023-10-11T06:46:03 | 2025-04-01T06:46:00.664294 | {
"authors": [
"Gk0Wk",
"linonetwo",
"oeyoews"
],
"repo": "tiddly-gittly/TiddlyWiki-CPL",
"url": "https://github.com/tiddly-gittly/TiddlyWiki-CPL/issues/148",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1782727848 | 条目标题中有空格时,会生成多余标题
比如这种,原条目是新条目 1,但添加子条目的时候就会变成三个标签了。
官方的添加子条目作为标签不会这样,估计要研究一下了。
<$button>
<$set name="tags" filter="[<currentTiddler>] [enlist{$:/plugins/linonetwo/intention-tower-knowledge-graph/Config/task-tag}]">
<$action-createtiddler $message="tm-new-tiddler" tags=<<tags>>
tmo_taskState="New">
<$action-sendmessage $message="tm-edit-tiddler" $param=<<createTiddler-title>>/>
</$action-createtiddler>
{{$:/plugins/linonetwo/intention-tower-knowledge-graph/Images/Task}} 子任务
</$set>
</$button>
还是官方厉害啊,没想到一抄就成功实现了。我把这段代码记到太微中文社区里去。
原来是先放到变量里
| gharchive/issue | 2023-06-30T15:27:10 | 2025-04-01T06:46:00.666550 | {
"authors": [
"dongrentianyu",
"linonetwo"
],
"repo": "tiddly-gittly/intention-tower-knowledge-graph",
"url": "https://github.com/tiddly-gittly/intention-tower-knowledge-graph/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
121046740 | Tandem v2 cleanup
WIP: Do not merge.
A couple of leftover tasks before we can push to prod.
<2> [uploader] Tandem v2 cleanup
Closing, because I once again used the wrong merge target.
| gharchive/pull-request | 2015-12-08T16:37:26 | 2025-04-01T06:46:00.668433 | {
"authors": [
"gniezen"
],
"repo": "tidepool-org/chrome-uploader",
"url": "https://github.com/tidepool-org/chrome-uploader/pull/214",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1044534902 | port path distance calculator
Description
path distance calculatorの.auto porting
autoware_utilsのポーティング待ち
Review Procedure
Remarks
Pre-Review Checklist for the PR Author
PR Author should check the checkboxes below when creating the PR.
[ ] Read pull request guidelines
[ ] Code follows coding guidelines
[ ] Assign PR to reviewer
Checklist for the PR Reviewer
Reviewers should check the checkboxes below before approval.
[ ] Commits are properly organized and messages are according to the guideline
[ ] Code follows coding guidelines
[ ] (Optional) Unit tests have been written for new behavior
[ ] PR title describes the changes
Post-Review Checklist for the PR Author
PR Author should check the checkboxes below before merging.
[ ] All open points are addressed and tracked via issues or tickets
[ ] Write release notes
CI Checks
Build and test for PR / build-and-test-pr: Required to pass before the merge.
Build and test for PR / clang-tidy-pr: NOT required to pass before the merge. It is up to the reviewer(s). Found false positives? See the guidelines.
Check spelling: NOT required to pass before the merge. It is up to the reviewer(s). See here if you want to add some words to the spell check dictionary.
Note:
Merge after https://github.com/tier4/AutowareArchitectureProposal.iv/pull/489
@kenji-miyake
follow to recent commit so please merge this if this is ok
Starting >>> path_distance_calculator
Finished <<< path_distance_calculator [3.49s]
Summary: 1 package finished [4.02s]
@taikitanaka3 Wait, I have to remove COLCON_IGNORE.
@TakaHoribe
merge ready desu!
| gharchive/pull-request | 2021-11-04T09:51:12 | 2025-04-01T06:46:00.751901 | {
"authors": [
"taikitanaka3",
"takayuki5168"
],
"repo": "tier4/AutowareArchitectureProposal.iv",
"url": "https://github.com/tier4/AutowareArchitectureProposal.iv/pull/498",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1114541854 | 🛑 PrivateBin is down
In a2ab8ba, PrivateBin (https://privatebin.my.id) was down:
HTTP code: 0
Response time: 0 ms
Resolved: PrivateBin is back up in d1f2a6c.
| gharchive/issue | 2022-01-26T01:52:05 | 2025-04-01T06:46:00.762604 | {
"authors": [
"tigefa4u"
],
"repo": "tigefa4u/status",
"url": "https://github.com/tigefa4u/status/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1644921964 | 背景色设置无效
我设置了animatedBackgroundColor,animatedColor 颜色之后,在骨架层上下滑动,会出先白色背景,旧版通过closeCache解决,在新版中貌似无效。
/
我设置了animatedBackgroundColor,animatedColor 颜色之后,在骨架层上下滑动,会出先白色背景,旧版通过closeCache解决,在新版中貌似无效。
貌似无效是指?能提供具体场景吗?
在你提供的demo tableView 中的 卡片布局。当我设置了animatedBackgroundColor,animatedColor不是白色,其他颜色。 然后再上下滑动的过程中会出现白色块
| gharchive/issue | 2023-03-29T02:15:53 | 2025-04-01T06:46:00.764660 | {
"authors": [
"XiaoHanL",
"tigerAndBull"
],
"repo": "tigerAndBull/TABAnimated",
"url": "https://github.com/tigerAndBull/TABAnimated/issues/220",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
340554534 | Added support for s390x
Added support for s390x by creating the s390x specific dockerfile which uses s390x/golang:1.9.2 as BASEIMAGE.
Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
@fasaxc could you please review above changes.
| gharchive/pull-request | 2018-07-12T09:16:10 | 2025-04-01T06:46:00.767385 | {
"authors": [
"CLAassistant",
"chakravarthy392"
],
"repo": "tigera/docker-protobuf",
"url": "https://github.com/tigera/docker-protobuf/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1553593394 | DOCS-1202: Confines Algolia crawlers to Docusaurus-native doc sets
https://tigera.atlassian.net/browse/DOCS-1202
The crawler crawls all of the old proxied sites. It's currently taking about 4 hours.
This config change restricts the crawler to /calico/ /calico-enterprise/ and /calico-cloud/.
With this change, the crawler takes minutes.
@denis-alkhelali I've tested this by running the crawler, and it seems to work. Can you see any problems with this?
@ctauchen perfect solution!
| gharchive/pull-request | 2023-01-23T18:54:01 | 2025-04-01T06:46:00.769828 | {
"authors": [
"ctauchen",
"denis-alkhelali"
],
"repo": "tigera/docs",
"url": "https://github.com/tigera/docs/pull/241",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2050948371 | [RS-1592] - Changes WAF from Coraza back to ModSecurity.
Description
This PR reverts the PR which changed from using ModSecurity to Coraza as the WAF engine. This is being done because WAF with Coraza doesn't currently work and is holding up Enterprise v3.19-EP1.
For PR author
[ ] Tests for change.
[ ] If changing pkg/apis/, run make gen-files
[ ] If changing versions, run make gen-versions
For PR reviewers
A note for code reviewers - all pull requests must have the following:
[ ] Milestone set according to targeted release.
[ ] Appropriate labels:
kind/bug if this is a bugfix.
kind/enhancement if this is a a new feature.
enterprise if this PR applies to Calico Enterprise only.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.Mike Stephen seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account.You have signed the CLA already but the status is still pending? Let us recheck it.
This is no longer needed.
| gharchive/pull-request | 2023-12-20T17:01:39 | 2025-04-01T06:46:00.775843 | {
"authors": [
"CLAassistant",
"mikestephen",
"tmjd"
],
"repo": "tigera/operator",
"url": "https://github.com/tigera/operator/pull/3078",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1402012341 | lsm_forest_fuzz: 11035627816572296482
At 2a134ecea5b546e138a2e781c0b8df50d981557b.
thread 18024 panic: reached unreachable code
/nix/store/g94blv8pk5z0h4maivk3ymcrafrcbwl8-zig-0.9.1/lib/zig/std/debug.zig:225:14: 0x23fa7b in std.debug.assert (lsm_forest_fuzz)
if (!ok) unreachable; // assertion failure
^
/home/jamie/tigerbeetle/src/lsm/level_iterator.zig:284:23: 0x283a3b in lsm.level_iterator.LevelIteratorType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage).peek (lsm_forest_fuzz)
assert(it.buffered_all_values());
^
/home/jamie/tigerbeetle/src/lsm/compaction.zig:154:52: 0x283775 in lsm.compaction.MergeStreamSelector.peek (lsm_forest_fuzz)
1 => compaction.iterator_b.peek(),
^
/home/jamie/tigerbeetle/src/lsm/k_way_merge.zig:105:28: 0x4754b8 in lsm.k_way_merge.KWayMergeIterator(lsm.compaction.CompactionType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage,lsm.table_immutable.TableImmutableIteratorType),u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,2,lsm.compaction.MergeStreamSelector.peek,lsm.compaction.MergeStreamSelector.pop,lsm.compaction.MergeStreamSelector.precedence).pop_internal (lsm_forest_fuzz)
if (stream_peek(it.context, root)) |key| {
^
/home/jamie/tigerbeetle/src/lsm/k_way_merge.zig:80:35: 0x4239c2 in lsm.k_way_merge.KWayMergeIterator(lsm.compaction.CompactionType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage,lsm.table_immutable.TableImmutableIteratorType),u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,2,lsm.compaction.MergeStreamSelector.peek,lsm.compaction.MergeStreamSelector.pop,lsm.compaction.MergeStreamSelector.precedence).pop (lsm_forest_fuzz)
while (it.pop_internal()) |value| {
^
/home/jamie/tigerbeetle/src/lsm/compaction.zig:485:49: 0x3b6ca6 in lsm.compaction.CompactionType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage,lsm.table_immutable.TableImmutableIteratorType).cpu_merge (lsm_forest_fuzz)
const value = merge_iterator.pop() orelse break;
^
/home/jamie/tigerbeetle/src/lsm/compaction.zig:455:37: 0x385a8e in lsm.compaction.CompactionType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage,lsm.table_immutable.TableImmutableIteratorType).cpu_merge_start (lsm_forest_fuzz)
compaction.cpu_merge();
^
/home/jamie/tigerbeetle/src/lsm/compaction.zig:434:71: 0x367fa5 in lsm.compaction.CompactionType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage,lsm.table_immutable.TableImmutableIteratorType).io_finish (lsm_forest_fuzz)
if (compaction.io_pending == 0) compaction.cpu_merge_start();
^
/home/jamie/tigerbeetle/src/lsm/compaction.zig:338:33: 0x367150 in lsm.compaction.CompactionType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage,lsm.table_immutable.TableImmutableIteratorType).iterator_b_io_callback (lsm_forest_fuzz)
compaction.io_finish();
^
/home/jamie/tigerbeetle/src/lsm/level_iterator.zig:247:28: 0x385760 in lsm.level_iterator.LevelIteratorType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage).table_iterator_callback (lsm_forest_fuzz)
it.callback(it);
^
/home/jamie/tigerbeetle/src/lsm/table_iterator.zig:229:29: 0x3b7909 in lsm.table_iterator.TableIteratorType(lsm.table.TableType(u64,tigerbeetle.Account,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).compare_keys,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).key_from_value,18446744073709551615,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone,lsm.groove.ObjectTreeHelpers(tigerbeetle.Account).tombstone_from_key),test.storage.Storage).on_read (lsm_forest_fuzz)
it.read_done(it);
^
/home/jamie/tigerbeetle/src/lsm/grid.zig:414:30: 0x2a7488 in lsm.grid.GridType(test.storage.Storage).start_read (lsm_forest_fuzz)
read.callback(read, block);
^
/home/jamie/tigerbeetle/src/lsm/grid.zig:176:32: 0x29d4db in lsm.grid.GridType(test.storage.Storage).tick (lsm_forest_fuzz)
grid.start_read(read);
^
/home/jamie/tigerbeetle/src/lsm/forest_fuzz.zig:105:22: 0x29179b in Environment.tick (lsm_forest_fuzz)
env.grid.tick();
^
/home/jamie/tigerbeetle/src/lsm/forest_fuzz.zig:112:52: 0x27f004 in Environment.tick_until_state_change (lsm_forest_fuzz)
while (env.state == current_state) env.tick();
^
/home/jamie/tigerbeetle/src/lsm/forest_fuzz.zig:163:36: 0x291b53 in Environment.compact (lsm_forest_fuzz)
env.tick_until_state_change(.forest_compacting, .forest_open);
^
/home/jamie/tigerbeetle/src/lsm/forest_fuzz.zig:211:32: 0x27b067 in Environment.run (lsm_forest_fuzz)
env.compact(op);
^
/home/jamie/tigerbeetle/src/lsm/forest_fuzz.zig:264:24: 0x2721b9 in fuzz (lsm_forest_fuzz)
try Environment.run(&storage, fuzz_ops);
^
/home/jamie/tigerbeetle/src/lsm/forest_fuzz.zig:414:13: 0x269c26 in main (lsm_forest_fuzz)
try fuzz(fuzz_ops);
^
/nix/store/g94blv8pk5z0h4maivk3ymcrafrcbwl8-zig-0.9.1/lib/zig/std/start.zig:561:37: 0x262d0a in std.start.callMain (lsm_forest_fuzz)
const result = root.main() catch |err| {
^
/nix/store/g94blv8pk5z0h4maivk3ymcrafrcbwl8-zig-0.9.1/lib/zig/std/start.zig:495:12: 0x2438fe in std.start.callMainWithArgs (lsm_forest_fuzz)
return @call(.{ .modifier = .always_inline }, callMain, .{});
^
/nix/store/g94blv8pk5z0h4maivk3ymcrafrcbwl8-zig-0.9.1/lib/zig/std/start.zig:409:17: 0x242996 in std.start.posixCallMainAndExit (lsm_forest_fuzz)
std.os.exit(@call(.{ .modifier = .always_inline }, callMainWithArgs, .{ argc, argv, envp }));
^
/nix/store/g94blv8pk5z0h4maivk3ymcrafrcbwl8-zig-0.9.1/lib/zig/std/start.zig:322:5: 0x2427a2 in std.start._start (lsm_forest_fuzz)
@call(.{ .modifier = .never_inline }, posixCallMainAndExit, .{});
^
The following command terminated unexpectedly:
cd /home/jamie/tigerbeetle && /home/jamie/tigerbeetle/zig-cache/o/7c816c2d3f9fb064b650a8d398257bf3/lsm_forest_fuzz --seed 11035627816572296482
error: the following build command failed with exit code 1:
/home/jamie/tigerbeetle/zig-cache/o/04e062ed50064d3f166819b2149e5107/build /nix/store/g94blv8pk5z0h4maivk3ymcrafrcbwl8-zig-0.9.1/bin/zig /home/jamie/tigerbeetle /home/jamie/tigerbeetle/zig-cache /home/jamie/.cache/zig lsm_forest_fuzz -- --seed 11035627816572296482
Possibly related to the heap update order issue; see https://github.com/tigerbeetledb/tigerbeetle/pull/177.
| gharchive/issue | 2022-10-08T17:33:34 | 2025-04-01T06:46:00.779735 | {
"authors": [
"jamii",
"sentientwaffle"
],
"repo": "tigerbeetledb/tigerbeetle",
"url": "https://github.com/tigerbeetledb/tigerbeetle/issues/186",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
753387031 | pass options to enable container without having to go through cli prompt
just some thoughts. i'm coming from laradock and i was looking for something lighter. takeout seems to be a great option, but from my tests and what i can get from the code, we have to go through the prompts. that makes it hard to automatize.
to give you an idea of my workflow, i run one command through itermocil, and it opens a bunch of terminal panels, goes into the right directories, boot up docker containers with the right options, run npm run watch in them, etc... basically i call one command, i wait for 20 seconds, and my project is up with the right version of PHP, mysql, redis, etc, etc. that would be pretty nice to be able to do this with takeout. any thought?
I love that idea! :) Do you think you could propose a syntax?
I love that idea! :) Do you think you could propose a syntax?
(Sorry for the delay in response. It's been... quite a year.)
(Sorry for the delay in response. It's been... quite a year.)
I can help with a PR if there is any specific syntax in mind.
no worries.
the syntax, good question. kinda subjective i think. personally i'm a big fan of two things: intuitive, but also following a used consensus. i guess it'd be nice to have both. like the typical -p for port, etc..., and something more intuitive too (in the case of port, well, like --port, which in this case is also the consensus 😅️). i'll do more research and come back to you. (currently i'm automatically printing the port i need in a tab panel and starting takeout in another one. i still have to choose manually but i have all the info i need so it's still pretty straightforward.)
no worries.
the syntax, good question. kinda subjective i think. personally i'm a big fan of two things: intuitive, but also following a used consensus. i guess it'd be nice to have both. like the typical -p for port, etc..., and something more intuitive too (in the case of port, well, like --port, which in this case is also the consensus 😅️). i'll do more research and come back to you. (currently i'm automatically printing the port i need in a tab panel and starting takeout in another one. i still have to choose manually but i have all the info i need so it's still pretty straightforward.)
Maybe a possible solution to this would be something like takeout taste-test, which is almost the same as takeout enable in that it runs you through the prompts, but then instead of actually executing the docker run command it just echos it out to the terminal. Someone could then just copy-paste that into their automation scripts and modify as needed.
@wunc We could certainly enable that using something like --dry-run to any of the commands. That doesn't perfectly solve this issue, but it would help.
i'll try to come up with something when i have a bit more time. currently not the priority (for me). thanks!
I love this suggestion. But it feels very similar to what docker-compose does. Once we start getting into configuration files to orchestrate containers, it might be too complex for most takeout users.
fair enough.
the idea though was not to mess with configuration files, but to give the ability to pass more arguments.
currently the way i handle my workflow is echoing the ports i'm using for the current project, and starting the containers manually:
the level of friction is still acceptable so i hadn't dig yet. guess closing this issue removes this from my todo list 😂️
| gharchive/issue | 2020-11-30T11:14:17 | 2025-04-01T06:46:00.787933 | {
"authors": [
"godbout",
"josecanhelp",
"mabadir",
"mattstauffer",
"wunc"
],
"repo": "tighten/takeout",
"url": "https://github.com/tighten/takeout/issues/189",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
330013887 | Why are assets compiled in to my source directory?
From the docs:
By default, once your assets are compiled, they will be placed in their corresponding directories, directly under the source folder:
What is the reasoning behind the built assets being duplicated in source and also in the build_local folder?
Ideally I would only have the built assets residing in the build_local or build_production folders which I can ignore from git.
This is because Webpack runs before the Jigsaw build.
Webpack needs an output path (mix.setPublicPath('source/assets/') in webpack.mix.js), but there's no build_local or build_production at that point in time.
You could simply gitignore the source/assets directory if you don't want it in version control.
Okay I get that, thanks.
Is it possible to build the client assets outside the source directory? I tried the following:
// webpack.mix.js
mix.setPublicPath('dist/');
// config.php
return [
'build' => [
'source' => 'dist'
],
...
];
The build was completing but the HTML files weren't being generated. It seemed like maybe the 'source' in config.php relates to where the view assets and webpack assets are?
It feels like I need a setting like:
// config.php
return [
'build' => [
'publicPath' => 'dist'
]
...
];
Thanks for your help, I really like the project so far!
The source key in config.php tells Jigsaw where your source files are located (layouts, partials, your .md files), so leave it as 'source' => 'source',.
publicPath won't do anything, the key you're looking for in there is destination.
So with that in mind:
// config.php
'build' => [
'source' => 'source',
'destination' => 'dist',
],
That would output the Jigsaw build where you want it.
But Jigsaw cleans up the destination directory when it builds, and because it runs after Webpack I think that would result in your dist/assets directory being removed. Haven't tested this though, try it out :)
Ah sorry, you're right, I did have 'destination' => 'dist'. I think the problem is I don't know how to configure where jigsaw looks for assets when it runs.
What I'm trying to achieve is:
Webpack runs first and outputs my built JS/SASS to /dist
Jigsaw should run over the views in source but look for the mix-manifest.json in /dist
Clears /build_local
Outputs the compiled views + copies built webpack assets from /dist in to /build_local
Do you know if this configuration is possible? The reason being that I only ever want /source to actually contain 'source' files to prevent confusion.
@hellocosmin's description of why it builds assets to source/assets is correct; Webpack runs first to compile assets, then Jigsaw runs its build process and simply copies the compiled assets to the build folder.
I can't think of a way to accomplish what you're hoping, at least without some additional Webpack build step after jigsaw build is run. In your step 2), Jigsaw only parses and copies whatever files it finds in source.
To help prevent confusion in /source/assets, I typically have a /build directory in there that non-static assets (JS, SASS, Less files, etc.) in /source/_assets get built to, and I gitignore that build directory. That makes it a little clearer that the generated files in /source/assets/build shouldn't be edited manually. If you do that, you'll need to also update your publicPath in webpack.mix.js to:
mix.setPublicPath('source/assets/build');
...and update Mix references in your templates:
<link rel="stylesheet" type="text/css" href="{{ mix('css/style.css', 'assets/build') }}">
Thanks @damiani, that looks pretty similar to the solution I have settled on for now.
| gharchive/issue | 2018-06-06T20:08:05 | 2025-04-01T06:46:00.799471 | {
"authors": [
"damiani",
"hellocosmin",
"jameshulse"
],
"repo": "tightenco/jigsaw",
"url": "https://github.com/tightenco/jigsaw/issues/224",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
162268636 | simple form generations fils
when having simple form in the gemfile the generators file with the below message.
without the simple form gem in the gemfile the generation of the templates is working fine.
Running via Spring preloader in process 34125
invoke Devise::I18n::SharedViewsGenerator
exist app/views/devise/shared
identical app/views/devise/shared/_links.html.erb
invoke Devise::I18n::MailerViewsGenerator
exist app/views/devise/mailer
identical app/views/devise/mailer/confirmation_instructions.html.erb
identical app/views/devise/mailer/password_change.html.erb
identical app/views/devise/mailer/reset_password_instructions.html.erb
identical app/views/devise/mailer/unlock_instructions.html.erb
invoke i18n:simple_form_for
Could not find "confirmations" in any of your source paths. Your current source paths are:
/Users/xxx/yyy/lib/templates/devise/simple_form_for
/Users/xxx/.rbenv/versions/2.2.4/lib/ruby/gems/2.2.0/gems/devise-i18n-1.1.0/lib/generators/devise/templates/simple_form_for
Same issue here.
In $GEM_HOME/gems/devise-i18n-1.1.0/lib/generators/devise there is only i18n folder (no templates folder).
In git repository there are folders under devise-i18n/lib/generators/devise/templates/simple_form_for.
Maybe a problem with packed gem?
Those files were indeed missed in the list of files to include in the packaged gem. It should be fixed for the next release.
Any chance for a new release? This is blocking a thing I'm writing over at https://github.com/thredded
@JasonBarnabe Thank you! 🎉
| gharchive/issue | 2016-06-25T07:24:26 | 2025-04-01T06:46:00.803021 | {
"authors": [
"Gee-Bee",
"JasonBarnabe",
"glebm",
"koenhandekyn"
],
"repo": "tigrish/devise-i18n",
"url": "https://github.com/tigrish/devise-i18n/issues/161",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1042350889 | raftstore: add a new worker to dynamically update pool size (#11159)
Signed-off-by: Wenbo Zhang ethercflow@gmail.com
What problem does this PR solve?
Issue Number: close #11159
What is changed and how it works?
Proposal: #11159
What's Changed: #11159
Related changes
No
Check List
Tests
No code
Side effects
No
Release note
None.
@NingLin-P @5kbpers PTAL
@5kbpers @NingLin-P Fixed all comments, PTAL again, thanks!
/merge
/run-test
| gharchive/pull-request | 2021-11-02T13:30:05 | 2025-04-01T06:46:00.813894 | {
"authors": [
"5kbpers",
"ethercflow"
],
"repo": "tikv/tikv",
"url": "https://github.com/tikv/tikv/pull/11211",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
643573756 | engine_rocks: support change rate limit dynamically
What problem does this PR solve?
Part of https://github.com/tikv/tikv/issues/6484
What is changed and how it works?
What's Changed:
This PR is trying to support changing rate-bytes-per-sec dynamically. Related to tikv/rust-rocksdb#510
Related changes
Need to cherry-pick to the release branch
Check List
Tests
Unit test
Release note
Support changing rate-bytes-per-sec dynamically
/run-all-tests
@yiwu-arbug can it be picked to 4.0?
@Connor1996 are you able to help?
| gharchive/pull-request | 2020-06-23T06:38:35 | 2025-04-01T06:46:00.817528 | {
"authors": [
"BusyJay",
"rleungx",
"yiwu-arbug"
],
"repo": "tikv/tikv",
"url": "https://github.com/tikv/tikv/pull/8124",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
230131028 | cleanup invokeCallback
removed duplicating variables succeeded, failed
@bekzod your making many nice cleanups! This project and es6-promise are actually based on the same code base. And I try to keep both in sync, if you have spare cycles it would be very helpful if you could replicate applicable changes in es6-promise as well
@stefanpenner thanks, yeah sure thing will look into es6-promise
| gharchive/pull-request | 2017-05-20T05:13:02 | 2025-04-01T06:46:00.818901 | {
"authors": [
"bekzod",
"stefanpenner"
],
"repo": "tildeio/rsvp.js",
"url": "https://github.com/tildeio/rsvp.js/pull/464",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1348117807 | Remove dependency on bootstrap
Handling of column metadata can basically stay the same, but implementing it by hand removes a lot of uneccessary stuff.
This is actually not necessary as bootstrap provides a css file with only the layouting rules.
| gharchive/issue | 2022-08-23T15:12:31 | 2025-04-01T06:46:00.839616 | {
"authors": [
"tim-fabian"
],
"repo": "tim-fabian/ngx-material-entity",
"url": "https://github.com/tim-fabian/ngx-material-entity/issues/101",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
159174947 | Timber initializing too early
Expected behavior
Timber to initialize after all plugins or mu-plugins are able to hook into WordPress functions.
This bug is specific for timber-library (WP plugin), since initialization for timber/timber will have be done by the theme developer.
Actual behavior
Timber initializes before this happens (at inclusion). This is especially a problem in Multisite, because network plugins are loaded before muplugins_loaded is called.
Right now, plugins are forced to be a mu-plugin AND cannot use muplugins_loaded to register their actions. Possibly, Timber should even be initialized on setup_theme instead or at least at plugins_loaded.
Use case
Related to #1056: Timber\ImageHelper::add_constants() is run on inclusion as well (when initializing Timber). This does not allow plugins to hook into home_url() and change it.
The specific use-case would be when mapping a WordPress Multisite install to a domain. In this case, a filter should be used to change the home_url().
I'm not sure if the problem is as broad as "Timber initializing too early" or if it's a specific bug for this use-case.
Steps to reproduce behavior
For the specific use-case:
Use WordPress Multisite
Create a subdirectory install
Map this website to a domain by filtering home_url() or using a plugin that does this
Sideload an image (for example, by resizing an external image)
What version of WordPress, PHP and Timber are you using?
WordPress 4.5.2, PHP 7.0, Timber 1.0.4.
How did you install Timber? (for example, from GitHub, Composer/Packagist, WP.org?)
Composer via WPackagist.
+1
I've experienced the same issue when running Wordpress with Bedrock, it changes the default content directory and install location for Wordpress.
+1
Its also an issue if you have Timber library installed, but local version in your own theme. Timber plugin registers and initializes itself as soon as the plugin php file gets loaded, which causes issues if theme uses different version of Timber.
@haroldangenent @mahagr Is this still something you're actively working on or need addressed? Please respond in two weeks or we'll close this thread
Yes, nothing has changed so I believe this is still an issue when using the plugin. It should at least be documented that timber/timber should be used for multisites that use domain mapping.
I'm using bedrock with Timber and didn't had any problems. I'm also using domain mapping. The only difference is the fact I'm not using the plugin but the composer module.
I'm also using "roots/multisite-url-fixer": "*".
Can you try using the module and this mu-plugin.
It's no issue if you're using Composer package, but you really shouldn't install the plugin or you're in trouble when it updates and you're still using the old version in the theme.
From what I know Timber 2.0 will be available only as a composer package.
So I'm not sure about fixing those bugs. Let's wait for Jared's or Lukas
opinion.
śr., 7 lis 2018, 18:54: Matias Griese notifications@github.com napisał(a):
It's no issue if you're using Composer package, but you really shouldn't
install the plugin or you're in trouble when it updates and you're still
using the old version in the theme.
—
You are receiving this because you were assigned.
Reply to this email directly, view it on GitHub
https://github.com/timber/timber/issues/1058#issuecomment-436716661, or mute
the thread
https://github.com/notifications/unsubscribe-auth/ACO-OhBmbfR248R3ice4ZLWhjPrFILGrks5usx5kgaJpZM4IxAdu
.
From what I know Timber 2.0 will be available only as a composer package.
Yes, that’s right. Releasing Timber 2.0 with the plugin version would break so many websites because of all the breaking changes. And Timber probably wouldn’t be accepted as a new plugin for version 2, because it’s not a "real" plugin. That’s why from 2.0, it will only be available as a Composer package.
Because the issue here is a case that can easily be circumvented by using the Composer package, I wouldn’t invest the time to fix this for the plugin version that will be outdated soon. But we could add a hint in the documentation for 1.x. Pull requests welcome 🙂.
@gchtr Well, the issue is opposite. If you have Timber plugin installed, Timber 2.0 based themes will not work. You need to uninstall Timber plugin to get them to work.
@mahagr That’s exactly what I meant or tried to say. Maybe I wasn’t clear enough.
| gharchive/issue | 2016-06-08T14:17:14 | 2025-04-01T06:46:00.858780 | {
"authors": [
"ChrisManganaro",
"gchtr",
"haroldangenent",
"mahagr",
"palmiak"
],
"repo": "timber/timber",
"url": "https://github.com/timber/timber/issues/1058",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
444322933 | ACF custom menu fields not included in TimberMenu object
Expected behavior
When I add custom fields to a Menu object (not the MenuItem object), I would expect to find these custom fields in {{menu.field}} or menu.meta('field'). For example, I've added a custom field that includes the logo in the menu for easy changing.
Actual behavior
I can't find where these custom fields are stored. There seems to be no documentation on this, so researching is difficult.
I'm using Wordpress 5.2 with ACF 5.7.9 and Timber 1.9.4
Thanks in advance for any help! If i can find the solution, I'd like to try and update the documentation for this!
EDIT: Issue #2000, what an amazing community ❤️
Looking for an answer to this myself. These menu fields (not menu item fields) are stored in the wp_termmeta table. Per the docs:
WP stores each Menu as a term object in the wp_terms table. ACF will store all custom field values in the wp_termmeta table.
https://www.advancedcustomfields.com/resources/adding-fields-menus/
@bramkorsten @bobhuf thanks for the notes on this — looking into it now!
@bramkorsten @bobhuf this might be a very simple fix. Check out #2043 and please let me know if that resolves the issue for you (it did in my tests)
It does. Thanks, Jared!
Works like a charm! Looking good Jared 👌👌
| gharchive/issue | 2019-05-15T09:11:16 | 2025-04-01T06:46:00.863867 | {
"authors": [
"bobhuf",
"bramkorsten",
"jarednova"
],
"repo": "timber/timber",
"url": "https://github.com/timber/timber/issues/2000",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2688755616 | Cleaning up Playgamepanel.
I cleaned up the files and made sure to use clean architecture. I made sure that every button has a use case and controller as stated in the slides. Can someone please review this for me. Thanks.
also there are merge issues, needs to be fixed
| gharchive/pull-request | 2024-11-25T01:22:31 | 2025-04-01T06:46:00.870678 | {
"authors": [
"alexanderlee10",
"timchen0326"
],
"repo": "timchen0326/HoopsHub",
"url": "https://github.com/timchen0326/HoopsHub/pull/10",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2261553708 | Multiple widgets used the same GlobalKey
Getting GlobalKey Error
Multiple widgets used the same GlobalKey.
════════ Exception caught by widgets library ═══════════════════════════════════
Multiple widgets used the same GlobalKey.
════════════════════════════════════════════════════════════════════════════════
`The following assertion was thrown while finalizing the widget tree:
Multiple widgets used the same GlobalKey.
The key [GlobalKey#beb1f] was used by multiple widgets. The parents of those widgets were:
CustomPaint(renderObject: RenderCustomPaint#830c9 NEEDS-PAINT)
CustomPaint(renderObject: RenderCustomPaint#e5d2c NEEDS-PAINT)
A GlobalKey can only be specified on one widget at a time in the widget tree.
When the exception was thrown, this was the stack:
#6 BuildOwner.finalizeTree. (package:flutter/src/widgets/framework.dart:3177:11)
framework.dart:3177
#7 BuildOwner.finalizeTree (package:flutter/src/widgets/framework.dart:3259:8)
framework.dart:3259
#8 WidgetsBinding.drawFrame (package:flutter/src/widgets/binding.dart:992:19)
binding.dart:992
#9 RendererBinding._handlePersistentFrameCallback (package:flutter/src/rendering/binding.dart:448:5)
binding.dart:448
#10 SchedulerBinding._invokeFrameCallback (package:flutter/src/scheduler/binding.dart:1386:15)
binding.dart:1386
#11 SchedulerBinding.handleDrawFrame (package:flutter/src/scheduler/binding.dart:1311:9)
binding.dart:1311
#12 SchedulerBinding.scheduleWarmUpFrame. (package:flutter/src/scheduler/binding.dart:1034:7)
binding.dart:1034
#16 _RawReceivePort._handleMessage (dart:isolate-patch/isolate_patch.dart:184:12)
isolate_patch.dart:184
(elided 3 frames from class _Timer and dart:async-patch)`
How to reproduce:
SizedBox(
width: MediaQuery.of(context).orientation == Orientation.portrait ? MediaQuery.of(context).size.height - 120 : MediaQuery.of(context).size.width - 120,
height: MediaQuery.of(context).orientation == Orientation.portrait ? MediaQuery.of(context).size.width - 120 : MediaQuery.of(context).size.height - 140,
child: AspectRatio(
aspectRatio: 16 / 9,
child: Card(
clipBehavior: Clip.hardEdge,
surfaceTintColor: EmintColors.emerald,
elevation: 0.2,
child: Scribble(
notifier: notifier,
drawPen: true,
),
),
),
),
Hi, thanks for opening up an issue. Could you tell me which version of scribble you are on? I can't reproduce this on the latest version.
Also, is it possible, that you are using the same ScribbleNotifier for multiple Scribble widgets? That could lead to the issue you're describing
Hi,
We are using scribble: ^0.10.0+1 version,
We are not using multiple scribble widgets but screen orientations happen when building widget. Sharing sample code.
class SignatureView extends StatefulHookConsumerWidget {
const SignatureView({super.key});
@override
ConsumerState<ConsumerStatefulWidget> createState() => _SignatureViewState();
}
class _SignatureViewState extends ConsumerState<SignatureView> {
late ScribbleNotifier notifier;
@override
void initState() {
notifier = ScribbleNotifier();
super.initState();
}
@override
Widget build(BuildContext context) {
useEffect(() {
WidgetsBinding.instance.addPostFrameCallback((timeStamp) {
SystemChrome.setPreferredOrientations([
DeviceOrientation.landscapeRight,
DeviceOrientation.landscapeLeft,
]);
SystemChrome.setEnabledSystemUIMode(SystemUiMode.leanBack, overlays: []);
});
return () {
SystemChrome.setPreferredOrientations([
DeviceOrientation.portraitUp,
DeviceOrientation.landscapeLeft,
DeviceOrientation.portraitDown,
DeviceOrientation.landscapeRight,
]);
SystemChrome.setEnabledSystemUIMode(SystemUiMode.edgeToEdge, overlays: SystemUiOverlay.values);
};
}, []);
return Padding(
padding: const EdgeInsets.symmetric(horizontal: 20.0),
child: SafeArea(
child: Column(
children: [
SizedBox(
width: MediaQuery.of(context).orientation == Orientation.portrait ? MediaQuery.of(context).size.height - 120 : MediaQuery.of(context).size.width - 20,
height: MediaQuery.of(context).orientation == Orientation.portrait ? MediaQuery.of(context).size.width - 120 : MediaQuery.of(context).size.height - 120,
child: AspectRatio(
aspectRatio: 16 / 9,
child: Container(
clipBehavior: Clip.hardEdge,
decoration: ShapeDecoration(
shape: RoundedRectangleBorder(
side: BorderSide(width: 1, color: EmintColors.neutral.shade30),
borderRadius: BorderRadius.circular(12),
),
),
child: Scribble(
notifier: notifier,
drawPen: true,
),
),
),
),
Row(
mainAxisAlignment: MainAxisAlignment.center,
children: _buildActions(context, provider),
)
],
),
),
);
}
Any solution for this ? @timcreatedit
I'm also getting this same error.
@kushal2021 @anaghvj I still can't reproduce. Could you share a minimal example, ideally without any 3rd party dependencies?
This is the example I'm running now: When switching orientation, everything works fine, no exceptions in the console.
import 'package:flutter/material.dart';
import 'package:scribble/scribble.dart';
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({super.key});
// This widget is the root of your application.
@override
Widget build(BuildContext context) {
return MaterialApp(
title: 'Scribble',
theme: ThemeData.from(
colorScheme: ColorScheme.fromSeed(seedColor: Colors.purple)),
home: const SignatureView(),
);
}
}
class SignatureView extends StatefulWidget {
const SignatureView({super.key});
@override
State<StatefulWidget> createState() => _SignatureViewState();
}
class _SignatureViewState extends State<SignatureView> {
late ScribbleNotifier notifier;
@override
void initState() {
notifier = ScribbleNotifier();
super.initState();
}
@override
Widget build(BuildContext context) {
return Scaffold(
body: Padding(
padding: const EdgeInsets.symmetric(horizontal: 20.0),
child: SafeArea(
child: Column(
children: [
Text(MediaQuery.of(context).orientation.toString()),
SizedBox(
width:
MediaQuery.of(context).orientation == Orientation.portrait
? MediaQuery.of(context).size.height - 120
: MediaQuery.of(context).size.width - 20,
height:
MediaQuery.of(context).orientation == Orientation.portrait
? MediaQuery.of(context).size.width - 120
: MediaQuery.of(context).size.height - 120,
child: AspectRatio(
aspectRatio: 16 / 9,
child: Container(
clipBehavior: Clip.hardEdge,
decoration: ShapeDecoration(
shape: RoundedRectangleBorder(
side: BorderSide(width: 1, color: Colors.red),
borderRadius: BorderRadius.circular(12),
),
),
child: Scribble(
notifier: notifier,
drawPen: true,
),
),
),
),
],
),
),
),
);
}
}
I am not facing this error with minimal code, I am only facing this error with my project's code.
I'm implementing same functionality like instagram story create where user can select multiple images and can draw on particular images.
If you want then we can connect and trace the issue.
| gharchive/issue | 2024-04-24T15:18:31 | 2025-04-01T06:46:00.881239 | {
"authors": [
"anaghvj",
"kushal2021",
"timcreatedit"
],
"repo": "timcreatedit/scribble",
"url": "https://github.com/timcreatedit/scribble/issues/62",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2000523723 | after adding the database source, keep loading
Describe the bug
Steps to reproduce
adding a new database to local proton
set a source name and save
check the sync table dialog keep
As a result, no table is listed. Probably need to fix the show tables or show streams
https://www.metabase.com/docs/latest/databases/sync-scan
metabase will run some queries with up to 10000 rows in the limit, but for streams, it will load the future data, if the total row is less than 10000, then it will keep waiting.
No good idea how to fix it now.
This issue is resolved with 0.0.2+, by updating the JDBC driver
| gharchive/issue | 2023-11-18T18:12:16 | 2025-04-01T06:46:00.887300 | {
"authors": [
"jovezhong"
],
"repo": "timeplus-io/metabase-proton-driver",
"url": "https://github.com/timeplus-io/metabase-proton-driver/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2400554657 | Format README
The rendered README is not formatted well. The two footprints are placed in the same line.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
| gharchive/pull-request | 2024-07-10T11:51:00 | 2025-04-01T06:46:00.936482 | {
"authors": [
"CLAassistant",
"niebayes"
],
"repo": "timescale/tsbs",
"url": "https://github.com/timescale/tsbs/pull/256",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
396338536 | breaks with django 2.1
Need to add renderer=None to forms render function to fix. See https://docs.djangoproject.com/en/dev/internals/deprecation/#deprecation-removed-in-2-1
Looks like the merge request targetting this issue has been accepted; could you bump the version and update the pip entry, please?
+1
+1
Version 1.1.0 with Django 2.1 support has just been released :+1:
| gharchive/issue | 2019-01-07T03:50:20 | 2025-04-01T06:46:00.958347 | {
"authors": [
"assembledadam",
"emg36",
"jLebioda",
"neoflex",
"timheap"
],
"repo": "timheap/django-enumchoicefield",
"url": "https://github.com/timheap/django-enumchoicefield/issues/10",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
953282296 | gif to video transcoding
Is your feature request related to a problem? Please describe.
I am creating a personal site with a blog and projects and I need to include animations in addition to static images. I prefer to use compressed MP4 looping videos for displaying the animations because they are far smaller in file size than animated gifs.
Describe the solution you'd like
A built in transcoder to turn GIFs into to muted looping autoplaying MP4 videos.
Describe alternatives you've considered
Alternatively, my problem could be solved by just installing a video player component, such as react-player or next-videos. However I don't know how to install and use these components. I've tried and failed.
Additional context
This feature exists on this 11ty blog template: https://github.com/google/eleventy-high-performance-blog, more specifically:
https://github.com/google/eleventy-high-performance-blog/blob/main/_11ty/video-gif.js
https://github.com/google/eleventy-high-performance-blog/blob/main/_11ty/img-dim.js
Thanks for the feature request. It should be technically feasible by extending the next image component and passing gifs through ffmpeg like the _11ty example link above. react-player and next-videos are different and won't work.
I won't prioritise it though as you can convert the gif to mp4 externally and just add the video. Will keep this issue open if anyone is interested in working on it.
Oh, I'm such a noob. I didn't realize I could just paste in the html element directly into the MDX post. Yes, I will just convert the gifs externally and add the videos to the posts, as you suggested. Thanks!
No worries! It's still a nice idea and enhancement if we can do it automatically.
| gharchive/issue | 2021-07-26T20:59:12 | 2025-04-01T06:46:00.971607 | {
"authors": [
"shanewasley",
"timlrx"
],
"repo": "timlrx/tailwind-nextjs-starter-blog",
"url": "https://github.com/timlrx/tailwind-nextjs-starter-blog/issues/152",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2255935505 | Help with banner to visible in the homepage of the blog and inside
How to post blog with banner in the home page and even in the blog. With the author details like posting default blog with postlayout you know what i mean!
How to post blog with banner in the home page and even in the blog. With the author details like posting default blog with postlayout you know what i mean!
I want to change this
To something like this:
I don't really know what you are asking for but if it is small visual changes, you probably need to know a little JS and React to modify the code.
| gharchive/issue | 2024-04-22T08:36:01 | 2025-04-01T06:46:00.974328 | {
"authors": [
"heyjumanji",
"timlrx"
],
"repo": "timlrx/tailwind-nextjs-starter-blog",
"url": "https://github.com/timlrx/tailwind-nextjs-starter-blog/issues/899",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
284358565 | Frequent connection reset errors from Poloniex
Version 4.3.1
I am getting frequent "connection reset" errors at initialization time:
Exception in thread "main" org.knowm.xchange.exceptions.ExchangeException: java.net.SocketException: Connection reset
at org.knowm.xchange.BaseExchange.applySpecification(BaseExchange.java:113)
at org.knowm.xchange.ExchangeFactory.createExchange(ExchangeFactory.java:98)
at org.knowm.xchange.ExchangeFactory.createExchange(ExchangeFactory.java:147)
Caused by: java.net.SocketException: Connection reset
at java.base/java.net.SocketInputStream.read(SocketInputStream.java:210)
at java.base/java.net.SocketInputStream.read(SocketInputStream.java:141)
at java.base/sun.security.ssl.SSLSocketInputRecord.read(SSLSocketInputRecord.java:425)
at java.base/sun.security.ssl.SSLSocketInputRecord.decode(SSLSocketInputRecord.java:154)
at java.base/sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1031)
at java.base/sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:973)
at java.base/sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1402)
at java.base/sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1429)
at java.base/sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1413)
at java.base/sun.net.www.protocol.https.HttpsClient.afterConnect(HttpsClient.java:567)
at java.base/sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:185)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream0(HttpURLConnection.java:1581)
at java.base/sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1509)
at java.base/java.net.HttpURLConnection.getResponseCode(HttpURLConnection.java:527)
at java.base/sun.net.www.protocol.https.HttpsURLConnectionImpl.getResponseCode(HttpsURLConnectionImpl.java:329)
at si.mazi.rescu.HttpTemplate.receive(HttpTemplate.java:121)
at si.mazi.rescu.RestInvocationHandler.receiveAndMap(RestInvocationHandler.java:156)
at si.mazi.rescu.RestInvocationHandler.invoke(RestInvocationHandler.java:120)
at com.sun.proxy.$Proxy13.getCurrencyInfo(Unknown Source)
at org.knowm.xchange.poloniex.service.PoloniexMarketDataServiceRaw.getPoloniexCurrencyInfo(PoloniexMarketDataServiceRaw.java:36)
at org.knowm.xchange.poloniex.PoloniexExchange.remoteInit(PoloniexExchange.java:58)
at org.knowm.xchange.BaseExchange.applySpecification(BaseExchange.java:111)
... 4 more
Is this a rate-limiting thing on their end? Instability? Or some sort of bug in this library? Would it make sense to add a retry loop here?
It's tough to say. It seems like a network issue to me but they could have also changed something on their API, and Xchange hasn't yet been updated.
| gharchive/issue | 2017-12-24T11:54:56 | 2025-04-01T06:46:00.986271 | {
"authors": [
"cowwoc",
"timmolter"
],
"repo": "timmolter/XChange",
"url": "https://github.com/timmolter/XChange/issues/2042",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
639625200 | Опечатка в документации
Опишите баг
Опечатка в документации про бренчи
Что должно было случится
В пунтке про создание обработчика в примерах с росток и веткой есть опечатка в await ans("Окей, вывожу!"), должно быть выхожу т.к. производится выход из бренчи
Заполните поля:
OS: Windows 10
Version of vkbottle: 2.7.5
Так же в README лицензии указана GPL 3.0, когда судя по LICENCE лицензия на самом деле MIT
Спасибо за бдительность, скоро исправим!
| gharchive/issue | 2020-06-16T12:29:04 | 2025-04-01T06:46:00.990233 | {
"authors": [
"dwapper",
"timoniq"
],
"repo": "timoniq/vkbottle",
"url": "https://github.com/timoniq/vkbottle/issues/134",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
112485185 | #isort:imports-thirdparty removes imports
I am opening a new issue as it appears my comment in closed issue #177 has been missed. This still applies to 4.2.2
I just tried using those comments with isort 4.0 and adding either #isort:imports-thirdparty or #isort:imports-firstparty causes imports to be removed:
-#isort:imports-thirdparty
-import gobject
-import gtk
-from twisted.internet import gtk2reactor
-from twisted.internet.error import ReactorAlreadyInstalledError
-
try:
# Install twisted reactor, before any other modules import reactor.
reactor = gtk2reactor.install()
@@ -26,31 +20,6 @@
# Running unit tests so trial already installed a rector
pass
-#isort:imports-firstparty
-import deluge.common
-import deluge.component as component
...
Thanks for reporting, I'll be looking into this, sorry for the long wait I've been heads down working on hug: https://github.com/timothycrosley/hug
Thanks!
~Timothy
This has been fixed in develop here: https://github.com/timothycrosley/isort/pull/407/files and will make its way into the next release.
Thanks!
~Timothy
| gharchive/issue | 2015-10-21T00:05:40 | 2025-04-01T06:46:01.000856 | {
"authors": [
"cas--",
"timothycrosley"
],
"repo": "timothycrosley/isort",
"url": "https://github.com/timothycrosley/isort/issues/357",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
944767966 | New Design
This incorporates a number of features from the Tailwind starter and ditches the existing styled-jsx CSS in favour of Tailwind. The aim is to create a functional, good looking starter that will serve as an introduction to Tina for new users.
@spbyrne updated the screenshot in the README based on the one you shared.
What's supposed to happen when you try to access https://tina-cloud-starter-orcin.vercel.app/admin ?
| gharchive/pull-request | 2021-07-14T20:22:10 | 2025-04-01T06:46:01.053790 | {
"authors": [
"DirtyF",
"spbyrne"
],
"repo": "tinacms/tina-cloud-starter",
"url": "https://github.com/tinacms/tina-cloud-starter/pull/99",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
593417353 | panic: required envvar is unset (ROLLBAR_TOKEN)
Successful build on Ubuntu 18.04 for aarch64 under WSL, but when I follow the instructions in the README I get this:
ed@iyengar:~/src/github.com/tinkerbell/boots$ ./boots
{"level":"panic","ts":1585923538.9905274,"caller":"rollbar/rollbar.go:20","msg":"required envvar is unset","service":"github.com/tinkerbell/boots","pkg":"log","envvar":"ROLLBAR_TOKEN"}
panic: required envvar is unset
goroutine 1 [running]:
go.uber.org/zap/zapcore.(*CheckedEntry).Write(0x4000204000, 0x400016a300, 0x1, 0x2)
/home/ed/go/pkg/mod/go.uber.org/zap@v1.10.0/zapcore/entry.go:229 +0x40c
go.uber.org/zap.(*SugaredLogger).log(0x40001b0018, 0x4, 0x641d77, 0x18, 0x0, 0x0, 0x0, 0x40000e9a78, 0x2, 0x2)
/home/ed/go/pkg/mod/go.uber.org/zap@v1.10.0/sugar.go:234 +0xd4
go.uber.org/zap.(*SugaredLogger).Panicw(...)
/home/ed/go/pkg/mod/go.uber.org/zap@v1.10.0/sugar.go:204
github.com/packethost/pkg/log/internal/rollbar.Setup(0x40001b0018, 0x64397e, 0x1b, 0x2)
/home/ed/go/pkg/mod/github.com/packethost/pkg@v0.0.0-20190715213007-7c3a64b4b5e3/log/internal/rollbar/rollbar.go:20 +0x3a8
github.com/packethost/pkg/log.configureLogger(0x40001a0120, 0x64397e, 0x1b, 0x633420, 0x4, 0x632e61, 0x3, 0x633eef, 0x5)
/home/ed/go/pkg/mod/github.com/packethost/pkg@v0.0.0-20190715213007-7c3a64b4b5e3/log/log.go:69 +0x1c8
github.com/packethost/pkg/log.Init(0x64397e, 0x1b, 0x0, 0x0, 0x0, 0x0, 0x400008b838, 0x0)
/home/ed/go/pkg/mod/github.com/packethost/pkg@v0.0.0-20190715213007-7c3a64b4b5e3/log/log.go:87 +0xd4
main.main()
/home/ed/src/github.com/tinkerbell/boots/main.go:41 +0xb8
@vielmetti , boots runs in a docker container and while starting a container with boots image we take care of all the environment variables which are required to run the boots binary. Please refer docker-compose.yml file in tink repo to know all the envs which are required to run boots docker image.
@vielmetti @parauliya Considering the scenario where a user may not want to use Docker for the whole provisioned setup, I think it would be nice to have a list of all the environment variables.
We can have a list of environment variables required for each application to be ready, in the README. And, it would be nice if we can briefly mention what each is required for.
I ran into this too when testing this morning.
Maybe this requirement changed after the README was written? I'd suggest giving a sample that works in the README, even if the context for the env-var choices are written up somewhere else.
@parauliya the documentation doesn't mention a Docker container, so my understanding from the README is that it should run outside, otherwise the README should prefix the instructions with docker run golang:latest or similar before having us install git-lfs.
@nathangoulding do you see a path for usage inside and outside of the container, or should users only use compose?
I tried copying the instructions from https://github.com/tinkerbell/boots/blob/9d6cded511d1e7678ec601a29aac4137f25ecb5a/installers/rancher/main_test.go, but got as far as:
ROLLBAR_DISABLE=1 API_AUTH_TOKEN=token API_CONSUMER_TOKEN=test1234 PACKET_VERSION=0 PACKET_VERSION=0 PACKET_ENV=test ROLLBAR_TOKEN=1 ./boots
{"level":"info","ts":1586167193.6481194,"caller":"boots/main.go:54","msg":"starting","service":"github.com/tinkerbell/boots","pkg":"main","version":"unknown (use make)"}
{"level":"error","ts":1586167193.6482196,"caller":"boots/main.go:58","msg":"FACILITY_CODE env must be set","service":"github.com/tinkerbell/boots","pkg":"main","error":"FACILITY_CODE env must be set","errorVerbose":"FACILITY_CODE env must be set\ngithub.com/tinkerbell/boots/packet.NewClient\n\t/home/alex/go/src/github.com/tinkerbell/boots/packet/client.go:42\nmain.main\n\t/home/alex/go/src/github.com/tinkerbell/boots/main.go:56\nruntime.main\n\t/usr/lib/go/src/runtime/proc.go:203\nruntime.goexit\n\t/usr/lib/go/src/runtime/asm_amd64.s:1373"}
panic: FACILITY_CODE env must be set
goroutine 1 [running]:
github.com/packethost/pkg/log.Logger.Fatal(0xa96141, 0x1b, 0xc000010058, 0xc00011e9e0, 0xb62700, 0xc00000e820, 0x0, 0x0, 0x0)
/home/alex/go/pkg/mod/github.com/packethost/pkg@v0.0.0-20190715213007-7c3a64b4b5e3/log/log.go:119 +0xe4
main.main()
/home/alex/go/src/github.com/tinkerbell/boots/main.go:58 +0x954
[alex@nuc boots]$ FACILITY_CODE=test ROLLBAR_DISABLE=1 API_AUTH_TOKEN=token API_CONSUMER_TOKEN=test1234 PACKET_VERSION=0 PACKET_VERSION=0 PACKET_ENV=test ROLLBAR_TOKEN=1 ./boots
panic: looking up ip of mirror url: lookup install.test.packet.net on 8.8.8.8:53: no such host
goroutine 1 [running]:
github.com/tinkerbell/boots/env.mustFindMirrorIPBase(0xc000150880, 0xc000028380)
/home/alex/go/src/github.com/tinkerbell/boots/env/mirror.go:85 +0x20a
github.com/tinkerbell/boots/env.init()
/home/alex/go/src/github.com/tinkerbell/boots/env/mirror.go:23 +0x2b8
@alexellis The puzzling thing to me is the information about what should be used as the "facility code" - that appears in the current code to resolve to a hard-coded lookup to a install.$(FACILITY_CODE).packet.net address, but for on-prem or homelab install there would need to be a change to that mirror url.
Perhaps we are now going outside of the happy path?
We should remove this restriction.
Closing this as the original issue is no longer present (ROLLBAR_TOKEN is not required).
https://github.com/tinkerbell/boots/blob/65d8c38/cmd/boots/main.go#L67-L78
Please open additional images and refer back to this issue if other problems expressed in this thread need to be resolved.
| gharchive/issue | 2020-04-03T14:20:39 | 2025-04-01T06:46:01.066125 | {
"authors": [
"alexellis",
"displague",
"gauravgahlot",
"nathangoulding",
"parauliya",
"vielmetti"
],
"repo": "tinkerbell/boots",
"url": "https://github.com/tinkerbell/boots/issues/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
767801675 | tools/ci: add verify rule to check lint / boilerplate / go mod / check crds and go generate
Description
This PR adds makefile rule to verify the following
golangci-lint
boilerplate
go mod
crds and generated files
Also fixed some misalignments in go mod and some dirty data in gitignore
cc @invidian @detiber
Why is this needed
Fixes https://github.com/tinkerbell/cluster-api-provider-tink/issues/10
How Has This Been Tested?
n/a
How are existing users impacted? What migration steps/scripts do we need?
n/a
Checklist:
I have:
[ ] updated the documentation and/or roadmap (if required)
[ ] added unit or e2e tests
[ ] provided instructions on how to upgrade
ping @detiber
I have other PRs to push, but i would like to have this in place first :)
| gharchive/pull-request | 2020-12-15T16:45:32 | 2025-04-01T06:46:01.070641 | {
"authors": [
"cpanato"
],
"repo": "tinkerbell/cluster-api-provider-tink",
"url": "https://github.com/tinkerbell/cluster-api-provider-tink/pull/18",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
471264480 | Movie named "Climax" gets detected as "IMAX" Edition
Movie named "Climax" gets detected as "IMAX" Edition
reopen just as a reminder to myself to move this to gitlab (v3) later...
should be fixed in 3.0.4
| gharchive/issue | 2019-07-22T19:10:04 | 2025-04-01T06:46:01.075850 | {
"authors": [
"axbmcuser",
"mlaggner"
],
"repo": "tinyMediaManager/tinyMediaManager",
"url": "https://github.com/tinyMediaManager/tinyMediaManager/issues/494",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
898996165 | [meta] Upgrade setup steps in the GitHub actions file
Describe the feature
There are new versions available for action/checkout and action/setup-java. We should upgrade to benefit from bug fixes.
Additional context
As far as I can tell there are no changes necessary to upgrade the checkout step. For the setup-java step we have to choose a Java distribution. V1 used zulu there, so that might be a useful value there as well.
I plan to test the changes today in my Javalin-Rocker-Example repository.
The changes are really minimal. For my repository I also introduced caching for dependencies downloaded by Maven. That might also be a useful change for the Javalin repository.
@Playacem sounds like a good idea, please go ahead if you have time. If not I'll get to it eventually :)
| gharchive/issue | 2021-05-23T11:18:57 | 2025-04-01T06:46:01.126199 | {
"authors": [
"Playacem",
"tipsy"
],
"repo": "tipsy/javalin",
"url": "https://github.com/tipsy/javalin/issues/1254",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1079012070 | MINIO support by reacting to S3_HOST
The proper way to use MINIO or other S3-compatible object stores is by using the --endpoint-url parameter on the aws CLI.
All the configuration was well defined and prepared, but the S3_HOST environment variable was not used and thus MINIO users were not able to use this container. With this PR, that should be addressed.
I am myself a MINIO user and, after the proposed modifications, I was able to use this.
A comment that may be related to this PR: if I'm not mistaken S3_URI_STYLE is another variable that is not being used at the moment... but I have no idea on its meaning (I don't really know a whole lot of S3 and how it works and how compatible are "S3-compatible" object stores).
Great! I believe all of this went a bit sideways when some providers stopped accepting v2 of the API and only accepting v4. The S3_URI_STYLE if i recall was a way to access S3 buckets either as "bucketname.somehugecloudprovider.com" as opposed to "somehugecloudprovider.com/bucketname". I feel that Minio defaults always to Path. If its working for you, then great! I will merge and cleanup README.
Tagged as tiredofit/db-backup:2.9.5
| gharchive/pull-request | 2021-12-13T20:51:52 | 2025-04-01T06:46:01.129832 | {
"authors": [
"alexbarcelo",
"tiredofit"
],
"repo": "tiredofit/docker-db-backup",
"url": "https://github.com/tiredofit/docker-db-backup/pull/91",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
635646074 | h5py error: RuntimeError: Unable to create link (Name already exists)
This happens when I perform a model.save in h5 format through the default Model API. A possible source of error according to me seems to be https://github.com/tensorflow/hub/issues/287 but I am not able to wrap around on how to solve this issue for FastDeconv2D.
cc: @titu1994
@titu1994 you are right , but what I mean is all deconv layers are not being uniquely identified thus probably it error occurs when h5py comes across the second deconv layer in the network with the same signature as the first one.
You can explicitly pass a name to each layer to fix that. The default and preferred way to save models is with saved model or adsTF checkpoint.
| gharchive/issue | 2020-06-09T17:56:07 | 2025-04-01T06:46:01.144452 | {
"authors": [
"forkbabu",
"titu1994",
"ucalyptus"
],
"repo": "titu1994/tf_neural_deconvolution",
"url": "https://github.com/titu1994/tf_neural_deconvolution/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2381418052 | Problema Extra
DESCRIPTION
Nombre: Daniel Valdivia López
Email: dvaldivia98@hotmail.com
WhatsApp: 4494257983
Ciudad: Aguascalientes
Pais: México
¿Como explicarias el proposito de este proyecto? (opcional)
Básicamente trata de implementar una función que encuentre todas las combinaciones de números que sean únicas de un conjunto de candidatos. Estos números deben de sumar un objetivo especifico.
¿Cuál es tu stack tecnológico preferido? ¿Por qué? (opcional)
El stack que he manipulado más recientemente es Angular/Java(SpringBoot)/MySQL, ya que son tecnologías en las que
más he desarrollado y en las que he encontrado bastantes retos y soluciones a diversos problemas, Angular me ha encantado
por el desarrollo de su entorno Frontend y manejo de rutas, MySQL lo utilizo desde que tengo memoria (Aunque puede cambiar a MongoDB) y Java fue el primer lenguaje de programación que aprendí, entonces saber que lo siguen utilizando me agrada bastante.
Hola Daniel,
¡Gracias por tu participación! Aquí tienes el feedback sobre tu solución:
Funcionalidad: Tu implementación cumple correctamente con los requisitos del problema.
Algoritmo: Tu enfoque recursivo es adecuado y eficiente para este tipo de problema.
Pruebas: Asegúrate de incluir casos de prueba en el repositorio para validar tu solución.
Calidad del Código: El código es claro y bien estructurado, pero faltan algunos comentarios y explicaciones.
Documentación y Estructura: Sería útil agregar un .gitignore y una mejor estructura de carpetas.
Sugerencia de Optimización: Considera agregar validaciones adicionales para mejorar la robustez.
¡Buen trabajo!
Saludos,
Alejandro
| gharchive/pull-request | 2024-06-29T02:38:46 | 2025-04-01T06:46:01.151574 | {
"authors": [
"Defrahnz",
"alejozepol"
],
"repo": "tiui-fulfillment/tiui-combination-sum-",
"url": "https://github.com/tiui-fulfillment/tiui-combination-sum-/pull/58",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
86255904 | Updates to the marker to allow custom images with different anchors
Updates include allowing the setting of a google map icon which will allow more custom settings of how the marker is handled. As well, allowing for the zIndex of the marker to be set.
Sorry for the two commits. I submitted a prior change, but back tracked on some of that implementation to make the implantation for fluent with Google's API.
| gharchive/pull-request | 2015-06-08T17:02:23 | 2025-04-01T06:46:01.175602 | {
"authors": [
"andyphillips404"
],
"repo": "tjkaal/GoogleMapsVaadin7",
"url": "https://github.com/tjkaal/GoogleMapsVaadin7/pull/26",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2214666894 | 🛑 yt.tjo.space is down
In 7e6b502, yt.tjo.space (https://yt.tjo.space/feed/trending) was down:
HTTP code: 404
Response time: 547 ms
Resolved: yt.tjo.space is back up in 9172fc3 after 12 minutes.
| gharchive/issue | 2024-03-29T05:13:33 | 2025-04-01T06:46:01.178085 | {
"authors": [
"mentos1386"
],
"repo": "tjo-space/status",
"url": "https://github.com/tjo-space/status/issues/1082",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1213506444 | filter dropdown fix
Description
How Has This Been Tested?
Types of changes
[x] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
Checklist:
[x] The PR is submitted to the correct branch (dev).
[x] My code follows the code style of this project.
[ ] I've added relevant tests.
latest grid filter dropdown broken, it repeats the current column filter name over and over instead of displaying the other column names. easy fix. please accept
| gharchive/pull-request | 2022-04-24T02:29:59 | 2025-04-01T06:46:01.186225 | {
"authors": [
"segfault-"
],
"repo": "tjscience/MudBlazor",
"url": "https://github.com/tjscience/MudBlazor/pull/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
223561435 | Keyboard does not appear with the Search bar
Hi the keyboard only seem to work with textview and not with the search bar. so currently dfferent keyboards appear depending on where the user types. It would be good if the plugin worked with the search bar too.
@anuragd7 Sorry for the late response here.
Not sure what we can do about this as this is behavior of the CocoaPod itself. We do now expose the raw plugin typings, so you have a little more power to configure things (see https://github.com/tjvantoll/nativescript-IQKeyboardManager#advanced-usage).
I’m going to close this as I’m not sure what we can really do in this plugin as its just a simple wrapper of the CocoaPod, but if anyone stumbles upon this and has any ideas please comment.
| gharchive/issue | 2017-04-22T12:29:51 | 2025-04-01T06:46:01.192712 | {
"authors": [
"anuragd7",
"tjvantoll"
],
"repo": "tjvantoll/nativescript-IQKeyboardManager",
"url": "https://github.com/tjvantoll/nativescript-IQKeyboardManager/issues/17",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
104066856 | local strategy, POST /auth/local return 403
Hello,
I tried sails-auth (v1.3.1), and followed the README.
I created a new user thanks to POST /user, with {"email":"admin@example.com","password":"admin123"}, and then tried to login thanks to POST /auth/local with the same body, and sails-auth returns a 403 error. What am I doing wrong ?
Thanks,
Sylvain
https://github.com/tjwebb/sails-auth/wiki/Logging-In
To login, you should send identity and password.
Yes, it worked. Thanks !
And is that possible to set the identifier as another attribute, for example the username instead of the email ?
I know this is old but it's still an open discussion. Can we change that request to a 400 error instead of a 403? 403 is "you do not have permission" and 400 is "bad request" (which missing a required field is)
how to redirect to an error page if the login fails?
| gharchive/issue | 2015-08-31T14:39:40 | 2025-04-01T06:46:01.196488 | {
"authors": [
"Simone-cogno",
"ryanwilliamquinn",
"sylvainlap",
"westlakem"
],
"repo": "tjwebb/sails-auth",
"url": "https://github.com/tjwebb/sails-auth/issues/87",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
87107805 | .v8flags.json on Linux
Could the .v8flags.(version).(username).json file be written to the standard directory ~/.cache on Linux?
Sure could. Feel free to PR for that.
#21
| gharchive/issue | 2015-06-10T21:07:42 | 2025-04-01T06:46:01.201065 | {
"authors": [
"thesbros",
"tkellen"
],
"repo": "tkellen/js-v8flags",
"url": "https://github.com/tkellen/js-v8flags/issues/20",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1458125249 | 肉鸽bug 自动结束并退出
多次出现这两种情况:
选完人一进去就退了
打完第一关/第二关, 后面全是不期而遇+商店的情况, 直接退出+结束肉鸽
不是重启程序, 是直接结束肉鸽+重进
录屏反馈
| gharchive/issue | 2022-11-21T15:23:06 | 2025-04-01T06:46:01.205412 | {
"authors": [
"DrKenther",
"tkkcc"
],
"repo": "tkkcc/ArkLights",
"url": "https://github.com/tkkcc/ArkLights/issues/92",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
308060769 | Spring Stroke Offset
Hi @tkoziara,
Would it be possible to include some additional functionality in the spring curve command to modify spring curves with time?
For example, we would like to be able to adjust the red dashed spring curve in the image below to become the blue solid spring curve over time.
The change to the API could be the following:
sprnum = SPRING (part1, point1, part2, geom2, spring | dashpot, direction, planar, unload, ylim, inactive, strokeoffset)
Where "strokeoffset" is an adjustment applied to all stroke values of the spring curve (dashpot unchanged), defined as a list or TSERIES number: [time1,strokeoffset1,time2,strokeoffset2,...,timen,strokeoffsetn]. The time values must be monotonically increasing. Ideally no limitation is imposed on strokeoffset values (however we currently intend on using either monotonically increasing or monotonically decreasing values).
Or perhaps you could suggest a better alternative?
The reason for this request is that we are intending on running a model of an array where components positioned "on-pitch" at t=0s would experience an initial overclosure with adjacent components, i.e. spring forces would be non-zero at t=0s. To avoid having initial strain energy in the model, we would like to position components on pitch with spring curves defined such that spring forces are 0 at t=0s, then gradually adjust these spring curves over time until the target initial configuration is achieved.
Best regards,
Ben
(mention @stevebrasier)
Hi @bencannell , @stevebrasier ; Yes, this should be OK to implement; I will look into this this and in the coming week:)
Have a look: http://parmes.org/parmec/input_commands.html#spring
This is on GitHub now, but has not been tested yet:) T.
FYI, initially made work; Try on this example:) closing the issue; T.
Hi @tkoziara, I've tested this implementation and it produces expected results.
Thank you!
Good :)
| gharchive/issue | 2018-03-23T14:53:50 | 2025-04-01T06:46:01.212423 | {
"authors": [
"bencannell",
"tkoziara"
],
"repo": "tkoziara/parmec",
"url": "https://github.com/tkoziara/parmec/issues/8",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
327220168 | Migration fails, every remote revision setting lost!
I updated from commit 478d418
# git fetch
remote: Counting objects: 45, done.
remote: Total 45 (delta 17), reused 17 (delta 17), pack-reused 28
Unpacking objects: 100% (45/45), done.
From https://github.com/tleish/redmine_remote_revision_url
478d418..f0b6c33 master -> origin/master
Had a look at the commit history and decided it's worth a try to update.
l# git pull
Updating 478d418..f0b6c33
Fast-forward
README.md | 5 +++-
.../redmine_remote_revision_url/_settings.html.erb | 9 ++++----
db/migrate/001_update_saved_settings.rb | 18 +++++++++++++++
init.rb | 9 ++++++--
.../application_helper_patch.rb | 8 ++-----
.../repositories_helper_patch.rb | 2 +-
.../repository_patch.rb | 24 +++++++++++++++-----
7 files changed, 54 insertions(+), 21 deletions(-)
create mode 100644 db/migrate/001_update_saved_settings.rb
Then I ran the migrations:
# bundle exec rake redmine:plugins:migrate RAILS_ENV=production
(in /var/www/projekte/redmine-3.4)
...
Migrating redmine_remote_revision_url (Remote Revision URL)...
== 1 UpdateSavedSettings: migrating ===========================================
== 1 UpdateSavedSettings: migrated (0.0162s) ==================================
...
Looks good. Then I restarted Redmine.
And all plugin fields in all repositories are empty. Every Git remote is lost!
And I thought there had been proper migrations with #10 ?!
When looking at the database tables I can still see all my settings in the field "extra_info". So it's obviously not completely lost...
How can I restore it?
So I updated the 20 repositories in my DB manually and added a colon : before the extra_info field names. Works again.
However, I still think that this is a regression from previous bugfixes. None of the other extra_info fields are starting with a colon, it's just a plain string token.
I no longer use Redmine and cannot effectively update and maintain this repo. Let me know if interested in taking it over?
| gharchive/issue | 2018-05-29T08:35:32 | 2025-04-01T06:46:01.246946 | {
"authors": [
"tleish",
"tofi86"
],
"repo": "tleish/redmine_remote_revision_url",
"url": "https://github.com/tleish/redmine_remote_revision_url/issues/14",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1655837393 | About padding_side='left'
Hi,
I noticed that you set padding_side='left' in finetune.py. However, in llama the default padding side seems to be 'right'. Would this inconsistency causes certain problems such as performance drop?
I think default llama doesn't even use padding tokens.
how can it train without padding? I thought padding is necessary to collate sentences of different lengths into the same batch?
@stellaludai correct but during pre-training you usually don't use different lenghts. See HF tutorial for more details: https://huggingface.co/course/chapter7/6?fw=pt#preparing-the-dataset
yes, exactly
@chrisociepa @ElleLeonne Thank you very much for sharing!
I think @ElleLeonne just refers to the fact that there is no pad_token in the LlamaTokenizer config, the default value is None.
@stellaludai correct but during pre-training you usually don't use different lenghts. See HF tutorial for more details: https://huggingface.co/course/chapter7/6?fw=pt#preparing-the-dataset
@chrisociepa Thank you for sharing the link. But I don't think they are about the same thing as the question itself:
The example in the link manually throws away chunks where length != context_length but the alpaca-Lora did not do this when creating dataset. Rather, it does the following padding in datacollator:
data_collator=transformers.DataCollatorForSeq2Seq(
tokenizer, pad_to_multiple_of=8, return_tensors="pt", padding=True
)
I think padding_side = left/right should not matter but attention_mask should be implemented which is not the case in alpaca-lora IMHO.
@Nsigma-Bill Thanks for sharing your opinion, but actually the above answers solved my question. In the link, the last remaining tail are thrown away if it is shorter than the previous chunks, which aims to avoid padding issues. As for attention mask, the transformers package should already have done for this purpose.
| gharchive/issue | 2023-04-05T15:24:39 | 2025-04-01T06:46:01.259101 | {
"authors": [
"ElleLeonne",
"Nsigma-Bill",
"chrisociepa",
"stellaludai"
],
"repo": "tloen/alpaca-lora",
"url": "https://github.com/tloen/alpaca-lora/issues/277",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1603893068 | chat: return of the double render
Now seeing double renders in some group chat channels:
This seems to only happen in Groups, not Talk.
This also only seems to happen after a reconnect, but I can't repro it locally on master. I'm attempting to repro by making sure the disconnect button is always visible and clicking the Reconnect button to call bootstrap(true). Everything seems fine?
Weird thing is that when an actual disconnect happens on my ship (which is running on wanec, therefore master), I get the double message renders.
| gharchive/issue | 2023-02-28T22:12:28 | 2025-04-01T06:46:01.261098 | {
"authors": [
"patosullivan"
],
"repo": "tloncorp/landscape-apps",
"url": "https://github.com/tloncorp/landscape-apps/issues/2026",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2423064769 | activity: web match mobile bell tab
Fixes TLON-2419 by mimicking what mobile does, basically clearing the bell any time you visit and bringing it back upon new relevant activity. This also fixes optimistic reads which were not happening because we defaulted variables.action in the mutation function and didn't account for that later on.
Button moved to settings but not sure what language we want here:
PR Checklist
[ ] Includes changes to desk files
[ ] Describes how you tested the PR locally (test ship vs livenet)
[ ] If a new feature, includes automated tests
[ ] Comments added anywhere logic may be confusing without context
diff is coming up weird because it's including the base branch changes for some reason
| gharchive/pull-request | 2024-07-22T14:44:36 | 2025-04-01T06:46:01.268982 | {
"authors": [
"arthyn"
],
"repo": "tloncorp/tlon-apps",
"url": "https://github.com/tloncorp/tlon-apps/pull/3780",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1064078134 | Help:How to replay a message?
How can i send the same message packet twice in a row?
I want to simulate a replay attack.
Use Java programming or XML
`Configurator.setAllLevels("", Level.INFO);
Security.addProvider(new BouncyCastleProvider());
Config config = Config.createConfig();
config.setDefaultRunningMode(RunningModeType.CLIENT);
OutboundConnection outboundConnection = config.getDefaultClientConnection();
outboundConnection.setPort(Integer.parseInt(args[0]));
config.setDefaultClientConnection(outboundConnection);
config.setDefaultSelectedCipherSuite(CipherSuite.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256);
config.setAddSessionTicketTLSExtension(true);
WorkflowTrace trace = new WorkflowTrace();
ClientHelloMessage clientHello = new ClientHelloMessage(config);
MessageAction messageAction = MessageActionFactory.createAction(
config, outboundConnection, ConnectionEndType.CLIENT, clientHello);
trace.addTlsAction(messageAction);
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ServerHelloMessage(config),
new CertificateMessage(config),
new ECDHEServerKeyExchangeMessage(config),
new ServerHelloDoneMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
new ECDHClientKeyExchangeMessage(config),
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
State state = new State(config, trace);
WorkflowExecutor workflowExecutor =
WorkflowExecutorFactory.createWorkflowExecutor(config.getWorkflowExecutorType(), state);
workflowExecutor.executeWorkflow();
System.out.println(clientHello.toString());
System.out.println(ArrayConverter.bytesToHexString(state.getTlsContext().getClientRandom()));
WorkflowTrace trace2 = new WorkflowTrace();
SendAction sendaction = new SendAction(clientHello);
sendaction.setRecords(messageAction.getRecords());
trace2.addTlsAction(sendaction);
trace2.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ServerHelloMessage(config),
new CertificateMessage(config),
new ECDHEServerKeyExchangeMessage(config),
new ServerHelloDoneMessage(config)));
trace2.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
new ECDHClientKeyExchangeMessage(config),
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
trace2.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
State state2 = new State(config, trace2);
// state2.replaceTlsContext(state.getTlsContext());
WorkflowExecutor workflowExecutor2 =
WorkflowExecutorFactory.createWorkflowExecutor(config.getWorkflowExecutorType(), state2);
workflowExecutor2.executeWorkflow();`
this is my code.
i want to handeshake first, and resend the client hello which had send at the first handshake in second handshake.
Do you want to replay the client hello encrypted or do you want to replay it plain?
Also - do you want to replay it as separate session (this is what you are doing now) or do you want to replay it in the same session?
Ok I think I get what you are trying to do now. You simply have to set: setUseFreshRandom(false) in the Config. That should do the trick
thank you very much
another question is: how can i resend a same app message packet in one session
Config config = Config.createConfig();
config.setDefaultRunningMode(RunningModeType.CLIENT);
OutboundConnection outboundConnection = config.getDefaultClientConnection();
outboundConnection.setPort(Integer.parseInt(args[0]));
config.setDefaultClientConnection(outboundConnection);
config.setDefaultSelectedCipherSuite(CipherSuite.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256);
config.setAddSessionTicketTLSExtension(true);
WorkflowTrace trace = new WorkflowTrace();
trace.addTlsAction(MessageActionFactory.createAction(
config, outboundConnection, ConnectionEndType.CLIENT, new ClientHelloMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ServerHelloMessage(config),
new CertificateMessage(config),
new ECDHEServerKeyExchangeMessage(config),
new ServerHelloDoneMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
new ECDHClientKeyExchangeMessage(config),
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
config.setUseFreshRandom(false);
config.setUseAllProvidedRecords(true);
ApplicationMessage app1 = new ApplicationMessage(config);
MessageAction messageAction = MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
app1);
trace.addTlsAction(messageAction);
MessageAction messageAction2 = MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
app1);
messageAction2.setMessages(messageAction.getMessages());
messageAction2.setRecords(messageAction.getRecords());
System.out.println(messageAction.getReadableString());
trace.addTlsAction(messageAction2);
State state = new State(config, trace);
WorkflowExecutor workflowExecutor =
WorkflowExecutorFactory.createWorkflowExecutor(config.getWorkflowExecutorType(), state);
workflowExecutor.executeWorkflow();
I refer to this issue: https://github.com/tls-attacker/TLS-Attacker/issues/98, and Modify the code as follows.
But the following error is reported:
Exception in thread "main" de.rub.nds.tlsattacker.core.exceptions.ContextHandlingException: Cannot replace TlsContext because the new TlsContext defines another connection.
at de.rub.nds.tlsattacker.core.state.ContextContainer.replaceTlsContext(ContextContainer.java:176)
at de.rub.nds.tlsattacker.core.state.State.replaceTlsContext(State.java:166)
at TLSS_TestSuite1_TestCase4_3a.main(TLSS_TestSuite1_TestCase4_3a.java:109)
Config config = Config.createConfig();
config.setDefaultRunningMode(RunningModeType.CLIENT);
OutboundConnection outboundConnection = config.getDefaultClientConnection();
outboundConnection.setPort(Integer.parseInt(args[0]));
config.setDefaultClientConnection(outboundConnection);
config.setDefaultSelectedCipherSuite(CipherSuite.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256);
config.setAddSessionTicketTLSExtension(true);
de.rub.nds.tlsattacker.core.state.TlsContext ctx = new de.rub.nds.tlsattacker.core.state.TlsContext(config, outboundConnection);
WorkflowTrace trace = new WorkflowTrace();
trace.addTlsAction(MessageActionFactory.createAction(
config, outboundConnection, ConnectionEndType.CLIENT, new ClientHelloMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ServerHelloMessage(config),
new CertificateMessage(config),
new ECDHEServerKeyExchangeMessage(config),
new ServerHelloDoneMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
new ECDHClientKeyExchangeMessage(config),
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
trace.addTlsAction(
MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.SERVER,
new ChangeCipherSpecMessage(config),
new FinishedMessage(config)));
config.setUseFreshRandom(false);
config.setUseAllProvidedRecords(true);
ApplicationMessage app1 = new ApplicationMessage(config);
Chooser chooser = ChooserFactory.getChooser(ChooserType.DEFAULT, ctx, config);
ApplicationMessagePreparator appPreparator = new ApplicationMessagePreparator(chooser, app1);
appPreparator.prepare();
MessageAction messageAction = MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
app1);
trace.addTlsAction(messageAction);
MessageAction messageAction2 = MessageActionFactory.createAction(
config,
outboundConnection,
ConnectionEndType.CLIENT,
app1);
messageAction2.setMessages(messageAction.getMessages());
messageAction2.setRecords(messageAction.getRecords());
System.out.println(messageAction.getReadableString());
trace.addTlsAction(messageAction2);
State state = new State(config, trace);
state.replaceTlsContext(ctx);
WorkflowExecutor workflowExecutor =
WorkflowExecutorFactory.createWorkflowExecutor(config.getWorkflowExecutorType(), state);
workflowExecutor.executeWorkflow();
| gharchive/issue | 2021-11-26T03:47:21 | 2025-04-01T06:46:01.279549 | {
"authors": [
"dennisokko",
"ic0ns"
],
"repo": "tls-attacker/TLS-Attacker",
"url": "https://github.com/tls-attacker/TLS-Attacker/issues/118",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.