id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
155847262 | Occasional build failure in wseb acceptor spec test ControlIT.shouldCloseConnectionOnReceivingInvalidPingFromClient
This failure occurred for example in build #1540 on Travis. Here's an extract from the build output showing the failure:
shouldCloseConnectionOnReceivingInvalidPingFromClient(org.kaazing.gateway.transport.wseb.specification.wse.acceptor.ControlIT) Time elapsed: 0.147 sec <<< FAILURE!
org.junit.ComparisonFailure: Specified behavior did not match expected:<...close
read status "[200" /.+/
read header "Content-Type" "application/octet-stream"
read header "Connection" "close"
read [0x01 0x30 0x32 0xFF]
read [0x01 0x30 0x31 0xFF]
read closed]
# Upstream
connect...> but was:<...close
read status "[404"]
Looking at the specification script: client.send.invalid.ping/request.rpt, it is clear this is an issue with the script. There is a race between the establishment of the upstream and the sending of the invalid ping frame and the establishment of the downstream. If the downstream connects after the upstream then the observed behavior is correct (status 404 response, not found, on the downstream, because the logical wseb connection was already closed). This needs to be fixed in the spec scripts: use barriers to force downstream to be established before the invalid ping frame is sent on the upstream. If have filed k3po issue https://github.com/k3po/k3po/issues/339 for this.
k3po PR https://github.com/k3po/k3po/pull/369
| gharchive/issue | 2016-05-19T22:12:41 | 2025-04-01T06:44:39.731770 | {
"authors": [
"cmebarrow",
"jitsni"
],
"repo": "kaazing/gateway",
"url": "https://github.com/kaazing/gateway/issues/580",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
352062302 | Reminder slack notifications are sent twice
This might be a configuration issue. Find the root cause and fix this.
Probably related to how kabisa team's team memberships and duplicate accounts are stored (e.g. the recent change from google to move from alias to account name (ariejan vs. ariejan.de.vroom).
Ignore the previous comment. It has nothing to do with accounts.
We use Rufus to schedule cronjobs for us, straight from the Ruby code. On boot-up, Rufus will update cron to schedule the appropriate tasks.
Since we run two containers, one for web, one for worker, we schedule these jobs on both containers, hence we get two notifications.
Rufus is not a recommended way of scheduling tasks. I've updated this issue's original message with TODO's on how to resolve this.
/cc @pascalw
| gharchive/issue | 2018-08-20T09:52:16 | 2025-04-01T06:44:39.736816 | {
"authors": [
"ariejan"
],
"repo": "kabisa/kudo-o-matic",
"url": "https://github.com/kabisa/kudo-o-matic/issues/372",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1293139633 | calcpriceとは?
calcpriceとは?
【質問】計算用現値CalcPriceについて · Issue #35 · kabucom/kabusapi
@hori-ryota さんの方でご回答いただいておりますが、こちら、計算用現値のこととなります
解決済みのようですので、クローズさせていただきます。
| gharchive/issue | 2022-07-04T12:58:33 | 2025-04-01T06:44:39.738606 | {
"authors": [
"Uchiii7",
"cyber-minion",
"hori-ryota",
"yasuyuki-nakazawa"
],
"repo": "kabucom/kabusapi",
"url": "https://github.com/kabucom/kabusapi/issues/562",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1812075742 | 225MINIでの返済について(/sendorder/future)
225MINIでの返済について(/sendorder/future)
売玉有の時 返済買いの注文を出しましたがエラーになってしまいます。
新規注文(TradeType: 1 , Side: '1')は問題なくできました。
以下のようなデータを送信していますがなにか間違いがありますでしょうか?
payload = {
Password: 'XXXXXXX',
Symbol: '168090019',
Exchange: 23,
TradeType: 2,
Side: '2',
Qty: 1,
ClosePositions: [{ HoldID: '20230719XXXXXXX', Qty: 1 }],
FrontOrderType: 120,
Price: 0,
ExpireDay: 0,
TimeInForce: 2
}
エラーコードは8で返済建玉情報不正エラー のようです。
アドバイスいただければと思います。
以下のパラメータの設定について、ご確認をお願いいたします。
ClosePositions
HoldIDは、残高照会APIで取得したExecutionIDを指定していただく必要がございます。
下記の例のように、先頭に「E」が付いているかと思いますので、ご確認をお願いします。
例:E20230719xxxxx
ExecutionIDの件、ありがとうございます。確認してみます
保有数が3枚ありどれかを返済したいときにOrderIdに紐づけれないとなれば
損益が真ん中のものを決済を最初にはできないとのことで良いでしょうか?
ExcutionIDは約定ごとに採番されています。
お問い合わせのように保有数が3枚あり、それぞれが別々に約定したものであれば、
それぞれに異なるExcutionIDがありますので、損益が真ん中のものを選択することも可能です。
確認してみます。ありがとうございました
| gharchive/issue | 2023-07-19T14:25:11 | 2025-04-01T06:44:39.741916 | {
"authors": [
"app-yama",
"satosato125"
],
"repo": "kabucom/kabusapi",
"url": "https://github.com/kabucom/kabusapi/issues/717",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
337286891 | Commands
어떤 명령어들을 만들까요?
[ ] 계정 이름으로 전적 검색
[ ] 계정의 최근 경기 검색?
뭔차이가 있는 건가용?
@kagong
계정 이름으로 전적 검색은 kda, rating이라던가 승률 등등 계정의 정보고요.
최근 경기 검색은 해당 계정의 최근 경기의 결과라던가 여러가지 정보가 들어있을거 같네요.
API Document를 좀 뒤적여봐야겠어요.
@JimJeon
경기 검색에서는
"attributes": {
"createdAt": "string",
"duration": 0,
"gameMode": "duo",
"mapName": "Desert_Main",
"isCustomMatch": true,
"patchVersion": "string",
"shardId": "string",
"stats": {},
"tags": {},
"titleId": "string"
},
이런 속성들이 있는 것 같은데
"relationships": {
"assets": {
"data": [
{
"type": "string",
"id": "string"
}
]
},
"rosters": {
"data": [
{
"type": "string",
"id": "string"
}
]
},
"rounds": {},
"spectators": {}
},
여기서 rosters 와 asset object들이 뭘 위한건지 이해가 안가네요 한번 봐주세용
https://documentation.playbattlegrounds.com/en/matches-endpoint.html#/Matches/get_matches__id_
{
"data": {
"type": "match",
"id": "e8af5c90-c784-400e-952e-f2949833dfc9",
"attributes": {
"shardId": "pc-kakao",
"tags": null,
"mapName": "Savage_Main",
"isCustomMatch": false,
"stats": null,
"duration": 1432,
"gameMode": "squad",
"titleId": "bluehole-pubg",
"createdAt": "2018-07-16T23:32:52Z"
},
"relationships": {
"assets": {
"data": [
{
"type": "asset",
"id": "1f46ec59-8954-11e8-b3ec-0a586461c617"
}
]
},
"rosters": {
"data": [
{
"type": "roster",
"id": "0d55d986-ba6f-49fe-9e32-00e9c933d384"
},
{
"type": "roster",
"id": "9035ab94-d2de-4e29-bc71-8aa6919dc2a1"
},
{
"type": "roster",
"id": "f3eb1099-7f02-4793-9f84-b2315984b26a"
}
]
}
},
"links": {
"self": "https://api.playbattlegrounds.com/shards/pc-kakao/matches/e8af5c90-c784-400e-952e-f2949833dfc9",
"schema": ""
}
}
}
이런 느낌으로 값을 받을 수 있네요. 응답이 너무 길어서 대부분 잘라내고 말씀해주신 부분만 남겨보았습니다.
"relationships": {
...
"assets": {
"data": [
{
"type": "asset",
"id": "1ad97f85-cf9b-11e7-b84e-0a586460f004"
}
]
}
...
}
위와 같은 Assets Object에 대해서 아래와 같이 telemetry.json을 접근 할 수 있다고 합니다.
curl "https://telemetry-cdn.playbattlegrounds.com/pc-krjp/2018/01/01/0/0/1ad97f85-cf9b-11e7-b84e-0a586460f004-telemetry.json" -H "Accept: application/vnd.api+json"
| gharchive/issue | 2018-07-01T12:21:39 | 2025-04-01T06:44:39.832174 | {
"authors": [
"JimJeon",
"kagong"
],
"repo": "kagong/PyGS",
"url": "https://github.com/kagong/PyGS/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
210353384 | Mount google cloud storage?
Any idea on how to mount google cloud storage using goofys?
it should just work if you set --endpoint to the S3 endpoint after you enable it in the google dashboard.
I see, thanks.
There's some info here for future reference:
https://cloud.google.com/storage/docs/migrating
https://cloud.google.com/storage/docs/interoperability
Thanks for the links!
Could someone post a CLI example of mounting Google Drive using goofys? Where do we need to adjust or set something on our Google Dashboard, for example (not everyone will know this).
Could someone post a CLI example of mounting Google Drive using goofys? Where do we need to adjust or set something on our Google Dashboard, for example (not everyone will know this).
I dont think goofys works with google drive. This issue is about https://cloud.google.com/storage.
I dont think goofys works with google drive. This issue is about https://cloud.google.com/storage.
| gharchive/issue | 2017-02-26T22:14:41 | 2025-04-01T06:44:39.836461 | {
"authors": [
"dotslash",
"forrie",
"kahing",
"peixotorms"
],
"repo": "kahing/goofys",
"url": "https://github.com/kahing/goofys/issues/158",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
334641832 | What is the path to the syslog file
After I ran $GOBIN/goofys --debug_s3 bucket_name dir, it said:
2018/06/21 15:13:47.098950 main.FATAL Unable to mount file system, see syslog for details
Which file should I look?
hi @taozhaolib , on Ubuntu this should be /var/log/syslog
| gharchive/issue | 2018-06-21T20:17:59 | 2025-04-01T06:44:39.838238 | {
"authors": [
"taozhaolib",
"weirdbricks"
],
"repo": "kahing/goofys",
"url": "https://github.com/kahing/goofys/issues/332",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
479326030 | goofys not work and not print log in wsl(Windows Subsystem Linux) Ubuntu 18.04
I downloaded goofys binary and run in Windows 2019 wsl(Windows Subsystem Linux) Ubuntu 18.04. It's not worked and print error message:
2019/08/11 02:43:20.681328 main.FATAL Unable to mount file system, see syslog for details
But unlucky, there's also no output in /var/log/syslog .
My questions are:
How can Goofys support in wsl ubuntu 18.04 in Windows Server 2019?
Why there's not error log in /var/log/syslog?
Thanks a lot.
Try -f to run the process in the foreground which will print the logs.
Thanks yuvisara.
I tried with -f parameters like below:
wsl-user@windows-wsl:~$ sudo ./goofys -f s3-temp-driver /mnt/goofys/
shown below:
2019/08/12 15:29:20.708697 main.FATAL Mounting file system: Mount: mount: running fusermount: exit status 1
stderr:
fusermount: fuse device not found, try 'modprobe fuse' first
Then I tried sudo modprobe fuse, got:
ec2-user@EC2AMAZ-TAE8VA7:~$ sudo modprobe fuse
modprobe: ERROR: ../libkmod/libkmod.c:586 kmod_search_moddep() could not open moddep file '/lib/modules/4.4.0-17763-Microsoft/modules.dep.bin'
modprobe: FATAL: Module fuse not found in directory /lib/modules/4.4.0-17763-Microsoft
From https://github.com/microsoft/WSL/issues/2869 I can see, no fuse support in wsl now.
Thanks.
I don't use windows personally so someone else will have to do the investigations. It's perhaps possible to port the underlining fuse library to use windows specific mechanism instead but again someone else will have to look into it
Got it and thanks for you guys replies.
| gharchive/issue | 2019-08-11T02:45:43 | 2025-04-01T06:44:39.844061 | {
"authors": [
"Allen2Git",
"kahing",
"yuvisara"
],
"repo": "kahing/goofys",
"url": "https://github.com/kahing/goofys/issues/443",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
284112933 | 清除全选方法
提供清除全选的方法。
https://github.com/kairi1227/react-drag-picker/commit/44d47efa1415a2f283d0f3577e19221ef888a19d
| gharchive/issue | 2017-12-22T09:20:34 | 2025-04-01T06:44:39.849337 | {
"authors": [
"kairi1227"
],
"repo": "kairi1227/react-drag-picker",
"url": "https://github.com/kairi1227/react-drag-picker/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
989750231 | Question about datasets
Why did you use Republic of Korea datasets?
We first tried to use other datasets from other countires, but it was impossible because we could not access to the actual driving record to classify dangerous driving behavior of each trajectory. However, fortunately, the KoROAD was able to provide such dataset to us, so we only use Korea datasets.
| gharchive/issue | 2021-09-07T08:53:08 | 2025-04-01T06:44:39.853644 | {
"authors": [
"itouchz",
"young-eun-nam"
],
"repo": "kaist-dmlab/DF-TAR",
"url": "https://github.com/kaist-dmlab/DF-TAR/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
614207956 | Fix typo in day3's and day4's documents
day3
どれだけ効率が良くなかったを示す指標 to どれだけ効率が良くなったかを示す指標
day4
全プロセス勝手に吐くく to 全プロセス勝手に吐く
Thanks @qb0C80aE !
| gharchive/pull-request | 2020-05-07T17:00:57 | 2025-04-01T06:44:39.864902 | {
"authors": [
"kaityo256",
"qb0C80aE"
],
"repo": "kaityo256/sevendayshpc",
"url": "https://github.com/kaityo256/sevendayshpc/pull/13",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1557895577 | Add Unit tests to book and label
In this PR I have made the following changes:
Added unit tests for Book class in book_spec.rb
Added unit tests for Label class in label_spec.rb
Removed read_booklabel method from app.rb
Hi @kaizipaul
Your changes are approved! There is nothing else to say other than... it's time to merge the branch :shipit:
Congratulations! 🎉
Status: Approved ✔️✔️✔️
Cheers and Happy coding!👏👏👏
| gharchive/pull-request | 2023-01-26T10:01:05 | 2025-04-01T06:44:39.883687 | {
"authors": [
"DJ-MrJay",
"kaizipaul"
],
"repo": "kaizipaul/ruby-capstone-project",
"url": "https://github.com/kaizipaul/ruby-capstone-project/pull/34",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1906168329 | fix bug during w2v training with utf8 characters
bug
when training w2v with Korean words(utf-8 characters), idmap['cols'] couldn't get utf-8
UnicodeEncodeError: 'ascii' codec can't encode characters in position 0-1: ordinal not in range(128)
as-is
dtype=Sx
S: data type should be ascii characters
to be
dtype=h5py.string_dtype('utf-8')
docs
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
| gharchive/pull-request | 2023-09-21T05:55:03 | 2025-04-01T06:44:39.888002 | {
"authors": [
"CLAassistant",
"hugh-ga"
],
"repo": "kakao/buffalo",
"url": "https://github.com/kakao/buffalo/pull/75",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
216133341 | Разработка полной схемы аналоговой части
Исчерпывающая схема с полной распиновкой.
Включая:
Входные каскады
Выходные каскады
Усилители из корпусе
Интерфейсы
Обвеска всех чипов (микроконтроллер, драйвер RS232, корпус с усилителями)
Подведение итогов по аналоговой части.
Реализует коммит https://github.com/kalaider/act-photo/commit/494b4d950dabe0811470aebe8d9b43f5b2261073.
Описание
На рисунке представлена полная на данный момент схема аналоговой части.
Отображено:
Первичный входной каскад (диод в генераторном включении)
Фильтр нижних частот
Конечный усилитель
Каскад установки опорного уровня освещенности
Выходной каскад
Интерфейсные схемы (интерфейсы I2C, SPI, RS232 через драйвер ST232 с обвеской)
Обвеска ATmega8A
Напряжения питания (VCC = +5V, VC = +12V, VE = -12V)
Не отображено:
Ввод питания и земли
Отдельные транзисторы заменены на микросхему TL084 из четырех транзисторов. Схемы фильтров упрощены руками.
Провод земли для наглядности помечен пурпурным цветом.
Модель
Свет, попадая на фотодиод, вызывает в нем фототок, который, проходя через резистор обратной связи первичного усилителя (а точнее преобразователя "ток-напряжение") преобразуется в падение напряжения, которое затем фильтруется и усиливается.
Микроконтроллер по разности опорного сигнала и сигнала с фотодиода по специальному алгоритму вырабатывает ШИМ, которая управляет свечением диода выходного каскада. Светодиод управляется током, чем обусловлен вариант его включения.
Для связи с компьютером, программатором и другими микроконтроллерами реализуется набор интерфейсов. RS232 требует драйвер (ST232) для стабильной работы. Все интерфейсные выводы выводятся в отдельные выходы на схеме.
Обвеска микросхем обеспечивает их стабильную работу.
Замечания
Конденсаторы большой емкости (~0.1u) -- полярные.
Резисторы, помеченные меткой Trimmer -- подстроечные.
Напряжение с подстроечного резистора, задающего опорный уровень, снимается с третьего ("реостатного") вывода.
Номиналы других подстроечных резисторов могут варьироваться. Эти резисторы служат лишь только для подгона параметров схемы. После подгона на их место будут вмонтированы уже обычные резисторы.
Метки в виде красных ромбов -- места, которые в данный момент уточняются / могут поменяться.
| gharchive/issue | 2017-03-22T16:44:36 | 2025-04-01T06:44:39.939136 | {
"authors": [
"kalaider"
],
"repo": "kalaider/act-photo",
"url": "https://github.com/kalaider/act-photo/issues/26",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1522968397 | Switch to functional/stateless architecture
Currently, BYOC is tightly coupled to the app object. This has some downsides:
It's not possible to use BYOC without an app object, e.g. in simple scripts where a dictionary would otherwise suffice.
The app objects end up keeping a bunch of intermediate state for their entire lifetimes, even though (in principle) that state can be dropped once the "final" values are calculated.
I believe that moving to a functional/stateless architecture would address both of these points, and make the code more extensible in general. By "functional/stateless architecture", I mean roughly that there would be a load(params, configs) function that would return a dict/object with all the specified information.
However, it's not clear how all of the useful features of BYOC would be reimplemented in such a system. The purpose of this thread is to enumerate and discuss such features.
DocoptConfig: Using parameters when loading configs
UpstreamConfig: Reusing configs from other apps
Part: Deriving multiple parameters from a single config value
Never overwrite manually specified values
This is more of a danger with the load() function, since the natural thing to do would just be to overwrite everything.
I'd also need to take care to distinguish between values set by the user, and values set by a previous call to load.
Automatic initialization
One feature of the current system is that you don't need to manually initialize things. Just add a Template to a Pcr reaction, for example, and it will automatically fill itself when you go to use it. This is convenient, but it adds a lot of complexity and (in some cases) duplicate effort.
A functional architecture would instead calculate everything at once.
Config
Currently, all the config classes are initialized with the app object.
This isn't ideal:
All the configs really need is to load values from somewhere. They don't need the app to do this, they just need some of the information that might be contained in the app.
Move all this to the factories:
Configs have no knowledge of the apps
Config factories do, though, and use that knowledge to correctly instantiate configs.
Three ways to specify values for configs:
Factory kwargs
App attributes
Class variables
I can probably write some code to simplify this:
Define "setting" descriptors for config classes.
Factories find all settings and set as appropriate.
Can instruct factory to look for different attribute.
Class-based interface
The functional interface will require the configs to be completely configured and ready to go.
The class-based interface will try to be more streamlined:
The __config__ factories will take responsibility for actually configuring the configs.
All parameters are automatically calculated and installed when any one parameter is accessed.
Apps can't be half-initialized:
Either every parameter has a value, or none do.
I'm not totally sure if this will solve issues like #47. It will be necessary to set sample.parent before calling load(), but that doesn't seem to onerous. Configs are immutable, so it doesn't matter what order the configs are used in.
Apps can't be reloaded:
There's no perfect way to distinguish between a value set/modified by the user vs one set by load(), so once a value has been set, I need to respect it.
I could add a load(force=True) flag to override this.
Method getter
This intrinsically needs to know about the app.
Having thought about this for a while, I don't think I can transition stepwise to a stateless architecture:
Reagents can be added to samples at any time, so there's no way to know when a sample is "complete", i.e. ready to load. This means that I'd have to force the user to manually load each reagent and sample, and I just don't think that that API is acceptable.
Loading parameters at the time that they're accessed is a natural way around this problem. By the time you're accessing an attribute, it's reasonable (and intuitive to API users) to assume that all the prerequisite reagents have been added to the sample by then. Importantly, though, accessing one attribute doesn't imply that any others are ready, which is why it's not really tractable to have an API that automatically loads everything once the first attribute is accessed.
However, the stateless architecture makes much more sense for a general-purpose library: it's simpler, more efficient, and more extensible. Without a stateless architecture, I couldn't see myself recommending this library to other people, nor could I see myself using it for anything other than stepwise. So I think a stateless architecture is essential.
Putting this all together, I think I need to support both architectures. More specifically, I need to implement a stateful architecture using the stateless architecture. This would actually be a good way to demonstrate the extensibility of the stateless architecture.
Some implementation details I have in mind:
Apps would have some configuration option that specifies how parameters should behave when they need a value. The options would be (i) raise an exception explaining that load() needs to be called, (ii) attempt to load that parameter, or (iii) attempt to load all parameters.
Define a "context" object that would store all the state necessary to load parameters. This object would hold the app itself, config layers, resolved getters for each parameter, miscellaneous data structure to communicate between parameters (e.g. for Part), etc.
In stateless mode, the context object would be created within the load() function, stored in the app, and deleted once everything has been loaded.
In stateful mode, the context object would be permanently stored in the app and possibly updated each time a new parameter was accessed. It'd probably make sense to provide a cleanup() function that could get rid of it.
Currently I have the concept of bound/unbound getters/configs. Bound objects hold a pointer to the app itself, and can get access to all the caches stored in the app via that pointer. I'd get rid of this concept, and just have getters/configs receive the context object as an argument.
I'd want to have pretty smart cache objects. For example, say I want to cache the layers from a config. I'd pass the config the context object, which would contain the app being loaded. If the config doesn't actually use any information in the app, though, I could cache the resulting layers in such a way that they could be used by other apps.
Case study: the stepwise configuration file. Currently I have a stepwise config that looks in all the places where a stepwise config file could be found and returns a layer for each. This is app-independent. But the config also allows the user to specify a "root key", in which case the config becomes app-dependent. Maybe I want to allow configs to depend on other configs. In this example, there would be an app-independent stepwise config that just loads every file it can find, and another app-dependent stepwise config that loads the independent one then yields copies of all those layers with a different root key. Fundamentally this is just letting the config make use of the cache that's already there, but this seems like a good way to do it.
Regarding the docopt issue, where some parameters can be loaded multiple times:
Some options:
Store every value produced by a getter, and associate it with a version number. Every time a parameter value is needed, use the cached values unless they're out of date. This would require a very detailed cache, though, and I'd need to call the picker every time a value is accessed. I think this is too much.
Clear the parameter-value cache each time a new config is loaded. This is pretty simple, but wasteful because it doesn't make an effort to avoid recalculating things that aren't affected by the new config.
Allow getters to raise NotReadyYet. This would indicate that the parameter could change value in the future (e.g. once more configs are loaded), but for now this getter should just be skipped.
I'd like some way for this exception to indicate what it's waiting on, otherwise load() could get stuck in an infinite loop (by trying to keep going until all "pending" parameters have values). Maybe I could have the context keep an internal version number that increments every time something about the context changes. If I get NotReadyYet twice for the same parameter with the same context version number, that's an infinite loop.
| gharchive/issue | 2023-01-06T18:20:01 | 2025-04-01T06:44:40.000018 | {
"authors": [
"kalekundert"
],
"repo": "kalekundert/byoc",
"url": "https://github.com/kalekundert/byoc/issues/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2340614457 | feat: add random background image from commons
Pull Request
Description
Le background etait blanc
Changements Effectués
J'ai ajoute une image randomly from wikimedia commons
Liste de Vérification
[ X] Tester les changements localement.
[ X] Le code respecte les normes du projet.
Merci !
My suggestion: use a white background on the card. E.g.: chrome://whats-new/.
| gharchive/pull-request | 2024-06-07T14:35:11 | 2025-04-01T06:44:40.002920 | {
"authors": [
"Bam92",
"BirushaNdegeya"
],
"repo": "kaliacad/mostvisitedarticle",
"url": "https://github.com/kaliacad/mostvisitedarticle/pull/171",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
638763021 | Dual bootstrap support chapter in contributing docs is outdated
CONTRIBUTING.md does not reflect the changes to dual bootstrap support yet. Especially it still mentions the bootstrap version specific trees, which have been dropped in #992. On the other side it does not discuss the build macros introduced in #992 and follow-up merge requests.
Good catch! PRs welcome! 😜
| gharchive/issue | 2020-06-15T11:09:04 | 2025-04-01T06:44:40.004262 | {
"authors": [
"jelhan",
"simonihmig"
],
"repo": "kaliber5/ember-bootstrap",
"url": "https://github.com/kaliber5/ember-bootstrap/issues/1124",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1767504410 | Fix optional binding error in AppInfoListView.
Hello Maintainer,
I hope this message finds you well. While I was going through the codebase, I found a place where optional binding is being used even though it's not needed.
The relevant process will result in an error in the Xcode 14.3.1 development environment.
Allowing the licenseFileURL property to be passed directly, given that it is always non-nil
Best Regards,
@Nao-RandD Thank you for submitting your PR. I've looked over the error you had.
The policy of this library is to display cells when values are passed. So I've modified the licenseFileURL type. I've released 1.0.0.
I appreciate your patience!
| gharchive/pull-request | 2023-06-21T12:34:17 | 2025-04-01T06:44:40.108608 | {
"authors": [
"Nao-RandD",
"kamimi01"
],
"repo": "kamimi01/AppInfoList",
"url": "https://github.com/kamimi01/AppInfoList/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
51723550 | Mod FieldUtility.GetMediaUrl() to honor Media.AlwaysIncludeServerUrl
Modified FieldUtility.GetMediaUrl() to honor the Media.AlwaysIncludeServerUrl setting, if it is set. (Issue #13)
Thanks for the PR!
Per http://www.sitecore.net/Learn/Blogs/Technical-Blogs/John-West-Sitecore-Blog/Posts/2012/12/Sitecore-Idiosyncrasies-Media-URLs.aspx it's no longer needed to prefix the URL at all in 6.6+ (and it's for sure not on 7.2) so I was able to remove it all and just use MediaManager directly.
I believe you guys are still on 6.5, in which case you may need to keep the tweak in your fork.
| gharchive/pull-request | 2014-12-11T18:55:21 | 2025-04-01T06:44:40.122289 | {
"authors": [
"jeremyclifton",
"kamsar"
],
"repo": "kamsar/Synthesis",
"url": "https://github.com/kamsar/Synthesis/pull/14",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
129029185 | Additional configuration flags
It might be nice to have some additional flags that would further segregate many-configurationed projects. For example:
explicitSyncOnly => configuration is not synced when 'sync all' is used, and must be synced explicitly or included in a list of configs to sync
maxConcurrency => set the max number of threads to use when syncing or reserializing. Must be 1 if the config contains template items due to a Sitecore bug (see #80) but for content configs we can speed things up by allowing threading.
category => organize the control panel by categories of configuration, possibly collapsed by default, for organizational purposes. Should also be able to flag all configs in a category for sync.
Anyone else got ideas? Feedback? At present this is just an idea.
The most pressing feature I would have for this is the ability to specify wildcards as sync targets - and have Unicorn sort out the dependencies and ordering.
Something like:
/unicorn.aspx?verb=Sync&configuration=*
or even:
/unicorn.aspx?verb=Sync&configuration=Foundation.*^Feature.*^Project.Common.Website^Project.Habitat.Website
+1 We are in desperate need to be able to perform sync for all configurations available, i.e.:
unicorn.aspx?verb=Sync&configuration=All
That is already available, just don't specify the configuration parameter
and that implies all configurations.
On Thu, Mar 17, 2016 at 9:40 AM Dmitri Moore notifications@github.com
wrote:
+1 We are in desperate need to be able to perform sync for all
configurations available, i.e.:
unicorn.aspx?verb=Sync&configuration=All
—
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub
https://github.com/kamsar/Unicorn/issues/105#issuecomment-197966273
It is indeed! Thanks Kam.
3.1.5 will have sync all available with a single click :grinning:
Would that be in the right order based on dependencies?
Absolutely. Dependencies are resolved for all syncs.
On Mon, Mar 21, 2016 at 7:29 PM Thomas Eldblom notifications@github.com
wrote:
Would that be in the right order based on dependencies?
—
You are receiving this because you authored the thread.
Reply to this email directly or view it on GitHub
https://github.com/kamsar/Unicorn/issues/105#issuecomment-199592123
3.1.5 contains a sync all button with the URL mentioned above, as well as the maxConcurrency setting.
| gharchive/issue | 2016-01-27T04:48:24 | 2025-04-01T06:44:40.130904 | {
"authors": [
"Eldblom",
"demisx",
"kamsar"
],
"repo": "kamsar/Unicorn",
"url": "https://github.com/kamsar/Unicorn/issues/105",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1161931101 | New endpoint for dataset-download-options
Hi @kapadia, I have a use case where I need to be able to identify dataset download options (products) without knowing entity IDs. This isn't possible with the download-options endpoint used by api.download_options because it returns nothing if the entity ID list is empty, so usgs would need a new function and payload dataset_download_options that uses the dataset-download-options endpoint.
I'm happy to add those functions and make a PR if you think it's worth including in usgs. Let me know what you think!
@aazuspan - happy to take further contributions from you :) I believe that end point used to be implemented in an old version. Changes to the USGS API necessitated a re-write of this library, and that end point was left off in the process.
Great, I'll make a PR!
| gharchive/issue | 2022-03-07T21:13:56 | 2025-04-01T06:44:40.167596 | {
"authors": [
"aazuspan",
"kapadia"
],
"repo": "kapadia/usgs",
"url": "https://github.com/kapadia/usgs/issues/64",
"license": "isc",
"license_type": "permissive",
"license_source": "bigquery"
} |
2557951754 | Browser managed by organisation is crashing
Hi Team,
I am trying to automate Ui tests using karate on browsers managed by our organization. When I execute the test, Chrome browser opens but no page loads and could see the browser is crashed.
I have even set the http proxy in driver configuration.
I am getting error as below
4j.AssertionFailedError: driver config / start failed: failed to get reply for: [id: 4, sessionId: 2600608260560A0C50727E58FBEF7D3C, method: page.enable]
it is possible that there are security restrictions in your org. to reopen, please follow this process: https://github.com/karatelabs/karate/wiki/How-to-Submit-an-Issue
I feel the best way for you to proceed is to troubleshoot this at your end or consider using Docker
| gharchive/issue | 2024-10-01T01:07:44 | 2025-04-01T06:44:40.179187 | {
"authors": [
"Tagorekrn",
"ptrthomas"
],
"repo": "karatelabs/karate",
"url": "https://github.com/karatelabs/karate/issues/2611",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
644208935 | Error wrapping 3D keypoints, bounding boxes and camera intrinsics together
I get an error in step 6 in the pre processing scripts section - Wrapping 3D keypoints, bounding boxes and camera intrinsics together:
(other than the warning of 'camera 54138969 isn't present in S11/Directions-2)
Command:
python3 generate-labels-npy-multiview.py ../../../data/human36m/ ../../../data/human36m/extra/una-dinosauria-data/ ../../../data/human36m/extra/bboxes-Human36M-GT.npy
Error:
Traceback (most recent call last):
File "generate-labels-npy-multiview.py", line 58, in
cameras_params = h5py.File(os.path.join(una_dinosauria_root, 'cameras.h5'), 'r')
File "/home/jamal/.local/lib/python3.6/site-packages/h5py/_hl/files.py", line 408, in init
swmr=swmr)
File "/home/jamal/.local/lib/python3.6/site-packages/h5py/_hl/files.py", line 173, in make_fid
fid = h5f.open(name, flags, fapl=fapl)
File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
File "h5py/h5f.pyx", line 88, in h5py.h5f.open
OSError: Unable to open file (unable to open file: name = '../../../data/human36m/extra/una-dinosauria-data/cameras.h5', errno = 2, error message = 'No such file or directory', flags = 0, o_flags = 0)
Make sure you've extracted 'h36m.zip' at a previous step correctly so that ../../../data/human36m/extra/una-dinosauria-data/cameras.h5 exists.
Thats right. It was in a subfolder 'H36M' that was created when extracted.
Thanks
| gharchive/issue | 2020-06-23T23:22:23 | 2025-04-01T06:44:40.184131 | {
"authors": [
"jamalknight",
"shrubb"
],
"repo": "karfly/learnable-triangulation-pytorch",
"url": "https://github.com/karfly/learnable-triangulation-pytorch/issues/87",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1448082099 | Parse timestamp cookies for SCHEDULED and DEADLINE
Fix #15 completely.
Related: #50
This PR also follows setuptools_scm v7.0.0 to drop Python 3.6 support, since orgparse requires setuptools_scm.
Update: Python 3.6 support is already dropped in a later commit (0041aad535836b5faf540205c8c68d5e42b52e19).
I've slightly modify tox.ini in this PR to pass the tests (see 6426a6aee5fca36296fba874cdf8309331ec9542). Please let me know if I should open a separate PR for those modifications.
Just pinging this PR, as it's been 1.5 months since it was initially submitted.
released to pypi as 0.3.2 !
Thanks! I just tested the new version on PyPI and verified the fix.
| gharchive/pull-request | 2022-11-14T13:32:22 | 2025-04-01T06:44:40.195217 | {
"authors": [
"j3soon",
"karlicoss"
],
"repo": "karlicoss/orgparse",
"url": "https://github.com/karlicoss/orgparse/pull/60",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1795079100 | 🛑 Habilitación is down
In 82bb066, Habilitación (https://catalogo-vpfe-hab.dian.gov.co/User/Login) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Habilitación is back up in e8965da.
| gharchive/issue | 2023-07-08T19:38:58 | 2025-04-01T06:44:40.197769 | {
"authors": [
"karlos528"
],
"repo": "karlos528/fe",
"url": "https://github.com/karlos528/fe/issues/325",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
597185018 | Not finding my usb on HassOS 3.12
Hi,
I'm a total noob in both hassio and its addons.
I was very enthusiastic when I found this project, really looking forward to getting it to work. Yesterday I got my CKOZ 00/14 stick. Included it in the Eaton MRF and got it included in my setup. I went on to plug it into my rpi that is running HassOS 3.12, the USB lights up green and I see the little red light flickering when I turn on or off my devices.
The problem I'm facing is that I cant get the xcomfortd executable to run, most likely because it can't find my USB stick. I can't see it in the list of hardware:
The pi is definitely seeing the usb though, I'm just not able to communicate with it(?):
After creating the executable, adding the packages needed and running xcomfortd I'm getting this:
Please help me. What am I doing wrong?
I made some progress so this can be closed, decided to reinstall hassio on raspbian instead.
You are in luck, I'm working on a newer version of this code, which should be more stable and be more flexible. Another user is working on a hassio addon, which should make this much more user friendly as well. I can ping you when it's released to the public (mere days away now).
That is amazing news, yes please let me know. Looking forward to it!
@karloygard Any chance you will support heating sensors and/or the room controller touch in the new version? So that it will be possible turn on/off floor heating based on a temp read a sensor either in the room controller touch or a floor sensor?
And by the way, actually got it working after some fiddling around. So it's not that user unfriendly!
@karloygard Any chance you will support heating sensors and/or the room controller touch in the new version? So that it will be possible turn on/off floor heating based on a temp read a sensor either in the room controller touch or a floor sensor?
Might be possible with some assistance from you. Which devices do you have, and how have you configured them?
@karloygard
I have a few room controller touch devices, they have in built sensors in the devices, mine have attached an external sensor called pt1000 that monitors the temp of the actual floor (not the room).
These devices control heating actuators that turn on/off the floor heating in my bathrooms and hallways.
In addition to this i have something that came out before the room controller touch called temperature sensor, which is a device that is similar to the stand alone pt1000(i guess). It monitors and reports the temp in the floor so that it can send signals to the heating actuators to turn on/off.
My system is programmed using the ckoz 00/13, using the Eaton MRF software.
You can find my email adr in my profile page if you would like more details or want me to test something, I will happily contribute with whatever I can.
Check out https://github.com/karloygard/hassio-addons. Open up Home Assistant, go to Supervisor -> Add-On Store and add a new repository with that url.
If you send me the datapoints file mentioned in the documentation for the addon, I can try to make sure that your devices will work with this code.
Hi,
I have added a few of my varmekabler(!) and stuff here and exported the datapoints file from the usb stick.
Will try to get the add-on working today. Cool, thanks!
Geir
On 11 Apr 2020, 11:39 +0200, Karl Anders Øygard notifications@github.com, wrote:
Check out https://github.com/karloygard/hassio-addons. Open up Home Assistant, go to Supervisor -> Add-On Store and add a new repository with that url.
If you send me the datapoints file mentioned in the documentation for the addon, I can try to make sure that your devices will work with this code.
—
You are receiving this because you modified the open/close state.
Reply to this email directly, view it on GitHub, or unsubscribe.
1 bakdør utelys 5709360 16 0 0 0 #000#000#000#000#0#000#000#005#000#
2 1.etg sov 4 5740117 17 0 0 0 #000#000#000#000#0#000#000#006#000#
3 1.etg sov 3 5734031 17 0 0 0 #000#000#000#000#0#000#000#006#000#
4 1.etg sov 2 5740097 17 0 0 0 #000#000#000#000#0#000#000#006#000#
5 1.etg sov.1 5733990 17 0 0 0 #000#000#000#000#0#000#000#006#000#
6 1.etg bad speil 5736891 17 0 0 0 #000#000#000#000#0#000#000#006#000#
7 1.etg bad tak 5736925 17 0 0 0 #000#000#000#000#0#000#000#006#000#
8 1.etg stor gang 5734000 17 0 0 0 #000#000#000#000#0#000#000#006#000#
9 utelys hoveddør 5727252 16 0 0 0 #000#000#000#000#0#000#000#005#000#
10 1.etg tv stue 5733992 17 0 0 0 #000#000#000#000#0#000#000#006#000#
11 1.etg liten gang 5734068 17 0 0 0 #000#000#000#000#0#000#000#006#000#
12 1.etg kjøkken spotter 5734274 17 0 0 0 #000#000#000#000#0#000#000#006#000#
13 1.etg stue lampe spi 5741956 17 0 0 0 #000#000#000#000#0#000#000#006#000#
14 utelys terrasse 5733605 16 0 0 0 #000#000#000#000#0#000#000#005#000#
16 1.etg bad vk 5727387 16 0 0 0 #000#000#000#000#0#000#000#005#000#
20 1.etg dusj/bad vk 5727267 16 0 0 0 #000#000#000#000#0#000#000#005#000#
22 1.etg liten gang vk 5733608 16 0 0 0 #000#000#000#000#0#000#000#005#000#
24 1.etg stor gang vk 5733583 16 0 0 0 #000#000#000#000#0#000#000#005#000#
15 1.etg bad temp (Kanal B) 5881271 23 1 1 1 #000#000#000#000#0#000#000#006#000#
17 1.etg dusj/bad (ekstern Sensor (PT1000)) 6784078 78 6 1 1 #000#000#000#000#0#000#000#001#000#
18 1.etg dusj/bad (Touch-sensor A) 6784078 78 4 1 1 #000#000#000#000#0#000#000#001#000#
19 1.etg dusj/bad (Touch-sensor B) 6784078 78 5 1 1 #000#000#000#000#0#000#000#001#000#
21 1.etg liten gang temp (Kanal A) 5881353 23 0 1 1 #000#000#000#000#0#000#000#006#000#
23 1.etg stor gang temp (Kanal B) 5881341 23 1 1 1 #000#000#000#000#0#000#000#006#000#
Just replied to the email, guess we should move this discussion to the right place.
| gharchive/issue | 2020-04-09T10:30:26 | 2025-04-01T06:44:40.213411 | {
"authors": [
"geeewizzz",
"karloygard"
],
"repo": "karloygard/xcomfortd",
"url": "https://github.com/karloygard/xcomfortd/issues/9",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
173959272 | Error occurred when building node - bin/karma start karma.conf.js --no-auto-watch --single-run --no-color
Steps to reproduce the behaviour
1. I have the node version of v0.12.0
2. Karma version is 0.13.19
3. npm version is 2.5.1
when building the build fails saying
node node_modules/karma/bin/karma start karma.conf.js --no-auto-watch --single-run --no-color
[ERROR] Return code: 1
[ERROR] Command output:
[ERROR] | ...(1513 lines of output hidden)...
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 107 of 122 (17 FAILED) (skipped 1) (0 secs / 1.221 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 108 of 122 (17 FAILED) (skipped 1) (0 secs / 1.268 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 109 of 122 (17 FAILED) (skipped 1) (0 secs / 1.278 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 110 of 122 (17 FAILED) (skipped 1) (0 secs / 1.286 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 111 of 122 (17 FAILED) (skipped 1) (0 secs / 1.294 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 112 of 122 (17 FAILED) (skipped 1) (0 secs / 1.302 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 113 of 122 (17 FAILED) (skipped 1) (0 secs / 1.309 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 114 of 122 (17 FAILED) (skipped 1) (0 secs / 1.318 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 115 of 122 (17 FAILED) (skipped 1) (0 secs / 1.325 secs)
[ERROR] | PhantomJS 2.1.1 (Windows 8 0.0.0): Executed 115 of 122 (17 FAILED) (skipped 7) (0.932 secs / 1.325 secs)
Any suggestions why this is happening ?
What is the behavior that you are reporting here?
Is it the failure of the script from the test failures? If so, that is expected behavior by Karma. Otherwise, more details should probably be given, as there is not enough to adequately address what is being attempted to be asked, and guessing isn't a good approach :) .
No feedback in a year. Closing.
| gharchive/issue | 2016-08-30T08:29:43 | 2025-04-01T06:44:40.230135 | {
"authors": [
"EzraBrooks",
"SManorathna",
"wesleycho"
],
"repo": "karma-runner/karma",
"url": "https://github.com/karma-runner/karma/issues/2338",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
159702723 | fix(launcher): send sigkill on timeout when force killing
@dignifiedquire This is in response to bug we are seeing where karma hangs because it is unable to force kill chrome, possibly https://github.com/karma-runner/karma/issues/24#issuecomment-216249098
I signed it!
Thanks, can you change the commit message to befix(launcher: instead of bug please?
@dignifiedquire, i updated the commit message, thanks.
@dignifiedquire Any ETA on this one landing?
| gharchive/pull-request | 2016-06-10T19:13:20 | 2025-04-01T06:44:40.232711 | {
"authors": [
"dignifiedquire",
"kakadiya91",
"modulesio"
],
"repo": "karma-runner/karma",
"url": "https://github.com/karma-runner/karma/pull/2169",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
614448518 | Refactoring code is done
Hi Karthik,
Please review the code and approve.
Thanks
Likhil.
@Likhil Test the code and confirm.
| gharchive/pull-request | 2020-05-08T01:53:29 | 2025-04-01T06:44:40.310740 | {
"authors": [
"Likhil",
"karthikreddykuna"
],
"repo": "karthikreddykuna/DXC-Industrialized-AI-Starter",
"url": "https://github.com/karthikreddykuna/DXC-Industrialized-AI-Starter/pull/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2480565145 | Make it compatible with Bun.
Fast server, slow runtime. Doesn't that mismatch? Basically, uWS is the only package that complaining. But otherwise, it'd be damn good to run hyper on bun.
ahh, it was requested already.
| gharchive/issue | 2024-08-22T11:38:14 | 2025-04-01T06:44:40.318542 | {
"authors": [
"mr-moon"
],
"repo": "kartikk221/hyper-express",
"url": "https://github.com/kartikk221/hyper-express/issues/296",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
442538997 | Enable file based memory backend
Enable file based memory backing at /dev/shm on the host which will be used primarily (as of this writing) to be virtio-fs cache.
Hi,
Currently, I can only start kata container with virtiofs by setting "enable_hugepages = true" in /etc/kata-containers/configuration.toml. Otherwise, kata will fail to create container using virtiofs.
Do you think it's the same issue here?
Yes @renzhengeek - that’s what Ganesh is working on resolving.
| gharchive/issue | 2019-05-10T04:59:17 | 2025-04-01T06:44:40.374674 | {
"authors": [
"egernst",
"ganeshmaharaj",
"renzhengeek"
],
"repo": "kata-containers/runtime",
"url": "https://github.com/kata-containers/runtime/issues/1656",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
361593680 | netmon: Add signals handler support
This PR factorizes the signal handling code through a new
runtime package as a first step.
And the second commit relies on this new package in order
to setup the signal handling of the network monitor.
Fixes #718
LGTM
Thanks @sboeuf !
| gharchive/pull-request | 2018-09-19T06:16:08 | 2025-04-01T06:44:40.376228 | {
"authors": [
"WeiZhang555",
"jodh-intel",
"sboeuf"
],
"repo": "kata-containers/runtime",
"url": "https://github.com/kata-containers/runtime/pull/755",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1658483967 | destop verssion review
good job you did well on this project
✔️ GOOD JOB
[x] please kindly align your content on the about section yours is at center kindly align then at the start on "the language part"
[x] also align the drop down arrows to be same as figma design
[x] In your form kindly change the color of the place holder for "write your message here"
Thank you for your Recommendations, all effected
| gharchive/issue | 2023-04-07T07:37:07 | 2025-04-01T06:44:40.405458 | {
"authors": [
"katfogy",
"techmoves"
],
"repo": "katfogy/setup-and-mobile-first",
"url": "https://github.com/katfogy/setup-and-mobile-first/issues/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
540033309 | Setting the bar height after creation
Hi,
I know that the bar height can be set when creating a new Waveform object, but you can't adjust the height of the bars afterward.
I was wondering if it was it's possible to implement a new function, setBarHeight()?
It might work similar to the zoom() function, just with adjusting the bar heights then redrawing the waveform?
use wavesurer.setHeight(height) to set the height of the waveform. If it uses bars, those will be set as well.
https://wfplayer.js.org/
Here's an example of what I was more meaning. If you drag the Wave Scale range input at the bottom of the options given, it'll increase or decreaes the bar heights without changing the height of the canvas that holds it.
I was wondering if it was it's possible to implement a new function, setBarHeight()?
pull requests are welcome.
| gharchive/issue | 2019-12-19T02:19:28 | 2025-04-01T06:44:40.411933 | {
"authors": [
"RobTallGuy",
"thijstriemstra"
],
"repo": "katspaugh/wavesurfer.js",
"url": "https://github.com/katspaugh/wavesurfer.js/issues/1830",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
675557291 | Add a display option that disregards file structure
We we run jest from the command line, it displays the tests as they are logically structured, not as they are distributed on disk. I wish JTE had a display option similar to that.
I don't want to see a tree that reflects file organization. I want to see a tree composed of describe and it/test blocks, without any visual indication of the file structure.
Came here to say the same thing. I'm looking at the code now and it seems like a pretty straightforward thing to add, so I'll try to get a PR up shortly.
Hi @tomprogers / @dannymcgee,
Question for you both: can you look at the following page on the host extension site and tell me if it meets your needs?
https://github.com/hbenl/vscode-test-explorer
The setting testExplorer.mergeSuites
@rossknudsen Unfortunately not for me. I do have that option enabled and I can see why it's useful, but the pain point I'm trying to address is that my test files are all pretty deeply nested inside my workspace directory structure, so I need to expand 5 or 6 nodes in the UI to see my first actual test suite, which leads to a lot of wasted horizontal space in the sidebar and a lot of clicking to see anything. If you take a look at the screenshots in my PR, the issue I'm having should be more clear.
nah, that's fine. I just thought I'd check before merging. I'll have another look later on, but the initial review looked good.
Heya! Thanks for following up.
I think I'm with danny on this. I do use the mergeSuites setting, but it's still not enough. As an example, I'm working on an Electron app, which adds another layer of folders to accommodate the main and renderer threads. By the time an actual describe appears, it's already indented like 4 levels. When I run jest, I don't see any of that wrapping FS structure.
| gharchive/issue | 2020-08-08T17:10:35 | 2025-04-01T06:44:40.432229 | {
"authors": [
"dannymcgee",
"rossknudsen",
"tomprogers"
],
"repo": "kavod-io/vscode-jest-test-adapter",
"url": "https://github.com/kavod-io/vscode-jest-test-adapter/issues/79",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
624185086 | いいところ診断の結果に優しさを追加したい
以下の結果を追加したい
'{userName}のいいところは優しさです。あなたの優しい雰囲気や立ち振る舞いに多くの人が癒やされています。'
これから対応します。
| gharchive/issue | 2020-05-25T09:50:10 | 2025-04-01T06:44:40.433358 | {
"authors": [
"kawakami-takayuki"
],
"repo": "kawakami-takayuki/assessment",
"url": "https://github.com/kawakami-takayuki/assessment/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
193461804 | Fix for missing metric_key attribute in Totals
Fixes #70 Fixed the getattr expression for operations so that it can handle a missing metric_key such as in the case of Totals
Coverage increased (+0.09%) to 97.645% when pulling ebf2d97d1bb81683c5c5bb45dec0d27a32a3396b on totals_fix into 40cec66443ea494e559f68abe6a389b401acf562 on master.
| gharchive/pull-request | 2016-12-05T10:34:00 | 2025-04-01T06:44:40.442412 | {
"authors": [
"coveralls",
"twheys"
],
"repo": "kayak/fireant",
"url": "https://github.com/kayak/fireant/pull/71",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
265948319 | TASK-1161 Store sub events in database
Now stores sub events in database and generates an ID for the sub event.
Does not yet check to avoid storing duplicate events or reprocessing
successful sub events.
Removed the id field from StatusEvent (now in StatusEventWithID).
Remove a not particularly helpful level of indirection by marking event
processing state in the storage system with a direct storage call,
rather than via an iterator then a queue then the storage system
Merging per @sychan
| gharchive/pull-request | 2017-10-16T23:36:50 | 2025-04-01T06:44:40.453100 | {
"authors": [
"MrCreosote"
],
"repo": "kbase/KBaseSearchEngine",
"url": "https://github.com/kbase/KBaseSearchEngine/pull/48",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
111393784 | Does 'Turbolinks Support' section in the documentation require updating?
I was looking at the documentation about the forward/back button in the 'Turbolinks Support' and I was trying to ensure my code is executed in this same manner for my project. I added the following code (which was taken from the documentation) to my application js:
$(document).on('page:restore', function(){
// Manually evaluates the appended script tag.
Paloma.executeHook();
});
This would not work for me. I had to modify it as below:
$(document).on('page:restore', function(){
// Manually evaluates the appended script tag.
Paloma.executeHook();
Paloma.engine.start();
});
Is what I've done correct and the documentation just needs updating? Or have I missed something else?
This might be the case since Paloma.start() is no longer called automatically and needs to always be explicitly called.
| gharchive/issue | 2015-10-14T13:05:22 | 2025-04-01T06:44:40.478262 | {
"authors": [
"brendon",
"sb89"
],
"repo": "kbparagua/paloma",
"url": "https://github.com/kbparagua/paloma/issues/82",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1528558908 | CRD convert error: no openapi schema found
Bug Report
Please answer these questions before submitting your issue. Thanks!
1. Minimal reproduce step (Required)
unzip the following file and put it in /oam.yaml
oam.yaml.zip
run kcl-openapi generate model --crd --skip-validation -f Desktop/oam.yaml
2. What did you expect to see? (Required)
KCL files generated
3. What did you see instead (Required)
error:
could not generate swagger spec: Desktop/oam.yaml, err: no openapi schema found in the crd file. Please check following fields: spec.Validation.OpenAPIV3Schema, spec.Versions.0.Schema
4. What is your KusionStack components version? (Required)
This bug also extends to CRD from spec V1. It also seems to be that this condition is triggered for all CRD model generation.
eg. crd for external-secrets operator
kcl-openapi generate model --crd \
-f <(curl -sS "https://raw.githubusercontent.com/external-secrets/external-secrets/main/config/crds/bases/external-secrets.io_externalsecrets.yaml") \
--skip-validation
could not generate swagger spec: /dev/fd/11, err: no openapi schema found in the crd file. Please check following fields: spec.Validation.OpenAPIV3Schema, spec.Versions.0.Schema
| gharchive/issue | 2023-01-11T07:27:37 | 2025-04-01T06:44:40.491704 | {
"authors": [
"amyXia1994",
"prahaladramji"
],
"repo": "kcl-lang/kcl-openapi",
"url": "https://github.com/kcl-lang/kcl-openapi/issues/31",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1374707973 | :sparkles: Add cluster-wide resources support to the resource reconciler
Summary
This PR enables the Resource reconciler to label cluster-wide resources with the available synctargets for a workspace with the desired state of "pending".
Currently, the list of allowed cluster-wide resources is hardcoded to "persistentvolumes" in preparation for upcoming storage work.
Related issue(s)
Fixes https://github.com/kcp-dev/kcp/issues/1890
hack/verify-go-versions.sh
[54](https://prow.ci.openshift.org/view/gs/origin-ci-test/pr-logs/pull/kcp-dev_kcp/1998/pull-ci-kcp-dev-kcp-main-images/1572606912959090688#1:build-log.txt%3A54)
go build -ldflags="-X k8s.io/client-go/pkg/version.gitCommit=bf76c43a -X k8s.io/client-go/pkg/version.gitTreeState=clean -X k8s.io/client-go/pkg/version.gitVersion=v1.24.3+kcp-v0.8.0-136-gbf76c43a7d64c0 -X k8s.io/client-go/pkg/version.gitMajor=1 -X k8s.io/client-go/pkg/version.gitMinor=24 -X k8s.io/client-go/pkg/version.buildDate=2022-09-21T15:30:04Z -X k8s.io/component-base/version.gitCommit=bf76c43a -X k8s.io/component-base/version.gitTreeState=clean -X k8s.io/component-base/version.gitVersion=v1.24.3+kcp-v0.8.0-136-gbf76c43a7d64c0 -X k8s.io/component-base/version.gitMajor=1 -X k8s.io/component-base/version.gitMinor=24 -X k8s.io/component-base/version.buildDate=2022-09-21T15:30:04Z" -o bin ./cmd/...
[55](https://prow.ci.openshift.org/view/gs/origin-ci-test/pr-logs/pull/kcp-dev_kcp/1998/pull-ci-kcp-dev-kcp-main-images/1572606912959090688#1:build-log.txt%3A55)
/usr/local/go/pkg/tool/linux_amd64/link: signal: killed
[56](https://prow.ci.openshift.org/view/gs/origin-ci-test/pr-logs/pull/kcp-dev_kcp/1998/pull-ci-kcp-dev-kcp-main-images/1572606912959090688#1:build-log.txt%3A56)
make: *** [Makefile:92: build] Error 1
/retest
/hold adding missing functionality to make this PR complete
/unhold
| gharchive/pull-request | 2022-09-15T15:24:09 | 2025-04-01T06:44:40.494436 | {
"authors": [
"jmprusi"
],
"repo": "kcp-dev/kcp",
"url": "https://github.com/kcp-dev/kcp/pull/1998",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1479983741 | WIP 🌱 Syncer: Cleanup of DNS-related resources
Summary
Cleanup of DNS-related resources when no more downstream namespaces exist for a given upstream workspace.
Disclaimer: still missing tests + e2e tests.
Related issue(s)
Fixes #2358
Can we do this with actual garbage collection and owner refs?
I think yes. Namespaced dependents can have cluster-scoped owners (Namespaces in this case).
Not sure about how you would use garbage collection in the specific case.
Do you mean that every DNS-related resource for a given KCP workspace should have as many owner references as the number of downstream namespaces originating from the given KCP workspace ?
In order to do this, in the EnsureDNSUpAndReady function we would need to:
check that every DNS-related resource has the namespace of the currently-synced resource in its owner refs,
update every DNS-related resource if necessary to add the namespace owner ref.
This would be required during every resource sync. I'd rather do more work during cleanup, and keep the resource syncing main reconciliation loop as quick as possible.
Then I assume that the DNS-related resources deletion would be triggered by cascading deletion when all the namespaces related to the given KCP workspace would be deleted. However we wouldn't control when this cleaning would happen, which would make it hard to synchronize this cleaning with the on-the-fly creation of those DNS-related resources when they are necessary (in the EnsureDNSUpAndReady function). Such a synchronization is for now still missing from this PR, but might be necessary and should be thought of.
Last point: when we start implementing the network isolation through NetworkPolicies on the physical cluster, we will need this concept of KCP tenant on downstream (== a originating KCP workspace + related SyncTarget). That seemed quite natural to use it here as well.
So, unless I'm missing something, applying the standard Kubernetes garbage collection and owner refs mechanism here is not as simple as it may seem, and may bring more slowness and complexity.
Let me finally just point to a previous PR comment answer to remind that we are not in the standard Kubernetes case, because everything downstream is finally driven by the individual resources synced from upstream.
@davidfestal are you still interested in moving forward with this?
/hold
on hold until PR https://github.com/kcp-dev/kcp/pull/2675 is merged.
/hold
Wait until PRs #2675 and #2946 are merged.
/close
TMC
| gharchive/pull-request | 2022-12-06T19:27:55 | 2025-04-01T06:44:40.501502 | {
"authors": [
"davidfestal",
"mjudeikis",
"ncdc"
],
"repo": "kcp-dev/kcp",
"url": "https://github.com/kcp-dev/kcp/pull/2455",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1530268725 | 🐛 IndexByLogicalClusterPathAndName should return clusterpath and name
Signed-off-by: Jian Qiu jqiu@redhat.com
Summary
I think IndexByLogicalClusterPathAndName should return clusterPath+Name by default?
Related issue(s)
Fixes #
The existing code effectively seemed wrong. But are we aware of something broken by this erroneous code ?
If yes, shouldn't we add a test case (unit test or e2e) that corresponds to this fix ?
Nice finding.
/lgtm
/approve
| gharchive/pull-request | 2023-01-12T08:00:32 | 2025-04-01T06:44:40.504593 | {
"authors": [
"davidfestal",
"qiujian16",
"sttts"
],
"repo": "kcp-dev/kcp",
"url": "https://github.com/kcp-dev/kcp/pull/2606",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
565329709 | Update library to compile on upcoming 0.14.0 PS release
For context, see https://discourse.purescript.org/t/updating-the-ecosystem-for-upcoming-0-14-0-release/1144
remind me to look at this, since i'm not an admin of this repo
| gharchive/pull-request | 2020-02-14T13:34:02 | 2025-04-01T06:44:40.506013 | {
"authors": [
"JordanMartinez",
"justinwoo"
],
"repo": "kcsongor/purescript-record-format",
"url": "https://github.com/kcsongor/purescript-record-format/pull/7",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
801844483 | 🛑 营销平台API is down
In 3d6cde8, 营销平台API (https://api.dgshare.com) was down:
HTTP code: 502
Response time: 3258 ms
Resolved: 上海购物API is back up in 732cfb9.
| gharchive/issue | 2021-02-05T04:57:47 | 2025-04-01T06:44:40.513080 | {
"authors": [
"amorist"
],
"repo": "kdcer/upptime",
"url": "https://github.com/kdcer/upptime/issues/91",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
217728593 | can I use this to transpose a musicxml file
Trying to create an app for myself (browser-based) that will show musicxml docs (rendered in JS) in different keys (transpose).
If I can't do it with this gem do you know a tool in ruby or JS that can transpose musicxml?
Thanks
Yeah you could use this. It might be easier to check out something like vexflow (https://github.com/0xfe/vexflow).
| gharchive/issue | 2017-03-29T00:09:43 | 2025-04-01T06:44:40.514713 | {
"authors": [
"benlieb",
"kddeisz"
],
"repo": "kddeisz/musicxml",
"url": "https://github.com/kddeisz/musicxml/issues/9",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
195136884 | Zoom to current location
Query current device location and zoom to that point.
Done in cac39a7f65d570da19705bf46108e8a9cb72daab
| gharchive/issue | 2016-12-13T01:42:08 | 2025-04-01T06:44:40.515740 | {
"authors": [
"kdeloach"
],
"repo": "kdeloach/septa-viz",
"url": "https://github.com/kdeloach/septa-viz/issues/4",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2730754264 | MicroPython for ESP32
I know this is a long shot, but by any chance have you made the MicroPython to use this on an ESP32? I just come across this repository when searching for ways to communicate with my Amana Furnace w/ Comfortnet. So far, I've been able to get data from the diagnostic port using an ESP32 and RS-485 module using ESPHome. I'm was hoping to get some information into Home-Assistant, but I way beyond capabilities at this point. I can see usable information when I manually go into the comfortnet settings in the thermostats and select some of the status options.
[11:14:40][D][uart:136]: 0x01 | 0xFF | 0x02 | 0x02 | 0x0200 | 0xA5 | 0x41 | 0xA0 | 17 | 06 00 00 09 10 0C 03 3F 57 02 78 E9 BE 12 A1 29 1D | .......?W.x....). | 0x3A99
[11:14:41][D][uart:089]: Dest | Src | Subnet | Meth | Params | SrcNode | MsgType | PktNum | Len | Payload HEX | Payload ASCII | Checksum
[11:14:41][D][uart:136]: 0x01 | 0xFF | 0x02 | 0x02 | 0x0200 | 0x02 | 0xC1 | 0x20 | 45 | 01 05 00 00 00 F0 B6 C5 D5 53 54 41 54 55 53 B7 C2 E2 4D 4F 44 45 20 20 35 30 25 20 48 45 41 54 20 C3 C2 E2 43 46 4D 20 20 33 32 30 C3 | .........STATUS...MODE 50% HEAT ...CFM
I never finished this library. Originally what I had intended to do was to connect an ESP32 just as you have and have it send the packet data to a PC where this library would have been running.
The biggest issue with ComfortNet is while it was a standard it was not written in a manner that made manufacturers adhere to the standard. A lot of the manufacturers deviated from the original specification and added to it and those additions are specific to the make and model of furnace. This is something that was done almost across the board across all manufacturers. That is the reason why you cannot use a Rheem Communicating thermostat on an Amana furnace even tho both furnaces use ComfortNet. I didn't want to have to invest in a 600 dollar thermostat for Rheem in order to figure out what those custom commands are. That is why I didn't continue with the development of this library,...
I didn't wrote the library to run on the ESP32 because at the time the ESP32 didn't have enough memory to be able to do it properly. the addition of the PSRAM didn't happen until later.
Since technology has improved with the ESP32 I could start writing a version of this library that would run on MicroPython. There would be enough memory available to do that. It would actually pair together quite well with the GUI framework binding that I have written that runs on the ESP32 with MicroPython. This would be nice because a user could use an ESP32-S3n16r8 with a 7" RGB or a 5" I8080 touch panel and they could make a thermostat with full blown touch UI. This would be very easy to make.
If you are willing to spend the time to debug/test and data log I am willing to start over and write it so it will run on MicroPython.
Here is a link to that GUI Framework binding if you wanted to look at it.
https://github.com/lvgl-micropython/lvgl_micropython
| gharchive/issue | 2024-12-10T17:36:58 | 2025-04-01T06:44:40.561974 | {
"authors": [
"kdschlosser",
"smurf12345"
],
"repo": "kdschlosser/ClimateTalk",
"url": "https://github.com/kdschlosser/ClimateTalk/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
260942960 | Integrate Top Visited Blog posts onto website
Plugin is already part of the project, needs only to be added to the theme.
.depo
| gharchive/issue | 2017-09-27T11:40:49 | 2025-04-01T06:44:40.583114 | {
"authors": [
"kealsera"
],
"repo": "kealsera/rikki",
"url": "https://github.com/kealsera/rikki/issues/87",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
284471584 | Handle as User error
}},"exceptionId":"docker-bbe20a94b9c309ce2f3c960f089b9ca0","exception":{"class":"Keboola\Syrup\Exception\ApplicationException","message":"Application error: 147946154733.dkr.ecr.us-east-1.amazonaws.com/developer-portal-v2/keboola.ex-google-analytics-v4:2.2.3 container '341349562-341349546.341349563--0-keboola-ex-google-analytics-v4' failed: (2) [2017-12-25 22:25:31] ex-google-analytics.ERROR: <meta name=viewport ... temporary error and could not complete your request.Please try again in 30 seconds. That\u2019s all we know.</ins> [] []","code":0,"attachment":"https://connection.keboola.com/admin/utils/logs?file=2017/12/25/22/2017-12-25-22-25-32-5a416cccf2054-exception"},"attachment":"https://connection.keboola.com/admin/utils/logs?file=2017/12/25/22/2017-12-25-22-25-33-5a416ccd02f69-log"}
https://keboola.slack.com/archives/C09U3R1J4/p1514237228000045https://keboola.slack.com/archives/C09U3R1J4/p1514237228000045
"output": "Running query 'mall cz + sk + pl + hu + hr + sl'",
| "errorOutput": "[2017-12-26 12:17:54] ex-google-analytics.ERROR: <meta name=viewport content="initial-scale=1, minimum-scale=1, width=device-width"> Error 502 (Server Error)!!1</title>
JJ, o tomto viem:
https://issuetracker.google.com/issues/69349753
Bohuzial v Googli na to seru. A deje sa to pravidelne len u tohto jedneho klienta pokial viem.
Retry to robi, ale nepomoze to. Malo by to pak ale hodit User Error.
| gharchive/issue | 2017-12-25T22:25:19 | 2025-04-01T06:44:40.592811 | {
"authors": [
"MiroCillik",
"odinuv"
],
"repo": "keboola/google-analytics-extractor",
"url": "https://github.com/keboola/google-analytics-extractor/issues/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2334420243 | Allow CA dirs to be specified beyond /custom/ca/
https://github.com/kedacore/keda/pull/5859
Checklist
[X] I have verified that my change is according to the deprecations & breaking changes policy
[X] Commits are signed with Developer Certificate of Origin (DCO - learn more)
[X] README is updated with new configuration values (if applicable) learn more
[X] A PR is opened to update KEDA core (repo)
Fixes https://github.com/kedacore/keda/issues/5860
The helm chart will fail to deploy a working KEDA operator until https://github.com/kedacore/keda/pull/5859 is merged
Thanks!
| gharchive/pull-request | 2024-06-04T21:41:20 | 2025-04-01T06:44:40.599813 | {
"authors": [
"joelsmith",
"zroubalik"
],
"repo": "kedacore/charts",
"url": "https://github.com/kedacore/charts/pull/650",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2014964053 | Client-side GRPC load balancing
Proposal
Keda currently uses the pick_first load balancing policy for GRPC connections which is the grpc library default. This currently works in most setups because of default coredns configurations. However, if we're not using the default settings, Keda should perform client-side round-robin load balancing for grpc connections aswell.
Use-Case
Some Kubernetes environments customize the default DNS configurations; For example:
disabling the loadbalance CoreDNS plugin
and talking to a headless service
the above changes means DNS requests will always return the results in the same order and won't randomize the ordering, ex with 2 pods:
root@app-5877ff4bd9-p27mk:/# dig headless-svc.default.svc.cluster.local +short
10.1.3.215
10.1.3.210
In this case, Keda should round-robin load balance grpc connections... otherwise it's always going to talk to the same pod and not send any requests off to the second one; this is a simple 1-line config change built into the grpc library; and shouldn't affect existing setups.
Is this a feature you are interested in implementing yourself?
Yes
Anything else?
No response
I'd say that this is already done xD
| gharchive/issue | 2023-11-28T17:38:09 | 2025-04-01T06:44:40.603819 | {
"authors": [
"BojanZelic",
"JorTurFer"
],
"repo": "kedacore/keda",
"url": "https://github.com/kedacore/keda/issues/5224",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
895291712 | Pass deepCopy objects to the polling goroutines
Signed-off-by: Zbynek Roubalik zroubali@redhat.com
Pass deepCopy of ScaledObject/ScaledJob objects to the scaleLoop polling goroutines it's a precaution to not have global objects shared between threads.
It should potentially help with the issue, which occurs if there is a big number of scalers running. Sometimes there are periodic panics in the json Marshall code that appeared to be related to passing the scaler objects to the background go routines and causing race conditions between fetching and serializing in different goroutines.
Checklist
[x] Commits are signed with Developer Certificate of Origin (DCO - learn more)
[x] Changelog has been updated
Actually as I'm looking at this harder, do we want to copy them here or at the point we actually launch the goroutines?
Actually as I'm looking at this harder, do we want to copy them here or at the point we actually launch the goroutines?
https://github.com/kedacore/keda/blob/347e8975d5f7b7ccc8e2bfc15311c47fd267707d/pkg/scaling/scale_handler.go#L104-L105
I think the way this is now, both goroutines would get the same reference.
Yeah, you are right, that's much better spot. Let me change that.
| gharchive/pull-request | 2021-05-19T10:51:43 | 2025-04-01T06:44:40.607659 | {
"authors": [
"coderanger",
"zroubalik"
],
"repo": "kedacore/keda",
"url": "https://github.com/kedacore/keda/pull/1812",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1291612731 | chore: bump-deps
Signed-off-by: Jorge Turrado jorge_turrado@hotmail.es
Checklist
[x] Commits are signed with Developer Certificate of Origin (DCO - learn more)
Relates to https://github.com/kedacore/keda/pull/3310
Relates to https://github.com/kedacore/keda/pull/3311
Relates to https://github.com/kedacore/keda/pull/3312
Relates to https://github.com/kedacore/keda/pull/3317
Relates to https://github.com/kedacore/keda/pull/3319
Relates to https://github.com/kedacore/keda/pull/3316
/run-e2e
Tagged the PRs that we have closed
| gharchive/pull-request | 2022-07-01T16:36:34 | 2025-04-01T06:44:40.611778 | {
"authors": [
"JorTurFer",
"tomkerkhove"
],
"repo": "kedacore/keda",
"url": "https://github.com/kedacore/keda/pull/3321",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2454303471 | [feat] Implement fee function
Implement function to calculate transaction fee:
https://github.com/keep-starknet-strange/raito/blob/4bf3aaba4d04f53255f32d022d3ee860ad20da8d/src/validation.cairo#L44
ZeroSync version:
https://github.com/ZeroSync/ZeroSync/blob/f1bd4736b0bac190ac48cbc8d5fcd92e2e1be7b9/src/transaction/transaction.cairo#L262
Hi, can I take this?
| gharchive/issue | 2024-08-07T20:35:30 | 2025-04-01T06:44:40.657477 | {
"authors": [
"maciejka",
"stevencartavia"
],
"repo": "keep-starknet-strange/raito",
"url": "https://github.com/keep-starknet-strange/raito/issues/39",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2090303704 | Added X Threads Doc
Added Previously Published Threads on X.
However @oluwaseundotcom I think having a docs/ folder, with inside a x_threads.md file that lists tweets about unruggable can be a good idea!
However @oluwaseundotcom I think having a docs/ folder, with inside a x_threads.md file that lists tweets about unruggable can be a good idea!
yes please. docs/social/x/threads.md
| gharchive/pull-request | 2024-01-19T11:13:44 | 2025-04-01T06:44:40.659487 | {
"authors": [
"AbdelStark",
"enitrat",
"oluwaseundotcom"
],
"repo": "keep-starknet-strange/unruggable.meme",
"url": "https://github.com/keep-starknet-strange/unruggable.meme/pull/150",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
841725737 | Application pdf updated
Link to Issue
Description of changes
Screenshot(s) or GIF(s) of changes
@steffen12 tests are still failing btw
| gharchive/pull-request | 2021-03-26T08:43:15 | 2025-04-01T06:44:40.690347 | {
"authors": [
"crchong1",
"steffen12"
],
"repo": "keepid/keepid_client",
"url": "https://github.com/keepid/keepid_client/pull/290",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
577303024 | AU331 机器学习与知识发现 - IEEE.ICU
https://ieee.icu/#/courses/grade-3/AU331
上海交通大学 IEEE 试点班,一个充满了「神仙」课程的专业,本网站旨在记录这些课程的信息以及历届学生对其的评价
这门课和同期的AU332类似,都是属于老师讲的不错,然后任务量较多的类型。我认为值得好好学
| gharchive/issue | 2020-03-07T08:50:42 | 2025-04-01T06:44:40.702392 | {
"authors": [
"AlexanderKirei",
"ieee-icu"
],
"repo": "keithnull/ieee.icu",
"url": "https://github.com/keithnull/ieee.icu/issues/57",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2353956511 | HentaiNexus: unable to fetch some thumbnails
Source information
HentaiNexus 1.4.10
Source language
English
Steps to reproduce
Open HentaiNexus and browse a bit
Expected behavior
Every title has a thumbnail
Actual behavior
A few titles appear without a thumbnail
Mihon/Tachiyomi version
TachiSY 1.10.5
Android version
Android 13
Other details
Looking at the webpages, the thumbnails are not in a .card-image img anymore, I think that the extension is actually fetching the first page of the title instead, but I'm not sure why the behavior is inconsistent when fetching.
Acknowledgements
[X] I have searched the existing issues and this is a new ticket, NOT a duplicate or related to another open or closed issue.
[X] I have written a short but informative title.
[X] I have updated the app to version 0.15.3.
[X] I have updated all installed extensions.
[X] I have tried the troubleshooting guide.
[X] If this is an issue with the app itself, I should be opening an issue in the app repository.
[X] I will fill out all of the requested information in this form.
Oh sorry, I searched it a couple of days ago but only really had time to write the issue now. I'll close this issue and add the comment on the other issue.
Actually, searching is:issue is:open nexus still returns nothing, that's why I didn't see it, I should have searched the full name instead.
| gharchive/issue | 2024-06-14T19:25:44 | 2025-04-01T06:44:40.708573 | {
"authors": [
"MediocreLegion"
],
"repo": "keiyoushi/extensions-source",
"url": "https://github.com/keiyoushi/extensions-source/issues/3566",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2551151071 | Asura Scans 1.4.40 HTTP 503.
Source information
Asura Scans 1.4.40
Source language
English
Steps to reproduce
Visit Asura Scans Extension
Expected behavior
Should Fetch manhwas and chapter lists
Actual behavior
HTTP 503
Checking the website in webview redirects asuracomic.net to asurascans.com/$1
Error 503 Backend Fetch Failed.
Mihon/Tachiyomi version
0.16.5
Android version
13
Other details
This issue started with an error 404 and then progressed onto 503.
Acknowledgements
[X] I have searched the existing issues and this is a new ticket, NOT a duplicate or related to another open or closed issue.
[X] I have written a short but informative title.
[X] I have updated the app to version 0.15.3.
[X] I have updated all installed extensions.
[X] I have tried the troubleshooting guide.
[X] If this is an issue with the app itself, I should be opening an issue in the app repository.
[X] I will fill out all of the requested information in this form.
Found existing Bug that has been reported a few minutes before.
| gharchive/issue | 2024-09-26T17:23:27 | 2025-04-01T06:44:40.715069 | {
"authors": [
"SantanuLayek"
],
"repo": "keiyoushi/extensions-source",
"url": "https://github.com/keiyoushi/extensions-source/issues/5244",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2129302506 | Make IP bind configurable
Some Home Assistant (OS) installations may have more than one interface assigned to them, e.g., when IoT devices are located in their own subnet (e.g., to prevent the smart meter from establishing a connection with some cloud services).
During the multicast listener's setup, one needs to explicitly specify the local address in that case. This change allows users with such a setup to use this addon as well.
The commit should be backwards compatible as far as I can see, but I don't consider myself an expert on Home Assistant. Feedback welcome. The value should be optional, i.e., an empty string is the default value. There might be a more elegant solution to this.
P.S.: This is basically the same change I had to use with the original SMA-EM project to receive any data from the meter.
I'm not convinced config.ini is being used by the non-legacy addon (sma-em-legacy)
It is completely removed from the dev version & at some point I'll copy the entire sma-em-dev to sma-em to sync them.
Please only make changes to the sma-em-dev folder.
@kellerza well I don't think that is the case either. Maybe as the expert you should clean up such old files from the directory? It was further not clear to me what the state of these various directories is; I modified what was used on my installation of Home Assistant.
Also, I'd suggest a more Git-ish workflow using separate branches (e.g., keeping the current sma-em on master, moving sma-em-dev to a dev branch and the -legacy to a new legacy branch. This helps reduce confusion around what files are used for what purpose and would further allow you to offer a "dev" repository for contributors to easily test changes.
I might reorganize my fork repository accordingly to test the -dev files and make changes accordingly.
Let's stick to the directories, as HASS OS expects certain files in the main branch.
See sma-em-dev is the working "branch" and it gets synced to sma-em whenever it's been stable for a while / some big change is planned for the dev addon.
Thank you for th ePR. Willing to accept it, but only to the dev/edge version of the addon.
At some point I will copy over dev/edge to the other verison
| gharchive/pull-request | 2024-02-12T01:54:11 | 2025-04-01T06:44:40.726779 | {
"authors": [
"TheAssassin",
"kellerza"
],
"repo": "kellerza/hassio-sma-em",
"url": "https://github.com/kellerza/hassio-sma-em/pull/35",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2121389738 | Standalone, where do I define the USB device?
Issue related to
Sunsynk / mbusd Home Assistant Add-On
Describe the issue/bug
I've copied over options.yaml and the error I'm getting is:
sunsynk-multi-1 | 17:57:45 INFO No response on the Modbus interface tcp://mbusd:502, see https://kellerza.github.io/sunsynk/guide/fault-finding
I haven't defined the dev/USB** device anywhere, I've just used the bundled mbus.
Do I need to define the USB device anywhere or should it 'just work'? I'm using the WAVE to USB adaptor.
Expected behavior
Connect successfullly
Your environment
Home Assistant version: 2023.8.x
Addon:
Name: sunsynk-multi
Version: 2023-08-17
Inverter:
Make: Sunsynk
Model: Ecco 3.5
Firmware:
Adaptor details:
RS485 WAVE to USB
Your configuration
---
DRIVER: umodbus
INVERTERS:
- SERIAL_NR: '007'
HA_PREFIX: SS
MODBUS_ID: 1
PORT: tcp://mbusd:502
SENSOR_DEFINITIONS: single-phase
SENSORS:
- total_pv_energy
- total_battery_charge
- total_battery_discharge
- total_grid_export
- total_grid_import
- battery_soc
SENSORS_FIRST_INVERTER: []
MANUFACTURER: Sunsynk
READ_SENSORS_BATCH_SIZE: 60
NUMBER_ENTITY_MODE: auto
MQTT_HOST: 192.168.68.53
MQTT_PORT: 1883
MQTT_USERNAME: ''
MQTT_PASSWORD: ''
DEBUG: 0
DEBUG_DEVICE: "/dev/ttyAMA0"
Logs
sunsynk-multi-1 | 17:57:44 INFO Using Single phase sensor definitions.
sunsynk-multi-1 | 17:57:44 INFO Added hidden sensors as other sensors depend on it: Rated power, Serial
sunsynk-multi-1 | 17:57:44 INFO Connecting to tcp://mbusd:502
sunsynk-multi-1 | 17:57:44 INFO Reading startup sensors Serial, Rated power
sunsynk-multi-1 | 17:57:44 ERROR gaierror reading 15 registers from 3: [Errno -2] Name does not resolve
sunsynk-multi-1 | 17:57:44 ERROR gaierror reading 15 registers from 3: [Errno -2] Name does not resolve
sunsynk-multi-1 | 17:57:44 ERROR gaierror reading 15 registers from 3: [Errno -2] Name does not resolve
sunsynk-multi-1 | 17:57:44 WARNING Retrying individual sensors: ['serial', 'rated_power']
sunsynk-multi-1 | 17:57:45 ERROR gaierror reading 5 registers from 3: [Errno -2] Name does not resolve Serial
sunsynk-multi-1 | 17:57:45 ERROR gaierror reading 2 registers from 16: [Errno -2] Name does not resolve Rated power
sunsynk-multi-1 | 17:57:45 CRITICAL Could not read sensors: ['Serial', 'Rated power']
sunsynk-multi-1 | 17:57:45 INFO ############################################################
sunsynk-multi-1 | 17:57:45 INFO No response on the Modbus interface tcp://mbusd:502, see https://kellerza.github.io/sunsynk/guide/fault-finding
sunsynk-multi-1 | 17:57:45 INFO ############################################################
sunsynk-multi-1 | 17:57:45 CRITICAL This Add-On will terminate in 30 seconds, use the Supervisor Watchdog to restart automatically.
The USB port should be configured where you have the USB device plugged in.
If you refer to mbusd, you have to either run mbusd natively or on Docker. In this case you specify the USB config in mbusd
Otherwise you need to identify, and configure the USB port in the options.yaml
Thanks for the quick reply, I'm not running any additional services for mbus, just the hass-addon-sunsynk-multi image. Should it be something like this?
DRIVER: umodbus
INVERTERS:
- PORT: serial:///dev/ttyUSB0
Yes, that looks better. I would use the pymodbus driver though, not umodbus
Thanks, do I need to run any additional services in Docker or is just the main Sunsynk one enough?
I got it working, I had to pass in the device to the docker compose file too::
services:
sunsynk-multi:
restart: unless-stopped
image: ghcr.io/kellerza/hass-addon-sunsynk-multi/aarch64:stable
devices:
- /dev/ttyUSB0:/dev/ttyUSB0
volumes:
- ./options.yaml:/data/options.yaml
---
DRIVER: pymodbus
INVERTERS:
- SERIAL_NR: '007'
HA_PREFIX: SS
MODBUS_ID: 1
PORT: /dev/ttyUSB0
I could document this and raise a PR if you want?
A PR would be welcome & help people in the future. Thanks!
PErfect, I'll do that! thanks
That fixed my issue too. I was resorting to using TCP access via the Logger, but I wanted to use a cable.
| gharchive/issue | 2024-02-06T18:02:38 | 2025-04-01T06:44:40.735086 | {
"authors": [
"championc",
"gurmukhp",
"kellerza"
],
"repo": "kellerza/sunsynk",
"url": "https://github.com/kellerza/sunsynk/issues/230",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
311689235 | Too easy to use
We need to add more dependencies because this repo is too pick up and go. I vote for redux, webpack, and to not document setup.
What's bad about being too easy to use?
@MaxxiBoi , i can't run this code, can you help me to run this code, its too complicated for me
| gharchive/issue | 2018-04-05T16:50:28 | 2025-04-01T06:44:40.750544 | {
"authors": [
"BusbyActual",
"MaxxiBoi",
"tarikhagustia"
],
"repo": "kelseyhightower/nocode",
"url": "https://github.com/kelseyhightower/nocode/issues/2243",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
314354720 | Add integration for third party apps
One should be able to add integration for third party apps such as CI/CD
I've coded a script to run with Travis, check it out:
Good Solid script.
| gharchive/issue | 2018-04-14T19:20:47 | 2025-04-01T06:44:40.751732 | {
"authors": [
"lhcgoncalves",
"mwaz"
],
"repo": "kelseyhightower/nocode",
"url": "https://github.com/kelseyhightower/nocode/issues/2270",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
298699147 | Adding kubernetes support
Added kubernetes support.
Thats all
+1
up
apiVersion: apps/v1
kind: Deployment
metadata:
name:
labels:
app:
spec:
replicas: 256
selector:
matchLabels:
app:
template:
metadata:
labels:
app:
spec:
containers:
- name:
image:
ports:
- containerPort: 80
| gharchive/pull-request | 2018-02-20T18:24:54 | 2025-04-01T06:44:40.754860 | {
"authors": [
"amirbagh75",
"ericvinicius",
"rsaraiva",
"vtrduque"
],
"repo": "kelseyhightower/nocode",
"url": "https://github.com/kelseyhightower/nocode/pull/1929",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
982824201 | page-break add blank space on top
Describe the bug
When I use <div class="html2pdf__page-break" />, the plugin add blank space beetween the label 'Test' and the upper border
There are not margin on inline style
....
</el-row>
</section>
<el-divider />
<div class="html2pdf__page-break" />
<section class="pdf-item">
<h3>Test</h3>
.....
All parts of the file must be inside a section ? I put inside only some parts of the code
Screenshots
Package Version
1.8.0
Thanks
<vue-html2pdf
ref="html2Pdf"
:margin="50"
:show-layout="true"
:float-layout="false"
:enable-download="download"
:preview-modal="false"
:paginate-elements-by-height="1400"
:filename="dynamicFileName"
:pdf-quality="2"
:manual-pagination="true"
pdf-format="a4"
pdf-orientation="portrait"
pdf-content-width="780px"
@startPagination="beginPrint()"
@beforeDownload="beforeDownload($event)"
>
Solved with
.on-top{
margin-top: -120px
}
and
<div class="html2pdf__page-break" />
<section :class="savingPdf ? 'on-top' : '' ">
<h3>Test</h3>
The way that I have fixed this issue is adding css to page-break class like this:
html2pdf__page-break {
margin: 0 !important;
}
ehhh convert image to base 64 and done:
https://www.base64-image.de/
| gharchive/issue | 2021-08-30T13:59:26 | 2025-04-01T06:44:40.762357 | {
"authors": [
"DanieleMar",
"Dzeg",
"sannchiss"
],
"repo": "kempsteven/vue-html2pdf",
"url": "https://github.com/kempsteven/vue-html2pdf/issues/104",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
418137835 | Typo fix: "Requirments" -> "Requirements"
(Seems like the GitHub editor also modified the white space on the last line.)
Thanks!
| gharchive/pull-request | 2019-03-07T05:37:48 | 2025-04-01T06:44:40.772580 | {
"authors": [
"follower",
"minux"
],
"repo": "kendryte/kflash.py",
"url": "https://github.com/kendryte/kflash.py/pull/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1975993478 | Pie Chartsページにおける現在の表示がAll ContestsなのかRated Onlyなのか分かりづらいUIになっている
#1369 にて、Pie Chartsの表示をOnly RatedとAll Contestsで切り替え可能になりましたが、現在の表示がOnly RatedなのかAll Contestsなのか一目で分からないです(UIが悪いです)。
以下の画像では、現在の表示がOnly Rated Contestsなのか、押せばOnly Rated Contestsになるのか、判断が付きません。
二択なのだから、Daily Effortのような表示が分かりやすいと思います(SimpleなのかColoredなのかが分かりやすく良いUI)。
それか、Difficulty Statusのページのように、現在進行形を使えばまだマシになるとは思います(例: Only Rated Contests -> Showing Only Rated Contests)。
#1369 がrevertされたため、こちらもclose致します。
| gharchive/issue | 2023-11-03T11:34:55 | 2025-04-01T06:44:40.831424 | {
"authors": [
"your-diary"
],
"repo": "kenkoooo/AtCoderProblems",
"url": "https://github.com/kenkoooo/AtCoderProblems/issues/1454",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
581174914 | 解ける確率を List にも入れる
https://twitter.com/medo_program05/status/1238788786210783234
Userを基準に計算した予測回答確率を元に、ユーザに対する問題の難易度アイコンを作ってみました。
(1-解答確率)を問題に対するUserの相対難易度と解釈して、3色([Easy, Moderate, Difficult], 両側1/3を境界)のバーで表現
マウスオーバーでユーザの予測回答確率が確認可能
関連Issue
#643
これいいですね! #643 との比較で言えば正解率が解釈しやすい値なのがわかりやすさに効いていると思います。
これはやや好みの話ですが、Tip で正解確率を出しているので下のバーでも統一して正解確率が高いほうが長いほうが直感に合いそうです。あと、一本のバーの中で複数の色があると混乱する(私はしました)ので、色を長さで変えるのはわかりやすくてよいのですが1本のバーの中では一つの色を使うようにしたほうがよくなると思いました。
あと、一本のバーの中で複数の色があると混乱する(私はしました)ので、色を長さで変えるのはわかりやすくてよいのですが1本のバーの中では一つの色を使うようにしたほうがよくなると思いました。
難易度ごとに色を統一し,長さはそれぞれの区間内での位置に合わせました.(難易度アイコンと同じ感じ)
難易度感が直観的にわかりやすくなったと思います.
これはやや好みの話ですが、Tip で正解確率を出しているので下のバーでも統一して正解確率が高いほうが長いほうが直感に合いそうです。
解答確率をそのまま表示するということですね.
私が相対難易度(1-解答確率)で表示した理由は,解答確率を元にアイコン表示すると,簡単な(解答確率の高い)問題が目立ち,難しい(解答確率の低い)問題が目立たなくなってしまうのはユーザーにとってうれしくないのではないかと考えたからです.
難易度アイコンも難しい問題ほど面積が大きくなっていますので,それとも直感を揃えたいと思いました.
議論のポイントは
ユーザーが最も欲しい情報は何か
欲しい情報に適したデザインは何か
難易度区分は調整したほうがいいかも。
現在: 上下33%
提案: 上下25%
区間は上記に変更しました.
とりあえず3パターン試し
1
アイコニックなバー.
個性強め.
細かい改善で良くなる可能性もある.
2
セルのボトムに合わせたバー.
主張は強いが直観的にわかりやすい.
しいて言うならよくありそうなデザイン.
3
セルの縦に合わせたバージョン.
こちらも直観的に認識しやすいが,横とくっつくので見づらい人もいるかもしれない.
パターン2は確かに読みやすいですね。
新たに表示する機能を追加するのではなくて #401 + List ページのテーブルに確率カラムを追加 ではダメですかね?
discrimination が小さいやつは超低難度か超高難度の問題がほとんどであまりあてになりませんし、difficulty+200程度のレーティングを持つ人なら 60-70% の確率、 difficulty+400 程度なら 70-90%、+800 なら 90% 以上の確率で解ける (と予測される) 問題がほとんどです。 difficulty 自体厳密な難易度判定ではないですし、自分の色から1-2色下の問題を意識して解くようにすればそれでいいような気がします。
discrimination のヒストグラムと delta=(Rating - Difficulty) のときの予測正解率 - 0.5 の雰囲気
個人的にはダークモードのパターン3なんかとてもカッコいいと思うんですが、 予測正解率は difficulty との相関が強いので多分玉の方だけで機能としては足りていて、視覚に強く訴える割には冗長な情報かと思います。
Recommends 以外の場所でも目安として知りたい気持ちは分かるので、上述の #401 + List ページのテーブルに確率カラムを追加 は欲しいと思います。
#401 + List ページのテーブルに確率カラムを追加
は良さそうですね。
すでにRecommendationで実装されている情報ですから、Tableで弊害がないなら表示してもいいと思います。
予測正解率の情報が冗長かどうかは主観にもよる部分があると思います。
確かに予測正解率はdifficultyと相関がありますが、discriminationによる補正は直感に対し無視できない影響があると私は考えています。
ToolTipによる予測正解率表示は問題ごとの情報の提示としての要件は満たしますが、一覧性がありません。
それに対し、提案した3色表示は概ね実力相当の問題がどれであるかを視認する一覧性に長けているので、意味がある表示だと私は思います。
個人的にはレコメンドと得られる情報が変わらない気もするので必要性を感じていませんが、これは個人の感想で、欲しい人がいれば入れた方が良いと思います。どんなユーザーを想定していますか?
まず、難易度メータというアイデアは、「推定正解率を表示する機能」のリッチな拡張であることは明らかにしておきます。
メインの目的は、単純に問題の難しさを知りたいということです。
AtCoder Problems で今日のコンテストの問題のdiffを確認する楽しみ方をしているユーザも多いと思いますが、そういった楽しみ方の一つになるのではないかと思います。
「問題に対して難易度を提供する」という機能は一般的に理解しやすく需要の高い情報だと思いますが、このアイデアはよりユーザに最適化した情報を提供するという立場になります。
またこの情報は、アイコニックな表示で直感的に理解できることや、一覧性によって他の問題との比較ができることでより価値が上がるものだと考えています。
ときに、「思ってたよりもdiffが低い(高い)」という印象の問題はしばしば判別度(discrimination)が低いという状況もありますが、difficultyだけでは得られない情報で補足されるのは私は嬉しいです。(#643)
これはデザイン依存の楽しみ方になるかもしれませんが、例えば
メーターのModerate(黄色)相当は確実に解ける用になっておきたい
レートが上がったら前は難しかった問題がModerateになって嬉しい
といったことも想定できそうです。
これに関しては、現状より色分けを工夫してユーザーが注目しやすい問題に適した表示に改良できれば良いなと考えています。
Table のページは既に色で表現している情報が多量にあって、ここに更に3色を追加するとなると視認性が悪化するような気がします。 (要するにごちゃごちゃしちゃう)
私は https://github.com/kenkoooo/AtCoderProblems/issues/469#issuecomment-659079129 で述べたように、ほとんどの問題に関しては discrimination はある程度の範囲に収まっているので難易度の表現としては difficulty の値だけで事足りていると思っていて、少数の問題に対して difficulty + α の情報を視覚的に提供するためだけに全体の視認性を阻害するのはUX的に良くない、というスタンスです。 (というか、問題の"真の難易度"を算出すること自体が現実的に不可能なのでこの +α の情報にもそんなに意味はないと思っています)
difficulty による色表示を予測正答率で代替すれば表示上はスッキリするかなと思いましたが、当人のレーティング周辺の difficulty の問題以外に対しては情報量ゼロになってしまって、結局劣化 difficulty にしかならなそうです。 何より現在ユーザー間では "赤diff" "青diff" のような言い回しでコンセンサスが取れているので個人依存の指標に基づく色分けは混乱の元になりそうです。
UIがシンプルな方がいいのは同意します。
仮に実装するとしてもできるだけシンプルでわかりやすい物がいいですし、改善はしていくべきです。
それははそれとして、
必要な表示を選択する自由は"チェックボックスによる切り替え機能"で担保されているということにはならないでしょうか。
難易度メータは独立したアイデアとして別にIssue切っても良さそうな気がしてきました。
そうですね、はじめは optional な機能として実装してみてユーザーの反応をうかがってみるのも良いかもしれません。 デザインの改善案なんかもユーザーから出てきそうです。
まだマージされていないのでどうなるか分かりませんが、 #691 で <DifficultyCircle>, <ProblemLink> に ProblemModel, RatingInfo プロパティを追加しました。 予測正答確率の計算がこれらのコンポーネント上でできるようになるので、コンポーネントの一部としてメーターが実装されると Table 以外のページでも表示できて嬉しいかもしれません。 デザイン上の制限は生まれますが…。
ありがとうございます。
承知しました。
競合しそうなので対応しておきます。
メーターは独立したコンポーネントにしておいたほうが表示非表示の切り替えが楽でいいかなと思ってます。
このIssueの本来の目的は #401 に近いものであり、 #691 で達成されそうなので、
難易度メーターは別Issue #693 にしました
done in #691
| gharchive/issue | 2020-03-14T11:56:16 | 2025-04-01T06:44:40.851922 | {
"authors": [
"MatsuTaku",
"amylase",
"kenkoooo",
"koyumeishi",
"southball"
],
"repo": "kenkoooo/AtCoderProblems",
"url": "https://github.com/kenkoooo/AtCoderProblems/issues/469",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
89582912 | Curry methods by default
To avoid needing separate bind or partial calls (for methods that take more than 1 argument)
var isTitle = check.oneOf(['Mr', 'Dr, 'Ms', 'Mss']);
la(isTitle(foo), 'invalid title', foo);
Example already curried https://github.com/kensho/check-more-types#checkequal - not sure it is implemented well
| gharchive/issue | 2015-06-19T14:51:39 | 2025-04-01T06:44:40.865183 | {
"authors": [
"bahmutov"
],
"repo": "kensho/check-more-types",
"url": "https://github.com/kensho/check-more-types/issues/34",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
374462179 | fix typos in custom renderer example
This PR fixes the mistake in the custom renderer example as discussed on https://github.com/kentcdodds/react-testing-library/issues/204#issuecomment-433468647
:tada: This PR is included in version 5.3.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/pull-request | 2018-10-26T16:41:23 | 2025-04-01T06:44:40.868132 | {
"authors": [
"ankitsinghaniyaz",
"kentcdodds"
],
"repo": "kentcdodds/react-testing-library",
"url": "https://github.com/kentcdodds/react-testing-library/pull/212",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
657035224 | docs(readme): Delete reference to doclets.io
fixes #24
Remove invalid doclets.io url from API documentation
@kentcdodds not sure why this is failing the CI as only changes made were to text files
| gharchive/pull-request | 2020-07-15T03:37:41 | 2025-04-01T06:44:40.869690 | {
"authors": [
"gitKendra"
],
"repo": "kentcdodds/webpack-config-utils",
"url": "https://github.com/kentcdodds/webpack-config-utils/pull/33",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
69871369 | This needs a few tests
I try running the tests and:
[08:05:30] Using gulpfile ~/tmp/nuka-carousel/Gulpfile.js
[08:05:30] Starting 'lint'...
[08:05:30] Finished 'lint' after 506 ms
[08:05:30] Starting 'karma'...
ERROR [config]: File /Users/jherri1/tmp/nuka-carousel/karma.conf.js does not exist!```
Tests are in!
| gharchive/issue | 2015-04-21T15:06:31 | 2025-04-01T06:44:40.870633 | {
"authors": [
"jherr",
"kenwheeler"
],
"repo": "kenwheeler/nuka-carousel",
"url": "https://github.com/kenwheeler/nuka-carousel/issues/1",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
103001398 | Gallery dragging/sliding problems with allot optenet security proxy
Hi,
Vodafone Germany (along other opcos world wide) uses Allot's Optenet Service (http://www.allot.com/allot-optenet-security-solutions/) for safe browsing. It's basically a big packet inspecting proxy which scans your traffic for malware and injects javascript into web pages for users to feel secure.
I made a screencast to illustrate the problem:
https://www.youtube.com/watch?v=fqUOXN2G9hw
On top you see the injected "you are secure" banderole with its tiny shield icon. This icon is draggable in case of the user wants to click anywhere below it on page. When I drag the shield, all carousels on page are dragged as well. When I release the shield, gallerys stay attached to the mouse movements.
This happens on every page using slick carousel I visited.
Whoah thats nuts. What exactly does that shield do?
It sits there and gives you the warm feeling of safety.
You can drag it left and right in case it overlaps anything of importance on the page so that you can move it away and then click on whatever was beneath. Btw, that is why I'm investigating. That shield overlays the menu hamburger button on our page and when the user drags it away, all carousels move as well...
Additionally the shield catches onclick() events, on which a bigger banderole unrolls displaying this sites security status (you can see it in my video, sec 27). Everything but the dragging works fine.
I'm not sure this is a slick issue. The drag on that banderole thing shouldn't be so intrusive.
| gharchive/issue | 2015-08-25T11:18:45 | 2025-04-01T06:44:40.874104 | {
"authors": [
"FieserKiller",
"kenwheeler"
],
"repo": "kenwheeler/slick",
"url": "https://github.com/kenwheeler/slick/issues/1667",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
185250481 | Can't center items horizontally when width is bigger than container
I tried using flex in each item (wrapping them by another div), centering using align-content: center, but the next slide get's in the current one.
Has anyone succeeded doing this? I tried with flex, but I it breaks.
Steps to reproduce the problem
create a carousel with images with width bigger than container
center images using flex, applying justify-content: center to every parent of every image
Chrome v.54 Ubuntu 64 bit
Slick 1.6, jQuery v1.11.2
Did this work before? No
Problem resolution:
http://jsfiddle.net/tomyo/fmo50w7n/678/
Glad you got it resolved!
Sorry, I meant, 'Problem reproduction', haha
| gharchive/issue | 2016-10-25T22:26:15 | 2025-04-01T06:44:40.877257 | {
"authors": [
"leggomuhgreggo",
"tomyo"
],
"repo": "kenwheeler/slick",
"url": "https://github.com/kenwheeler/slick/issues/2587",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
218082321 | Take time to load first time
Hello Team,
I am using slick slider for banner display in my home page so i have one issues like
when my page is load then content of slider section show little bit slow like u know
first time page is open then u show that nothings in this area after some time u saw that banner in this section
see screen shot of first load page
After few time image
I don't feel that this is quite a slick issue, but something that could be solved via CSS. Have you examined any ways to optimize asset loading? Also, you could define a pre-set height for a container element for your banner assets, and have a .loading class indicating to the user that the banner is still loading/rendering. That way the page doesn't jump when new content is rendered.
i have define pre set height of container and also using loader after that i am getting this issues
basically i am using center mode also
in head?
| gharchive/issue | 2017-03-30T05:07:03 | 2025-04-01T06:44:40.880974 | {
"authors": [
"gavinlynch",
"lasbrojen",
"rahi-rajeev"
],
"repo": "kenwheeler/slick",
"url": "https://github.com/kenwheeler/slick/issues/2823",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
200712090 | Update README.md
resolves #26
updated, and I'll update in generator-keystone as well
| gharchive/pull-request | 2017-01-13T19:36:12 | 2025-04-01T06:44:40.882063 | {
"authors": [
"thescientist13"
],
"repo": "kenzanlabs/keystone",
"url": "https://github.com/kenzanlabs/keystone/pull/37",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
798614539 | required time for training on kinetics
hi
How long does it take to train model on kinetics dataset for a few examples of the different hardware you tested before?
Hi @parvanehabi, thanks for your interest! If I recall correctly it takes a while to train on Kinetics. I was sharing the training log in another issue earlier, see here. Seems like it takes 4 days to train on the whole dataset. I don't quite remember what hardware it was using, probably 1080Ti's. Hope this helps!
| gharchive/issue | 2021-02-01T18:45:25 | 2025-04-01T06:44:40.883606 | {
"authors": [
"kenziyuliu",
"parvanehabi"
],
"repo": "kenziyuliu/MS-G3D",
"url": "https://github.com/kenziyuliu/MS-G3D/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1584349665 | update to mermaid v9.x to get rid of some vulnerabilities
mermaid v9.1+ fixes some vulnerabilities that we see with the v8 line:
https://access.redhat.com/security/cve/CVE-2021-23648
https://github.com/mermaid-js/mermaid/security/advisories/GHSA-x3vm-38hw-55wf
Do you have any plans to update to v9?
I'll do that now. Sorry I'm falling behind for the OS community.
@hakandilek it looks like v9 broke something with many to many relations. I'll need to figure out what that is before merging. I don't have time to look into it at this moment though.
You can install v1.3.0 and it is updated with mermaid v10.
npm i -D prisma-erd-generator@alpha or npm i -D prisma-erd-generator@1.3.0
Great, thanks 😊
| gharchive/issue | 2023-02-14T15:17:32 | 2025-04-01T06:44:40.886753 | {
"authors": [
"hakandilek",
"keonik"
],
"repo": "keonik/prisma-erd-generator",
"url": "https://github.com/keonik/prisma-erd-generator/issues/173",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1381879545 | Problem with minimal theme and excalidraw plugin
Is the bug present when using the default Obsidian theme?
No
Is the bug present when snippets and plugins are disabled?
Yes
Minimal theme version
6.0.5
Describe the bug
Then I turn on dark mode in the obsidian and Excalidraw is in light mode, I get bad colours
Excalidraw author redirect me to you about this issue.
more details in a ticket:
https://github.com/zsviczian/obsidian-excalidraw-plugin/issues/817
Debug info
no one
Also experiencing this issue.
Also seeing this. Does anyone know of a workaround for this issue?
Here is a CSS snippet workaround to make it usable. Depending on your theme it may not look very pretty.
body .excalidraw, body .excalidraw.theme--dark { --island-bg-color: #F5F5F5; }
I am also having this issue - just want to clarify the issue.
It happens when I am using dark theme and I switch to excalidraw light mode (the dark mode = black paper background, works fine)
Issue is on mac OS
Obsidian v1.0
Here is a CSS snippet workaround to make it usable. Depending on your theme it may not look very pretty.
body .excalidraw, body .excalidraw.theme--dark { --island-bg-color: #F5F5F5; }
I had to change it to
body .excalidraw { --island-bg-color: #F5F5F5; }
to make the switching work properly
Fixed in 6.1.10
| gharchive/issue | 2022-09-22T06:01:20 | 2025-04-01T06:44:40.892420 | {
"authors": [
"The-Grand-Vizier",
"axelson",
"dariuszkowalski-com",
"insecurejezza",
"janLo",
"kepano"
],
"repo": "kepano/obsidian-minimal",
"url": "https://github.com/kepano/obsidian-minimal/issues/427",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1181916044 | [feature]: Add Datagrid to list all the test runs (UI improvement)
Is there an existing feature request for this?
[X] I have searched the existing issues
Summary
Adding Datagrid instead of cards on the "/testruns" page makes it easier to sort out test runs according to users need
Why should this be worked on?
Makes it easier to sort out required test run data based on their need.
A better UX experience for the user.
This is a duplicate issue https://github.com/keploy/ui/issues/4
| gharchive/issue | 2022-03-26T17:13:48 | 2025-04-01T06:44:40.901399 | {
"authors": [
"namantaneja167",
"nehagup"
],
"repo": "keploy/ui",
"url": "https://github.com/keploy/ui/issues/36",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2530868529 | fix: Fixed Recording of test cases
This pr fixes the behavior of recording when terminal is closed.
After recording test cases in zsh shell
Great work! Works great with replay, please check on the record side
also add semicolons in your files, will cause havoc later if not present
This is making, default running shell as bash, this isnt ideal,please look into the same
if any terminal is open it is also closed please try to fix it
Well done @ayush3160
| gharchive/pull-request | 2024-09-17T11:14:03 | 2025-04-01T06:44:40.904072 | {
"authors": [
"ayush3160",
"khareyash05",
"shivamsouravjha"
],
"repo": "keploy/vscode-extension",
"url": "https://github.com/keploy/vscode-extension/pull/50",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
722043023 | Misleading/incorrect information for quality gate only projects
Hi
I am testing the latest Keptn bridge 0.7.2 and have one project that I only use for Test Execution & Quality Gates - so - I am not using it for delivery.
As you can see from the screenshot the information is rather misleading and actually incorrect as this project has and will never see a deployment - yet - the message tells me that I should do a deployment. We should fix this by detecting whether a project is only used for e.g: quality gates or testing as a service and then show relevant information for those use cases
Here is what I see - it shows me no deployment and a final score of 0
And here is the view in my services view where I can see my latest test executions and quality gates
The initial event that I am sending is actually a deployment event as you can see here:
And here is the event content:
{
"contenttype": "application/json",
"data": {
"deploymentURIPublic": "http://simplenode.keptn07project-staging.keptn07-agrabner.demo.keptn.sh/",
"eventContext": null,
"image": "Keptn Performance as a Service",
"labels": {
"buildId": "47",
"jobname": "Keptn Performance as a Service",
"joburl": "https://angr-jenkins.apps.lab.dynatrace.org/job/Keptn%20Performance%20as%20a%20Service/47/"
},
"project": "perfaasproject",
"service": "perfaasservice",
"stage": "performance",
"tag": "47",
"teststrategy": "performance"
},
"id": "79000c5f-575f-4537-b0ec-2d32598fbe05",
"source": "jenkins-library",
"specversion": "0.2",
"time": "2020-10-14T14:18:57.857Z",
"type": "sh.keptn.events.deployment-finished",
"shkeptncontext": "4d90f1f3-284e-4b18-a6d7-9437f93a83ce"
}
Would be great if we could improve the environment screen
This issue is partly implemented by 0.8.x, but a final adjustment will happen with: https://github.com/keptn/keptn/issues/4172
| gharchive/issue | 2020-10-15T06:34:31 | 2025-04-01T06:44:40.909334 | {
"authors": [
"grabnerandi",
"johannes-b"
],
"repo": "keptn/keptn",
"url": "https://github.com/keptn/keptn/issues/2544",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
818751657 | helm-service: Delivery failed with "Error when installing/upgrading chart" ... "has no deployed releases"
Environment
Client OS (e.g., Linux, OSX, Windows): Linux
Keptn Version (keptn version): 0.8.0-rc1
Kubernetes Cloud Provider (e.g., GKE, AKS): GKE 1.18
Kubernetes version (kubectl version): 1.18
Affected Component
[ ] Docs
[ ] CLI
[ ] Kubernetes Integration
[ ] Openshift Integration
[x] Helm
[ ] Istio
[ ] Bridge
[ ] Approval
[ ] Datastore
[ ] REST API
[ ] eventbroker / distributor
[ ] jmeter
Describe the bug
When deploying to hardening (staging) with blue-green, the first deployment step in the shipyard fails with
Error when installing/upgrading chart tempberry-hardening-tempberry-backend-generated in namespace tempberry-hardening: "tempberry-hardening-tempberry-backend-generated" has no deployed releases
To Reproduce
Create a project using https://github.com/keptn/examples/blob/release-0.8.0/onboarding-carts/shipyard.yaml
Onboard services and trigger delivery
During my first deployment in that stage something went wrong with the Kubernetes API being down (unrelated to Keptn). Any deployments after that don't work anymore. I'm not sure if that is something we can reproduce, but my guess is that the helm chart tempberry-hardening-tempberry-backend-generated does not exist, and helm-service is currently not able to handle that case.
Expected behavior
IMHO helm-service needs to be able to recover from such an error somehow.
Current behavior
Delivery with helm-service fails.
Additional context
$ helm ls -n tempberry-hardening
NAME NAMESPACE REVISION UPDATED STATUS CHART APP VERSION
tempberry-hardening-postgres tempberry-hardening 2 2021-03-01 10:08:56.744060584 +0000 UTC deployed postgres-0.1.0
tempberry-hardening-postgres-generated tempberry-hardening 2 2021-03-01 10:09:01.997584865 +0000 UTC deployed postgres-generated-0.1.0
tempberry-hardening-tempberry-backend tempberry-hardening 3 2021-03-01 10:50:44.882586921 +0000 UTC deployed tempberry-backend-0.1.0
$ kubectl -n tempberry-hardening get deployments -owide
NAME READY UP-TO-DATE AVAILABLE AGE CONTAINERS IMAGES SELECTOR
postgres 1/1 1 1 60m postgres postgres:10.4 app=postgres
tempberry-backend 1/1 1 1 49m tempberry-backend ckreuzberger/tempberry-backend:0.1-master app=tempberry-backend
tempberry-backend-primary 1/1 1 1 48m tempberry-backend ckreuzberger/tempberry-backend:0.1 app=tempberry-backend-primary
Helm-Service Logs
{"timestamp":"2021-03-01T10:50:29.765242963Z","logLevel":"INFO","message":"Finished release for service tempberry-backend in stage dev and project tempberry"}
{"timestamp":"2021-03-01T10:50:34.967065024Z","logLevel":"DEBUG","message":"Got event of type sh.keptn.event.deployment.triggered"}
{"timestamp":"2021-03-01T10:50:34.969458043Z","logLevel":"INFO","message":"Starting deployment for service tempberry-backend in stage hardening of project tempberry"}
{"timestamp":"2021-03-01T10:50:34.972152075Z","logLevel":"INFO","message":"Updating values for service tempberry-backend in stage hardening of project tempberry"}
{"timestamp":"2021-03-01T10:50:44.592129Z","logLevel":"INFO","message":"Creating namespace tempberry-hardening if not present"}
{"timestamp":"2021-03-01T10:50:44.598135022Z","logLevel":"DEBUG","message":"Reuse existing namespace tempberry-hardening"}
{"timestamp":"2021-03-01T10:50:44.59860691Z","logLevel":"INFO","message":"Start upgrading release tempberry-hardening-tempberry-backend in namespace tempberry-hardening"}
{"timestamp":"2021-03-01T10:51:05.14470142Z","logLevel":"DEBUG","message":"---\n# Source: tempberry-backend/templates/tempberry-backend-service.yaml\n# [START service]\napiVersion: v1\nkind: Service\n..."}
{"timestamp":"2021-03-01T10:51:05.147570396Z","logLevel":"INFO","message":"Finished upgrading chart tempberry-hardening-tempberry-backend in namespace tempberry-hardening"}
{"timestamp":"2021-03-01T10:51:11.128708444Z","logLevel":"INFO","message":"Creating namespace tempberry-hardening if not present"}
{"timestamp":"2021-03-01T10:51:11.133444573Z","logLevel":"DEBUG","message":"Reuse existing namespace tempberry-hardening"}
{"timestamp":"2021-03-01T10:51:11.133476194Z","logLevel":"INFO","message":"Start upgrading release tempberry-hardening-tempberry-backend-generated in namespace tempberry-hardening"}
{"timestamp":"2021-03-01T10:51:11.147670555Z","logLevel":"ERROR","message":"Error when installing/upgrading chart tempberry-hardening-tempberry-backend-generated in namespace tempberry-hardening: \"tempberry-hardening-tempberry-backend-generated\" has no deployed releases"}
Workaround
I worked around this problem by deleting the service and onboarding it again
keptn delete service ...
keptn onboard service ...
keptn trigger delivery ...
It seems this issue has been fixed in the meantime - potentially by an update of the helm.sh/helm/v3 dependency, which was set to v3.1.2 in Keptn 0.8.0, but has been upgraded to v3.5.1 in the meantime.
With v0.8.0 of the helm service, I was able to reproduce the problem as follows, using the carts-db service of our examples repo:
# Onboard the service
keptn onboard service carts-db --project=sockshop --chart=./carts-db
# Trigger a delivery with a tag that does not exist:
keptn trigger delivery --project=sockshop --service=carts-db --image=mongo --tag=foo --sequence=delivery-direct
This will eventually lead to a time out in the helm service when waiting for the deployment to be ready. After that, the release is marked as failed when retrieving the releases using the helm CLI:
$ helm ls -n sockshop-dev -a
NAME NAMESPACE REVISION UPDATED STATUS CHART APP VERSION
sockshop-dev-carts sockshop-dev 33 2021-03-26 11:17:53.085282944 +0000 UTC deployed carts-0.1.0
sockshop-dev-carts-db sockshop-dev 2 2021-03-29 14:16:09.052321541 +0000 UTC failed carts-db-0.1.0
sockshop-dev-carts-generated sockshop-dev 30 2021-03-26 11:19:03.923766366 +0000 UTC deployed carts-generated-0.1.0
After that, I tried to trigger a delivery again:
keptn trigger delivery --project=sockshop --service=carts-db --image=mongo --tag=foo --sequence=delivery-direct
And got the following log output in the helm-service:
{"timestamp":"2021-03-30T06:41:24.87039808Z","logLevel":"ERROR","message":"Error when installing/upgrading chart sockshop-dev-carts-db in namespace sockshop-dev: \"sockshop-dev-carts-db\" has no deployed releases"}
I also tried this with the helm service built in PR #3466 but got the same result.
After upgrading the helm-service to 0.8.1 however, the upgrade was performed, even though the release was still in the failed state:
Trigger delivery:
keptn trigger delivery --project=sockshop --service=carts-db --image=mongo --tag=foo --sequence=delivery-direct
Helm-service logs:
{"timestamp":"2021-03-30T06:43:40.510320184Z","logLevel":"INFO","message":"Start upgrading release sockshop-dev-carts-db in namespace sockshop-dev"}
{"timestamp":"2021-03-30T06:46:41.351515486Z","logLevel":"ERROR","message":"Error when installing/upgrading chart sockshop-dev-carts-db in namespace sockshop-dev: timed out waiting for the condition"}
Eventuall, after providing the correct image/tag to the delivery sequence, the carts-db service could be deployed successfully (without deleting the service from Keptn first)
keptn trigger delivery --project=sockshop --service=carts-db --image=docker.io/mongo --tag=latest --sequence=delivery-direct
| gharchive/issue | 2021-03-01T11:03:29 | 2025-04-01T06:44:40.923855 | {
"authors": [
"bacherfl",
"christian-kreuzberger-dtx"
],
"repo": "keptn/keptn",
"url": "https://github.com/keptn/keptn/issues/3407",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
606142564 | Include versioning information in tutorials
Include some kind of versioning into the tutorials. Right now they are not targeting any particular keptn version.
fixed with #13
| gharchive/issue | 2020-04-24T08:37:03 | 2025-04-01T06:44:40.925490 | {
"authors": [
"jetzlstorfer"
],
"repo": "keptn/tutorials",
"url": "https://github.com/keptn/tutorials/issues/12",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1118277190 | Update TF Decision Forests example
Hello @fchollet and @achoum -
Please review the updates I made to the TF Decision Forests example.
Cheers
@achoum @ksalama are we good here?
@achoum - Thank you so much for your comments. I addressed then as follows:
I used CATEGORICAL_FEATURE_NAMES instead of CATEGORICAL_FEATURES_WITH_VOCABULARY
Create the feature semantics based on NUMERICAL_FEATURE_NAMES and CATEGORICAL_FEATURES_WITH_VOCABULARY, rather than the model input data types.
Deferred computing the vocabulary of the categorical features to the create_embedding_encoder() method.
Fixed the casting to float32 or float64 as we discussed
Updated the BinaryTargetEncoding to pre-compute all the required stats in the adapt() method, while the call() method would only do the lookup of the stats
Allow the user to provide vocabulary_size to avoid calling tf.unique(). However, if it is not passed, it's computed during the execution of the adapt() method. Also added a comment to recommend passing the vocabulary_size.
I kept the prepare_dataframe() method as it seems that tfdf.keras.pd_dataframe_to_tf_dataset doesn't do all the "magic" of converting the target column to integer etc. Am I missing something?
Also, in the last model, when I perform activation directly on the feature, then linear projection, the results are worse than using Dense(units, activation)(encoded_features). So I kept it as it was.
I haven't use tf.data.experimental.group_by_reducer, but I am planning to investigate it (in a subsequent PR).
Also, didn't use a dictionary for the parameters of the model as it, for some reasons, makes the example more than 300 lines, which is not allowed for the Keras examples!
Let me know if you have any further comments.
All good :) Thanks for the work. This is awesome.
| gharchive/pull-request | 2022-01-29T17:12:36 | 2025-04-01T06:44:40.947587 | {
"authors": [
"achoum",
"fchollet",
"ksalama"
],
"repo": "keras-team/keras-io",
"url": "https://github.com/keras-team/keras-io/pull/791",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
383361174 | How do keras calulate the metric?
I'm trying to write a new metric of f1-macro for the softmax, I have tried two methods, I can't understand why the result of them is different.
The first method, I defined a f1-macro function like that,
def f1_macro(y_true,y_pred):
y_pred_tmp=K.argmax(y_pred,axis=-1)
y_pred_tmp=K.one_hot(y_pred_tmp,4) (MULTI CLASS OF 4)
tp=y_pred_tmp*y_true
tp=tf.reduce_sum(tp,axis=0)
fp=K.greater(y_pred_tmp,y_true)
fp=tf.reduce_sum(tf.cast(fp,tf.float32),axis=0)
fn = K.greater(y_true,y_pred_tmp)
fn = tf.reduce_sum(tf.cast(fn, tf.float32), axis=0)
prec_list=[]
recal_list=[]
f1_list=[]
for i in range(0,4):
prec=tp[i]/(tp[i]+fp[i]+K.epsilon())
rec=tp[i]/(tp[i]+fn[i]+K.epsilon())
f1=2*prec*rec/(prec+rec+K.epsilon())
prec_list.append(prec)
recal_list.append(rec)
f1_list.append(f1)
return tf.reduce_mean(f1_list)
then use it like that model.compile(..., metric=[f1-macro])
The second method, I used the Callback to calculate the f1-macro by sklearn f1-score, the f1 calculated by f1_macro function is always less than the Callback. the Callback function is defined below:
class Metrics(Callback):
def on_train_begin(self, logs={}):
self.val_f1s = []
self.val_recalls = []
self.val_precisions = []
def on_epoch_end(self, epoch, logs={}):
pred=np.asarray(self.model.predict(self.validation_data[0]))
val_predict = np.argmax(pred, axis=-1)
val_targ = np.argmax(self.validation_data[1], axis=-1)
_val_recall = recall_score(val_targ, val_predict, average='macro')
_val_precision = precision_score(val_targ, val_predict, average='macro')
_val_f1 = f1_score(val_targ, val_predict, average='macro')
self.val_recalls.append(_val_recall, )
self.val_precisions.append(_val_precision)
self.val_f1s.append(_val_f1)
print('— val_f1: %f — val_precision: %f — val_recall %f' % (_val_f1, _val_precision, _val_recall))
return
and I use it by
metrics = Metrics()
model.fit(X_train, y_train,
epochs=10,
batch_size=batch_size,
validation_data=(X_valid, y_valid),
callbacks=[early_stopping, plateau, checkpoint, metrics],
verbose=2
)
In the two methods, is the validation data for keras to calculate (X_valid, y_valid)?
Is the calculation principle of sklearn f1 score the same with f1_macro ?
How do keras calculate the metrics, is it after each batch size, or after each epoch?
I'm trying to write a new metric of f1-macro for the softmax, I have tried two methods, I can't understand why the result of them is different.
The first method, I defined a f1-macro function like that,
def f1_macro(y_true,y_pred):
y_pred_tmp=K.argmax(y_pred,axis=-1)
y_pred_tmp=K.one_hot(y_pred_tmp,4) (MULTI CLASS OF 4)
tp=y_pred_tmp*y_true
tp=tf.reduce_sum(tp,axis=0)
fp=K.greater(y_pred_tmp,y_true)
fp=tf.reduce_sum(tf.cast(fp,tf.float32),axis=0)
fn = K.greater(y_true,y_pred_tmp)
fn = tf.reduce_sum(tf.cast(fn, tf.float32), axis=0)
prec_list=[]
recal_list=[]
f1_list=[]
for i in range(0,4):
prec=tp[i]/(tp[i]+fp[i]+K.epsilon())
rec=tp[i]/(tp[i]+fn[i]+K.epsilon())
f1=2*prec*rec/(prec+rec+K.epsilon())
prec_list.append(prec)
recal_list.append(rec)
f1_list.append(f1)
return tf.reduce_mean(f1_list)
then use it like that model.compile(..., metric=[f1-macro])
The second method, I used the Callback to calculate the f1-macro by sklearn f1-score, the f1 calculated by f1_macro function is always less than the Callback. the Callback function is defined below:
class Metrics(Callback):
def on_train_begin(self, logs={}):
self.val_f1s = []
self.val_recalls = []
self.val_precisions = []
def on_epoch_end(self, epoch, logs={}):
pred=np.asarray(self.model.predict(self.validation_data[0]))
val_predict = np.argmax(pred, axis=-1)
val_targ = np.argmax(self.validation_data[1], axis=-1)
_val_recall = recall_score(val_targ, val_predict, average='macro')
_val_precision = precision_score(val_targ, val_predict, average='macro')
_val_f1 = f1_score(val_targ, val_predict, average='macro')
self.val_recalls.append(_val_recall, )
self.val_precisions.append(_val_precision)
self.val_f1s.append(_val_f1)
print('— val_f1: %f — val_precision: %f — val_recall %f' % (_val_f1, _val_precision, _val_recall))
return
and I use it by
metrics = Metrics()
model.fit(X_train, y_train,
epochs=10,
batch_size=batch_size,
validation_data=(X_valid, y_valid),
callbacks=[early_stopping, plateau, checkpoint, metrics],
verbose=2
)
In the two methods, is the validation data for keras to calculate (X_valid, y_valid)?
Is the calculation principle of sklearn f1 score the same with f1_macro ?
How do keras calculate the metrics, is it after each batch size, or after each epoch?
Closing this issue since its been addressed. Feel free to reopen if have any further questions. Thanks!
| gharchive/issue | 2018-11-22T02:48:36 | 2025-04-01T06:44:40.956206 | {
"authors": [
"Suncicie",
"ymodak"
],
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/issues/11705",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1331980450 | Returning model with best weights when EarlyStopping is not triggered and restore_best_weights=True
TensorFlow version (you are using):
TF 2.5.0
Are you willing to contribute it (Yes) :
Describe the feature and the current behavior/state.
At this moment, EarlyStopping will only return the best performing model when EarlyStopping occurs due to no increase in performance over "patience" amount of epochs. Looking at the case where the complete training process is completed without invoking early stopping (e.g. patience is set way to high), most people would like to get the best performing model as a returned model. As the EarlyStopping callback tracks this best weights, it would be almost trivial to integrate this functionality.
Describe the feature clearly here. Be sure to convey here why the requested feature is needed. Any brief description about the use-case would help.
Updating the returned model when complete training occurs in the EarlyStopping callback. In this instant this is possible afterwards by using following code:
ES = EarlyStopping(patience=20, restore_best_weights=True)
callbacks = [ES]
model.fit(x=train_input, y=train_output, epochs=400, batch_size=64, callbacks=callbacks, validation_split=0.25)
if ES.stopped_epoch == 0: # ES did not halt training and thus not update model to best one seen
model.set_weights(ES.best_weights) # does however save the best weights under ES.best_weights
print(f"Restoring model weights from the end of the best epoch: {ES.best_epoch + 1}.")
Will this change the current api? How?
No undesired behaviour should occur
Who will benefit from this feature?
Will return desired behaviour for most people using EarlyStopping
Contributing
Do you want to contribute a PR? (No):
If yes, please read this page for instructions
Briefly describe your candidate solution(if contributing):
Update the on train end function as follows:
def on_train_end(self, logs=None):
if self.stopped_epoch > 0 and self.verbose > 0:
io_utils.print_msg( f"Epoch {self.stopped_epoch + 1}: early stopping")
else:
self.model.set_weights(self.best_weights)
io_utils.print_msg(
"Restoring model weights from "
"the end of the best epoch: "
f"{self.best_epoch + 1}."
)
This functionality is available in the ModelCheckpoint callback (doc) if you change the default params to save_best_only=True.
Agreed that the functionality is already provided in ModelCheckpoint, it just seemed a good addition to me to add it to the EarlyStopping as it would be the desired behaviour for most people and the best weights are already tracked so it's just a matter of returning them instead of having 2 callback functions tracking the same variables.
| gharchive/issue | 2022-08-08T14:44:26 | 2025-04-01T06:44:40.962705 | {
"authors": [
"cevheck",
"jbischof"
],
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/issues/16876",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
94198371 | Dynamic k-Max Pooling
I want to try the 1d CNN model in paper: A convolutional neural network for modelling sentences
the pdf:http://arxiv.org/pdf/1404.2188.pdf?utm_medium=App.net&utm_source=PourOver
How could I implement the Dynamic k-Max Pooling layer?
What have you tried so far?
I just know, it should select the activation of max top k units, and only update the weights of max top k units, and it should use mask. I don't know how to implement mask in theano.
You can try something like this
import numpy as np
import theano.tensor as T
from keras.layers.core import MaskedLayer
class KMaxPooling(MaskedLayer):
def __init__(self, pooling_size):
super(MaskedLayer, self).__init__()
self.pooling_size = pooling_size
self.input = T.tensor3()
def get_output_mask(self, train=False):
return None
def get_output(self, train=False):
data = self.get_input(train)
mask = self.get_input_mask(train)
if mask is None:
mask = T.sum(T.ones_like(data), axis=-1)
mask = mask.dimshuffle(0, 1, "x")
masked_data = T.switch(T.eq(mask, 0), -np.inf, data)
result = masked_data[T.arange(masked_data.shape[0]).dimshuffle(0, "x", "x"),
T.sort(T.argsort(masked_data, axis=1)[:, -self.pooling_size:, :], axis=1),
T.arange(masked_data.shape[2]).dimshuffle("x", "x", 0)]
return result
def get_config(self):
return {"name" : self.__class__.__name__, "pooling_size" : self.pooling_size}
Theano takes care of correct gradient propagation when you do array indexing, no need to worry about it
Did this ever make it into the main branch of Keras?
Is there a tensorflow implementation of the above code?
Not the most generic, well-engineered and thoroughly-tested solution, but this seems to do the trick for me:
from keras.engine import Layer, InputSpec
from keras.layers import Flatten
import tensorflow as tf
class KMaxPooling(Layer):
"""
K-max pooling layer that extracts the k-highest activations from a sequence (2nd dimension).
TensorFlow backend.
"""
def __init__(self, k=1, **kwargs):
super().__init__(**kwargs)
self.input_spec = InputSpec(ndim=3)
self.k = k
def compute_output_shape(self, input_shape):
return (input_shape[0], (input_shape[2] * self.k))
def call(self, inputs):
# swap last two dimensions since top_k will be applied along the last dimension
shifted_input = tf.transpose(inputs, [0, 2, 1])
# extract top_k, returns two tensors [values, indices]
top_k = tf.nn.top_k(shifted_input, k=self.k, sorted=True, name=None)[0]
# return flattened output
return Flatten()(top_k)
@arbackus please help figure out the right dimensions ?
https://stackoverflow.com/questions/45891148/k-maxpooling-layer-in-keras
`def Kmaxpooling(x,N_conv,present_conv):
#get shapes
inp_shape1=tf.shape(x)[1]
inp_shape0=tf.shape(x)[0]
inp_shape2=tf.shape(x)[2]
multiplier=N_conv-present_conv
k=inp_shape1*(multiplier)/(multiplier+1)
#transpose since topk works on last dim
transposed_x=tf.transpose(x, [0, 2, 1])
topk=tf.nn.top_k(transposed_x,k, sorted=True)
#sort the indices to maintain order
topk_inds=tf.negative(tf.nn.top_k(tf.negative(topk[1]),k, sorted=True)[0])
#Get the indices of all axis and merge them
indx_0=tf.reshape(tf.tile(tf.expand_dims(tf.range(inp_shape0),1),[1,tf.multiply(k,inp_shape2)]),[-1])
indx_2=tf.reshape(tf.tile(tf.expand_dims(tf.tile(tf.expand_dims(tf.range(inp_shape2),1),[1,k]),0),[inp_shape0,1,1]),[-1])
indx=tf.reshape(tf.stack([dim0,dim2,tf.reshape(topk_inds,[-1])],axis=1),[inp_shape0,inp_shape2,k,3])
res=tf.transpose(tf.gather_nd(transposed_x,indx), [0, 2, 1])
return res#tf.cast(indx,tf.float32)
@arbackus this solution works, but is very slow. In my case adding this more than doubles the entire network training time. A faster solution would be practical.
if the backend is tf, how to use the kmaxpooling ,for example:
conv1 = Conv2D(32, (3, 3), activation="relu", padding="same")(conv1)
pool1 = MaxPooling2D(pool_size=(2, 2))(conv1)
i replace MaxPooling2D to Kmaxpooling, but it don't work,so how to use the Kmaxpooling
@anttttti
Is this still an issue? Can you try with latest TF version? Thanks
Try this below code preserves the order in the of the original sequence. Not thoroughly tested but works fine.
`class KMaxPooling(layers.Layer):
"""
K-max pooling layer that extracts the k-highest activations from a sequence (2nd dimension).
TensorFlow backend.
"""
def init(self, k=1, axis=1, **kwargs):
super(KMaxPooling, self).init(**kwargs)
self.input_spec = layers.InputSpec(ndim=3)
self.k = k
assert axis in [1,2], 'expected dimensions (samples, filters, convolved_values),\
cannot fold along samples dimension or axis not in list [1,2]'
self.axis = axis
# need to switch the axis with the last elemnet
# to perform transpose for tok k elements since top_k works in last axis
self.transpose_perm = [0,1,2] #default
self.transpose_perm[self.axis] = 2
self.transpose_perm[2] = self.axis
def compute_output_shape(self, input_shape):
input_shape_list = list(input_shape)
input_shape_list[self.axis] = self.k
return tuple(input_shape_list)
def call(self, x):
# swap sequence dimension to get top k elements along axis=1
transposed_for_topk = tf.transpose(x, perm=self.transpose_perm)
# extract top_k, returns two tensors [values, indices]
top_k_vals, top_k_indices = tf.math.top_k(transposed_for_topk,
k=self.k, sorted=True,
name=None)
# maintain the order of values as in the paper
# sort indices
sorted_top_k_ind = tf.sort(top_k_indices)
flatten_seq = tf.reshape(transposed_for_topk, (-1,))
shape_seq = tf.shape(transposed_for_topk)
len_seq = tf.shape(flatten_seq)[0]
indices_seq = tf.range(len_seq)
indices_seq = tf.reshape(indices_seq, shape_seq)
indices_gather = tf.gather(indices_seq, 0, axis=-1)
indices_sum = tf.expand_dims(indices_gather, axis=-1)
sorted_top_k_ind += indices_sum
k_max_out = tf.gather(flatten_seq, sorted_top_k_ind)
# return back to normal dimension but now sequence dimension has only k elements
# performing another transpose will get the tensor back to its original shape
# but will have k as its axis_1 size
transposed_back = tf.transpose(k_max_out, perm=self.transpose_perm)
return transposed_back
class Folding(layers.Layer):
def __init__(self, **kwargs):
super(Folding, self).__init__(**kwargs)
self.input_spec = layers.InputSpec(ndim=3)
def compute_output_shape(self, input_shape):
return (input_shape[0], input_shape[1], int(input_shape[2]/2))
def call(self, x):
input_shape = x.get_shape().as_list()
# split the tensor along dimension 2 into dimension_axis_size/2
# which will give us 2 tensors
splits = tf.split(x, num_or_size_splits=int(input_shape[2]/2), axis=2)
# reduce sums of the pair of rows we have split onto
reduce_sums = [tf.reduce_sum(split, axis=2) for split in splits]
# stack them up along the same axis we have reduced
row_reduced = tf.stack(reduce_sums, axis=2)
return row_reduced`
Try this custom tensorflow keras layer. Not thoroughly tested but works fine.
`class KMaxPooling(layers.Layer):
"""
K-max pooling layer that extracts the k-highest activations from a sequence (2nd dimension).
TensorFlow backend.
"""
def init(self, k=1, axis=1, **kwargs):
super(KMaxPooling, self).init(**kwargs)
self.input_spec = layers.InputSpec(ndim=3)
self.k = k
assert axis in [1,2], 'expected dimensions (samples, filters, convolved_values),\
cannot fold along samples dimension or axis not in list [1,2]'
self.axis = axis
# need to switch the axis with the last elemnet
# to perform transpose for tok k elements since top_k works in last axis
self.transpose_perm = [0,1,2] #default
self.transpose_perm[self.axis] = 2
self.transpose_perm[2] = self.axis
def compute_output_shape(self, input_shape):
input_shape_list = list(input_shape)
input_shape_list[self.axis] = self.k
return tuple(input_shape_list)
def call(self, x):
# swap sequence dimension to get top k elements along axis=1
transposed_for_topk = tf.transpose(x, perm=self.transpose_perm)
# extract top_k, returns two tensors [values, indices]
top_k_vals, top_k_indices = tf.math.top_k(transposed_for_topk,
k=self.k, sorted=True,
name=None)
# maintain the order of values as in the paper
# sort indices
sorted_top_k_ind = tf.sort(top_k_indices)
flatten_seq = tf.reshape(transposed_for_topk, (-1,))
shape_seq = tf.shape(transposed_for_topk)
len_seq = tf.shape(flatten_seq)[0]
indices_seq = tf.range(len_seq)
indices_seq = tf.reshape(indices_seq, shape_seq)
indices_gather = tf.gather(indices_seq, 0, axis=-1)
indices_sum = tf.expand_dims(indices_gather, axis=-1)
sorted_top_k_ind += indices_sum
k_max_out = tf.gather(flatten_seq, sorted_top_k_ind)
# return back to normal dimension but now sequence dimension has only k elements
# performing another transpose will get the tensor back to its original shape
# but will have k as its axis_1 size
transposed_back = tf.transpose(k_max_out, perm=self.transpose_perm)
return transposed_back`
`class KMaxPooling(layers.Layer):
"""
K-max pooling layer that extracts the k-highest activations from a sequence (2nd dimension).
TensorFlow backend.
"""
def init(self, k=1, axis=1, **kwargs):
super(KMaxPooling, self).init(**kwargs)
self.input_spec = layers.InputSpec(ndim=3)
self.k = k
assert axis in [1,2], 'expected dimensions (samples, filters, convolved_values),\
cannot fold along samples dimension or axis not in list [1,2]'
self.axis = axis
# need to switch the axis with the last elemnet
# to perform transpose for tok k elements since top_k works in last axis
self.transpose_perm = [0,1,2] #default
self.transpose_perm[self.axis] = 2
self.transpose_perm[2] = self.axis
def compute_output_shape(self, input_shape):
input_shape_list = list(input_shape)
input_shape_list[self.axis] = self.k
return tuple(input_shape_list)
def call(self, x):
# swap sequence dimension to get top k elements along axis=1
transposed_for_topk = tf.transpose(x, perm=self.transpose_perm)
# extract top_k, returns two tensors [values, indices]
top_k_vals, top_k_indices = tf.math.top_k(transposed_for_topk,
k=self.k, sorted=True,
name=None)
# maintain the order of values as in the paper
# sort indices
sorted_top_k_ind = tf.sort(top_k_indices)
flatten_seq = tf.reshape(transposed_for_topk, (-1,))
shape_seq = tf.shape(transposed_for_topk)
len_seq = tf.shape(flatten_seq)[0]
indices_seq = tf.range(len_seq)
indices_seq = tf.reshape(indices_seq, shape_seq)
indices_gather = tf.gather(indices_seq, 0, axis=-1)
indices_sum = tf.expand_dims(indices_gather, axis=-1)
sorted_top_k_ind += indices_sum
k_max_out = tf.gather(flatten_seq, sorted_top_k_ind)
# return back to normal dimension but now sequence dimension has only k elements
# performing another transpose will get the tensor back to its original shape
# but will have k as its axis_1 size
transposed_back = tf.transpose(k_max_out, perm=self.transpose_perm)
return transposed_back`
Try this custom layer. Not thoroughly tested but works fine for me. Let me know how it works for you.
P.S. Latest tensorflow version
`class KMaxPooling(layers.Layer):
def init(self, k=1, axis=1, **kwargs):
super(KMaxPooling, self).init(**kwargs)
self.input_spec = layers.InputSpec(ndim=3)
self.k = k
assert axis in [1,2], 'expected dimensions (samples, filters, convolved_values),\
cannot fold along samples dimension or axis not in list [1,2]'
self.axis = axis
# need to switch the axis with the last elemnet
# to perform transpose for tok k elements since top_k works in last axis
self.transpose_perm = [0,1,2] #default
self.transpose_perm[self.axis] = 2
self.transpose_perm[2] = self.axis
def compute_output_shape(self, input_shape):
input_shape_list = list(input_shape)
input_shape_list[self.axis] = self.k
return tuple(input_shape_list)
def call(self, x):
# swap sequence dimension to get top k elements along axis=1
transposed_for_topk = tf.transpose(x, perm=self.transpose_perm)
# extract top_k, returns two tensors [values, indices]
top_k_vals, top_k_indices = tf.math.top_k(transposed_for_topk,
k=self.k, sorted=True,
name=None)
# maintain the order of values as in the paper
# sort indices
sorted_top_k_ind = tf.sort(top_k_indices)
flatten_seq = tf.reshape(transposed_for_topk, (-1,))
shape_seq = tf.shape(transposed_for_topk)
len_seq = tf.shape(flatten_seq)[0]
indices_seq = tf.range(len_seq)
indices_seq = tf.reshape(indices_seq, shape_seq)
indices_gather = tf.gather(indices_seq, 0, axis=-1)
indices_sum = tf.expand_dims(indices_gather, axis=-1)
sorted_top_k_ind += indices_sum
k_max_out = tf.gather(flatten_seq, sorted_top_k_ind)
# return back to normal dimension but now sequence dimension has only k elements
# performing another transpose will get the tensor back to its original shape
# but will have k as its axis_1 size
transposed_back = tf.transpose(k_max_out, perm=self.transpose_perm)
return transposed_back`
Check this out a custom keras layer. Not thoroughly tested but works fine for me. Let me know what you think. P.S. TF 2.1.0
import tensorflow as tf
from tensorflow.keras import layers
class KMaxPooling(layers.Layer):
"""
K-max pooling layer that extracts the k-highest activations from a sequence (2nd dimension).
TensorFlow backend.
"""
def __init__(self, k=1, axis=1, **kwargs):
super(KMaxPooling, self).__init__(**kwargs)
self.input_spec = layers.InputSpec(ndim=3)
self.k = k
assert axis in [1,2], 'expected dimensions (samples, filters, convolved_values),\
cannot fold along samples dimension or axis not in list [1,2]'
self.axis = axis
# need to switch the axis with the last elemnet
# to perform transpose for tok k elements since top_k works in last axis
self.transpose_perm = [0,1,2] #default
self.transpose_perm[self.axis] = 2
self.transpose_perm[2] = self.axis
def compute_output_shape(self, input_shape):
input_shape_list = list(input_shape)
input_shape_list[self.axis] = self.k
return tuple(input_shape_list)
def call(self, x):
# swap sequence dimension to get top k elements along axis=1
transposed_for_topk = tf.transpose(x, perm=self.transpose_perm)
# extract top_k, returns two tensors [values, indices]
top_k_vals, top_k_indices = tf.math.top_k(transposed_for_topk,
k=self.k, sorted=True,
name=None)
# maintain the order of values as in the paper
# sort indices
sorted_top_k_ind = tf.sort(top_k_indices)
flatten_seq = tf.reshape(transposed_for_topk, (-1,))
shape_seq = tf.shape(transposed_for_topk)
len_seq = tf.shape(flatten_seq)[0]
indices_seq = tf.range(len_seq)
indices_seq = tf.reshape(indices_seq, shape_seq)
indices_gather = tf.gather(indices_seq, 0, axis=-1)
indices_sum = tf.expand_dims(indices_gather, axis=-1)
sorted_top_k_ind += indices_sum
k_max_out = tf.gather(flatten_seq, sorted_top_k_ind)
# return back to normal dimension but now sequence dimension has only k elements
# performing another transpose will get the tensor back to its original shape
# but will have k as its axis_1 size
transposed_back = tf.transpose(k_max_out, perm=self.transpose_perm)
return transposed_back
@makaveli10 How to use it instead of Maxpooling?
| gharchive/issue | 2015-07-10T02:36:29 | 2025-04-01T06:44:40.983431 | {
"authors": [
"andcut",
"anttttti",
"arbackus",
"bicepjai",
"fchollet",
"iskandr",
"jvishnuvardhan",
"makaveli10",
"manan15105411262",
"ngragaei",
"peterGuang",
"rainmakerr",
"samskruthireddy"
],
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/issues/373",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
257616447 | Pad_sequences throwing error with texts_to_sequence_generator
After creating a text to sequence generator using texts_to_sequences_generator, when that generator object is passed as argument to the pad_sequences. It throws the following error.
ValueError Traceback (most recent call last)
in ()
1 seq2 = tokenizer.texts_to_sequences_generator(sent)
----> 2 data2 = pad_sequences(seq2,maxlen=maxSeqLength)
/opt/conda/lib/python3.5/site-packages/keras/preprocessing/sequence.py in pad_sequences(sequences, maxlen, dtype, padding, truncating, value)
35 """
36 if not hasattr(sequences, 'len'):
---> 37 raise ValueError('sequences must be iterable.')
38 lengths = []
39 for x in sequences:
ValueError: sequences must be iterable.
I reduced the code to bare minimum for my use case. Assuming post padding and int32 data type.
def pad_sequences(sequences, maxlen):
for s in sequences:
if len(s) == 0: yield np.zeros((1, maxlen), dtype='int32')
s = np.array(s, dtype='int32')
yield np.expand_dims(np.pad(s[:maxlen], (0, max(0, maxlen - len(s))), 'constant'), axis=0)
This doesn't return a matrix rather a generator! So, any of your following step need to take care of that.
Bonus tip:
To clone generators --
Use tee
from itertools import tee
| gharchive/issue | 2017-09-14T07:07:15 | 2025-04-01T06:44:40.989578 | {
"authors": [
"ishmeetSRaina",
"pbamotra"
],
"repo": "keras-team/keras",
"url": "https://github.com/keras-team/keras/issues/7894",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2457365697 | bigquery.Copy not working with validation error
Expected Behavior
No response
Actual Behaviour
This cannot be save due to this error:
Internal server error: HV000030: No validator could be found for constraint ‘io.kestra.plugin.gcp.bigquery.StoreFetchDestinationValidation’ validating type ‘io.kestra.plugin.gcp.bigquery.Copy’. Check configuration for ‘tasks[0]’
id: test_bq_copy
namespace: kestra.sandbox
tasks:
- id: snapshot_table
type: io.kestra.plugin.gcp.bigquery.Copy
destinationTable: kestra-dev.bpimpaud.new_data_test
operationType: COPY
sourceTables:
- kestra-dev.bpimpaud.data_from_db
projectId: kestra-dev
serviceAccount: "{{ secret('GCP_CREDS') }}"
The flow is fairly basic and use the example from plugin documentation
Steps To Reproduce
No response
Environment Information
Kestra Version: 0.18.0
Plugin version:
Operating System (OS / Docker / Kubernetes): docker
Java Version (If not docker):
Example flow
No response
switch REST or BULK, error persist
| gharchive/issue | 2024-08-09T08:09:17 | 2025-04-01T06:44:41.005667 | {
"authors": [
"Ben8t",
"fikafetsy"
],
"repo": "kestra-io/plugin-gcp",
"url": "https://github.com/kestra-io/plugin-gcp/issues/428",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
614339965 | Installation Steps?
Hi Ketan,
If Kindly you could help with how we can install and setup this project it would be really helpful and useful.
Thank you
You can refer README.md file I've updated it and you may need to make some changes in code to make it work in your region.
Thanks
| gharchive/issue | 2020-05-07T20:58:57 | 2025-04-01T06:44:41.007708 | {
"authors": [
"SuhailSumra",
"ketanchoyal"
],
"repo": "ketanchoyal/theParker",
"url": "https://github.com/ketanchoyal/theParker/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1646770600 | Catch pydantic validation errors with one validation type
I wonder if it is possible to modify the AsdfPydanticModel to raise the same ValidationError as ASDF does? That way libraries reading asdf files will on need to worry about catching one error type.
@WilliamJamieson
ASDF's error is a direct use of jsonschema's which is not a subclass of a standard type
from jsonschema import ValidationError
Pydantic's error is a subclass of TypeError
from pydantic.error_wrappers import ValidationError
I really would love to use ExceptionGroup (new in Python 3.11) for this because there are situations where I want to catch only pydantic validation error.
Maybe asdf should define a AsdfValidationError which inherits from both TypeError and jsonschema.ValidationError? I think that would be a reasonable improvement on asdf's part.
Desired result:
from asdf import ValidationError
try:
af.open(...)
except ValidationError:
...
After Python 3.11,
from asdf import ValidationError
# Very verbose, could forward this asdf_pydantic
from pydantic.error_wrappers import ValidationError as PydanticValidationError
try:
af.open(...)
except* ValidationError:
...
except* PydanticValidationError:
...
Good idea on inheriting TypeError. I'd keep the same name ValidationError for compatibilty.
Conversely, since this is such a new library you could just require Python 3.11 to use it.
I will consider it; I have things in Roman that are 3.10.
| gharchive/issue | 2023-03-30T01:36:56 | 2025-04-01T06:44:41.012498 | {
"authors": [
"WilliamJamieson",
"ketozhang"
],
"repo": "ketozhang/asdf-pydantic",
"url": "https://github.com/ketozhang/asdf-pydantic/issues/2",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
209680472 | Music similarity using music-auto_tagging for feature extraction
Hi ,
I have used the music-auto_tagging for feature extraction along with a classifier and come up with an approach to recommend music based on music similarity.
Accoustically similar music have some accoustic features which can be used to identify similar sounding music. We have considered the following four accoustic features in order to distinguish different kind of music.
1)Drag : Songs which are very slow moving
2)Beats : Songs in which rythmic beats are more prominent and highlighted
3)Melody : Songs without too much beats /slow or medium tempo (speed) /
4)Fast : Songs which have fast tempo (speed) .
We picked manually around 60-80 music samples under each of the above categories.
Then we formed a binary cluster for each category as follows :
For example in order to train for Drag category , we formed two groups with 65 samples in each group.
1)Drag group (65 samples) : all samples in this group are very slow moving songs.
2)Others group (65 samples ) : We mixed samples from beats melody and fast category which are not slow moving but with medium to fast tempo(speed).
Then we trained a convolutional neural network (CNN) model using normalized spectrograms of all samples from both groups. The trained model was used to identify the dragscore for a song.
Using the same approach as explained above , separate CNN models were trained in order to get beatsscore, melodyscore and fastscore from a song.
Now using the four separately trained CNN models , we obtained dragscore,beatsscore,melodyscore and fastscore for every song in the test data (around 3000+ music samples) which were not part of training data.
For any selected song from test set , based on the four scores of the selected sample song , songs which have similar scores in the test set were listed in the result page with option to listen to the selected and listed samples.
The listing based on score similarity seemed to match the selected song in terms of music similarity for 85-90% of the selections.
So it works well? then it's good news :)
Thanks. I have tried with very few samples only (around 60-70 samples per category). I will need to try with more samples per category and check if I can get better accuracy .
| gharchive/issue | 2017-02-23T07:00:30 | 2025-04-01T06:44:41.016914 | {
"authors": [
"keunwoochoi",
"srinidhikrs"
],
"repo": "keunwoochoi/music-auto_tagging-keras",
"url": "https://github.com/keunwoochoi/music-auto_tagging-keras/issues/17",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
251490451 | Deserializing array of arbitrary values
I am interested in deserializing arbitrary values contained in JSON array. The contents and data types inside the array may change anytime.
The objective is to deserialize data so it may be further processed and stored to disk using more effective serialization. The performance is key factor here, which brought me to investigate Jil.
Using Json.NET this can be achieved quite easily.
string json = @"[1, true, 1.2345, null, ""test""]";
var values = JsonConvert.DeserializeObject<object[]>(json);
Values array would then contain the values as .NET (boxed) primitive types (long, bool, double, object, string). The array would then be really easy to use for further processing and efficient storage.
However the Json.NET deserialization has quite large performance overhead.
Deserializing the same data with Jil seems to produce dynamic JsonObjects. I am having hard time figuring out what to do with the dynamic objects to get to the "raw" data. Is there feasible to way to achieve what I am trying to do using Jil?
It works, but it's quite a bit of work and quite error prone, especially if you don't know anything about the types you are deserializing (e.g. your null value above, decimal vs. double, Guids etc.).
The steps are:
obtain typeconverter: var converter = TypeDescriptor.GetConverter(dynamic);
Find out if the dynamic is a complex object by converting it to a Dictionary<string,object>
Find out if the dynamic is an array by converting it to an array etc.
Find out if the dynamic is a string, value types, primitve types by converting it to string etc.
And repeat that until your done, TypeConverters have a method CanConvertTo which will try the conversion without throwing an exception. This is not designed for performance e.g. CanConvertTo then doing ConvertTo basically amounts to doing the conversion twice. Personally I wouldn't try this if you don't know the types you're deserializing before hand, simply because JSON is not as typesafe as C#.
Thanks Tornhoof. I did try your suggestion and indeed I was able to get it working quite similarly as Json.NET. However the conversions via the TypeConverter absolutely ruined the performance.
Yes, as I wrote, it does the conversion twice, if you know the types, you can skip the CanConvertTo step and it's pretty much twice as fast, but even then it's only marginally faster than JSON.NET.
| gharchive/issue | 2017-08-20T14:02:59 | 2025-04-01T06:44:41.066050 | {
"authors": [
"Tornhoof",
"juho-hanhimaki"
],
"repo": "kevin-montrose/Jil",
"url": "https://github.com/kevin-montrose/Jil/issues/273",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1677312110 | Skip masking trace, etc when bias_correction is None; enable row-by-row BG subtraction; edit NIRCam extraction height
This resolves #522 by skipping a bunch of unnecessary steps in bias_sub.py when meta.bias_correction is None.
Got it! I had fixed a related problem, but not the exact one you were describing.
@taylorbell57 Do you have any other concerns about this PR before we merge?
| gharchive/pull-request | 2023-04-20T19:43:26 | 2025-04-01T06:44:41.067859 | {
"authors": [
"kevin218"
],
"repo": "kevin218/Eureka",
"url": "https://github.com/kevin218/Eureka/pull/524",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2376095618 | 🛑 FX blog (kty) is down
In 6b4d2c9, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in 9a12a12 after 6 minutes.
| gharchive/issue | 2024-06-26T19:17:49 | 2025-04-01T06:44:41.097693 | {
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/10654",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2383780673 | 🛑 FX blog (kty) is down
In 8175af1, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in 2b90f00 after 8 minutes.
| gharchive/issue | 2024-07-01T12:54:32 | 2025-04-01T06:44:41.100259 | {
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/12019",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2390134726 | 🛑 FX corretor (kty) is down
In e2e9936, FX corretor (kty) (https://corretores.foxterciaimobiliaria.com.br) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX corretor (kty) is back up in 2bddbb2 after 7 minutes.
| gharchive/issue | 2024-07-04T07:17:55 | 2025-04-01T06:44:41.102703 | {
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/12759",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2407389066 | 🛑 FX blog (kty) is down
In 9f29a0c, FX blog (kty) (https://foxter-blogeditor.konecty.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: FX blog (kty) is back up in f219cae after 26 minutes.
| gharchive/issue | 2024-07-14T11:23:24 | 2025-04-01T06:44:41.105165 | {
"authors": [
"keviocastro"
],
"repo": "keviocastro/upptime",
"url": "https://github.com/keviocastro/upptime/issues/15395",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.