id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
781818944 | Migrating from Travis CI to GitHub Actions
Travis CI stop working on open source projects.
Migrate the CI server from Travis CI to GitHub Actions, or circles CI alternatively.
done
done
| gharchive/issue | 2021-01-08T03:43:08 | 2025-04-01T06:46:05.749185 | {
"authors": [
"hrchu"
],
"repo": "twonote/radosgw-admin4j",
"url": "https://github.com/twonote/radosgw-admin4j/issues/37",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
111802954 | Maxy/2565
save easyform values in the notebook #2565
:+1:
| gharchive/pull-request | 2015-10-16T10:19:40 | 2025-04-01T06:46:05.750226 | {
"authors": [
"maxyeg",
"scottdraves"
],
"repo": "twosigma/beaker-notebook",
"url": "https://github.com/twosigma/beaker-notebook/pull/2739",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1995850101 | [要望]ログ検索条件のノード指定対象をソートしてもらえませんでしょうか
ログ検索条件指定時のノード選択は、現状ノード選択になっていますけれど、
ソートされていない為対象を探すのが大変です。ソートして頂けませんか。
テキスト入力で、例えば頭文字入力でドロップダウンの選択対象ノードにジャンプしてくれると非常に助かります。
他でも使っているautocompleteが使えると思うので検討します。
Autocompleteにしました。
ありがとうございました。
| gharchive/issue | 2023-11-16T01:08:53 | 2025-04-01T06:46:05.791603 | {
"authors": [
"sadayoshi-git",
"twsnmp"
],
"repo": "twsnmp/twsnmpfc",
"url": "https://github.com/twsnmp/twsnmpfc/issues/28",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
11356204 | Add funcionality to expand and collapse notes
It would be great to have the opportunity to collapse and expand the notes, one by one or all together at once.
Great job.
In the new version, you have the possibility to expand, minimize, hide and show notes.
Thanks for your interest :)
| gharchive/issue | 2013-02-25T12:03:38 | 2025-04-01T06:46:05.797025 | {
"authors": [
"rafagarcia",
"txusko"
],
"repo": "txusko/PostItAll",
"url": "https://github.com/txusko/PostItAll/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1898072222 | avoid needing Node to build wcc
I’m planning to package xcc and wcc for Guix. Besides the need for an install target, a big impediment is currently the dependency on Node to build wcc.
I know downstream packaging probably isn’t a very significant consideration for you, but if it would be possible to consider this kind of change, it is important to me! I can work on adding an install target later too, if you don’t have anything against it.
@zamfofex Thank you to reduce dependency!
| gharchive/pull-request | 2023-09-15T09:39:21 | 2025-04-01T06:46:05.801499 | {
"authors": [
"tyfkda",
"zamfofex"
],
"repo": "tyfkda/xcc",
"url": "https://github.com/tyfkda/xcc/pull/126",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
258524719 | Fix Module Index in the docs
http://eta-lang.org/docs/html/index.html
The Module Index link in the page above is not working.
We have moved to our own custom documentation system so this is no longer relevant.
| gharchive/issue | 2017-09-18T15:37:17 | 2025-04-01T06:46:05.885626 | {
"authors": [
"rahulmutt"
],
"repo": "typelead/eta",
"url": "https://github.com/typelead/eta/issues/522",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
991600795 | Update specs2-core to 4.12.11
Updates org.specs2:specs2-core from 4.12.10 to 4.12.11.
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Ignore future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "org.specs2", artifactId = "specs2-core" } ]
labels: library-update, semver-patch
Superseded by https://github.com/typelevel/cats-effect-testing/pull/203
@sh0hei actually, IMO you should still merge this. It's unclear when #203 will be merged and may involve a branching/version bump.
| gharchive/pull-request | 2021-09-08T23:09:43 | 2025-04-01T06:46:05.889361 | {
"authors": [
"armanbilge",
"scala-steward",
"sh0hei"
],
"repo": "typelevel/cats-effect-testing",
"url": "https://github.com/typelevel/cats-effect-testing/pull/204",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2247770684 | Release latest dependency updates
Hi, would be great if we could get a release of this lib, lots of updates have happened since 2.6.0. Thanks!
Yeah, it's been a while since the last release. Not only have the dependencies been updated, but we've even got some new features. @typelevel/log4cats thoughts?
https://github.com/typelevel/log4cats/releases/tag/v2.7.0
Artifacts should appear in Maven Central in 4-8 hours normally.
| gharchive/issue | 2024-04-17T09:04:47 | 2025-04-01T06:46:05.932259 | {
"authors": [
"danicheg",
"howyp"
],
"repo": "typelevel/log4cats",
"url": "https://github.com/typelevel/log4cats/issues/832",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1696202002 | Aliases for AskVault and LocalVault
I don't generally like aliases, but in the spirit of MonadCancelThrow, I think we need one here, else we get gangling lambdas.
I think this problem is more acute in libraries than apps, because the kind-projector syntax of Scala 2[^1] isn't supported in Scala 3. But, still.
Proposal:
type LocalVault[F] = Local[F, Vault]
type AskVault[F] = Ask[F, Vault]
[^1]: Local[*[_], Vault]
where do you think this should go?
I think a package object org.typelevel.otel4s in core-common.
We have an awkward split between all and common both using the top-level package. The Java API puts all the context-related stuff in a context subpackage, which we might want to consider mimicking. 🤔
May I also claim this one & learn more about type aliases along the way?
| gharchive/issue | 2023-05-04T15:01:51 | 2025-04-01T06:46:05.935790 | {
"authors": [
"AprilAtBanno",
"rossabaker",
"sherriesyt"
],
"repo": "typelevel/otel4s",
"url": "https://github.com/typelevel/otel4s/issues/198",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1389851393 | Consider choosing custom domain
One another thing, related to the name debate, you should pause and reflect before 1.0 is whether to keep the current website URL or move to your own domain name.
https://krzema12.github.io/github-actions-kotlin-dsl/
vs
https://gakd.io (not the actual name)
My experience on this issue:
The docs for refreshVersions are also stored on GitHub pages at https://jmfayard.github.io/refreshVersions/
Now I thought many times whether I should move away from GitHub pages or not.
Because to be frank it's more a quick and dirty solution than a great long term solution.
I looks more professional to have your own DNS like https://kotest.io/
Now you may think that you can just choose your DNS and keep GitHub pages.
But there are some support for it, the documentation and everything else is in fact quite bad, it was never designed for that.
My experience is that https://www.netlify.com/ takes care of everything (deployment, reporting, custom domain name, dns configuration, SSL, ...) in a much easier way.
I'm thinking about a compromise: creating a GitHub org named github-workflows-kt, thanks to this we could have docs hosted at https://github-workflows-kt.github.io/docs (in docs repository) and it's for free. For now I have mixed feeling when thinking about paying over 200 PLN for a custom domain.
A possibly bearable price:
That's the normal price. It's only .io which is expansive.
The domain name is a nice too have and optional though, don't worry too much about it.
Requesting new group ID in Maven Central: https://issues.sonatype.org/browse/OSSRH-90368
This is almost done, let's wait for publishing the next release to confirm that everything works fine after moving things around.
TODO: adjust the teaser image https://github.com/typesafegithub/github-workflows-kt/commit/1c612284ef71e1e7c44709485f28530d1c92fd7a
| gharchive/issue | 2022-09-28T19:31:39 | 2025-04-01T06:46:05.971805 | {
"authors": [
"jmfayard",
"krzema12"
],
"repo": "typesafegithub/github-workflows-kt",
"url": "https://github.com/typesafegithub/github-workflows-kt/issues/465",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
146630969 | Handling existing conf file upgrade
Has anyone implemented upgrade handling for conf files?
Example scenario:
Application has a hidden default conf file and a modifiable user conf file with the same properties
Conf files are loaded so that user conf file values override the default conf file values
Both conf files are validated using a schema
In the next version of the application the default conf file changes:
Properties are either added or removed
Conf file loading will still work
The user conf file validation will fail
The user conf file will now have missing or deprecated properties
So the user conf file has to be modified to keep it valid and in sync with the new default conf file, without losing any value changes the user might have done.
It isn't clear to me which part of the scenario you're suggesting the library could handle, vs. which part the application would handle, vs. which part user would have to do manually?
I'm not quite sure. Currently I have a custom implementation for the schema validation part.
The library could provide a function for comparing two conf files, writing missing property keys and removing unnecessary property keys. But this kind of functionality can't automatically handle all the possible changes in property values. It would have to know when a property value is valid within the new schema or not.
So I guess an ultimate solution would be a validator with automatic patching functionality.
| gharchive/issue | 2016-04-07T13:46:35 | 2025-04-01T06:46:05.975835 | {
"authors": [
"havocp",
"jmullo"
],
"repo": "typesafehub/config",
"url": "https://github.com/typesafehub/config/issues/389",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2266884672 | upgrade gradle action in server-check
There's a new recommended version (and it's a good excuse to check with the build isn't caching)
fixed in #105
| gharchive/issue | 2024-04-27T07:18:35 | 2025-04-01T06:46:06.048455 | {
"authors": [
"lucapette"
],
"repo": "typestreamio/typestream",
"url": "https://github.com/typestreamio/typestream/issues/104",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
83660017 | Getting a range
Is it possible to filter based on a range?
for example if I have items with a price field
id: 0,
price : 10
do something like get all items with price between 10 and 20? and paginate?
P.S. Great library!
+1, I'd love to be able to filter like that:
GET /products?price={gte:10,lte:20}
or maybe
GET /products?price__gte=10&price__lte=20
gt, lt, gte, lte are widely used (MongoDB, ElasticSearch, Django...)
Just added the feature in v0.7.28. Thank you for the feedback and suggestions :)
Awesome, you rock!
| gharchive/issue | 2015-06-01T19:40:37 | 2025-04-01T06:46:06.081222 | {
"authors": [
"ObjectiveTruth",
"fzaninotto",
"typicode"
],
"repo": "typicode/json-server",
"url": "https://github.com/typicode/json-server/issues/108",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
133175492 | Added a Typings, the Registry and Versions doc
This is to resolve https://github.com/typings/typings/issues/191
I think this is pretty good for now. I did wonder if it would be worth adding in a section at the end to outline the use case where someone wants to carry on evolving the type definition for an older version of a package (eg Angular 1.2.x users who can't upgrade due to IE support) and how the Typings Registry allows for that.
If you think that'd be handy I can submit a subsequent PR when I get some time to cater for that.
Is this all OK otherwise? I tweaked it a little from the original issue but not a great deal.
Suggestion: missing "," in the last example. :smile:
Suggestion 2: Don't write it too well, the one I wrote looks so amateur comparing to yours. :smirk: :+1:
Suggestion 3: you can use "```json" on the code block to enable syntax highlighting. e.g.:
{
"versions": {
"1.7.0": "github:some-kind-soul/moment-1-7-0-typing#iamahash1iamahash1iamahash1",
"2.10.5": "github:typed-typings/npm-moment#a4075cd50e63efbedd850f654594f293ab81a385"
}
}
All valid contributions @unional! I didn't know about the JSON syntax highlighting - I intend to use that.
Amended.
Nice :+1: Feel free to keep iterating, but I'm happy to merge already.
Sweet - thanks.
| gharchive/pull-request | 2016-02-12T07:21:52 | 2025-04-01T06:46:06.085928 | {
"authors": [
"blakeembrey",
"johnnyreilly",
"unional"
],
"repo": "typings/typings",
"url": "https://github.com/typings/typings/pull/199",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2374319222 | 按钮点击无响应
现在插件中,点击左侧 文件夹/列表 button没有反应,只有通过点击搜索之后再进入才能正确响应
请问有能够稳定复现的步骤吗?没有的话,没那么快搞好😂
哦,懂你意思了
折叠侧栏
点击“文件管理”或“大纲” -> 无法打开侧边栏
请更新到 2.0.13
可以在 左键点击右下角“设置” -> 选择“插件设置” -> “关于”Tab -> 点击“更新”按钮
| gharchive/issue | 2024-06-26T05:26:36 | 2025-04-01T06:46:06.091214 | {
"authors": [
"plylrnsdy",
"zzicarus"
],
"repo": "typora-community-plugin/typora-community-plugin",
"url": "https://github.com/typora-community-plugin/typora-community-plugin/issues/17",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2444516375 | academicons:0.1.0
I am submitting
[x] a new package
[ ] an update for a package
Description: This package provides functions to use icons of Academicons conveniently by leveraging the desktop fonts. One can simply add icons via their name once the font is installed.
I have read and followed the submission guidelines and, in particular, I
[x] selected a name that isn't the most obvious or canonical name for what the package does
[x] added a typst.toml file with all required keys
[x] added a README.md with documentation for my package
[x] have chosen a license and added a LICENSE file or linked one in my README.md
[x] tested my package locally on my system and it worked
[x] excluded PDFs or README images, if any, but not the LICENSE
[x] ensured that my package is licensed such that users can use and distribute the contents of its template directory without restriction, after modifying them through normal use.
Thanks for the quick review and sorry for the copy/paste error. Regarding authorship, I sent an email, but I'm waiting for the answer.
Thanks for the package submission. It all looks good to me, excepted for the name that is canonical (it is true fontawesome was accepted, but that was a mistake). Could you please change it to make it a bit more unique?
Thanks for the package submission. It all looks good to me, excepted for the name that is canonical (it is true fontawesome was accepted, but that was a mistake). Could you please change it to make it a bit more unique?
How about "using-academicons" or "import-academicons" ?
I personally feel like it would be a bit redundant with the #import keyword, but it would fit the naming guidelines I think, so if you want you can use that :)
| gharchive/pull-request | 2024-08-02T09:25:51 | 2025-04-01T06:46:06.108445 | {
"authors": [
"bpkleer",
"elegaanz"
],
"repo": "typst/packages",
"url": "https://github.com/typst/packages/pull/856",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1856530024 | Hanging-indent doesn't work correctly for lists inside term lists
Description
The following
#set page("a7")
/ Hello:
- #lorem(20)
- World
results in
Reproduction URL
No response
Operating system
No response
Typst version
[X] I am using the latest version of Typst
I can add that this is related to hang-indent. It doesn't seem like the lines that include the bullet points are indented at all.
I think this was fixed.
| gharchive/issue | 2023-08-18T11:05:07 | 2025-04-01T06:46:06.111658 | {
"authors": [
"Enivex",
"laurmaedje"
],
"repo": "typst/typst",
"url": "https://github.com/typst/typst/issues/1955",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
329142473 | Provide complete initial downloadable archives
@tzaeru : This is mostly a note for me unless you do it sooner (although you'd certainly be much more capable at this than me)
Goal: Host complete initial game versions on this backend as well
Currently we'd still need to distribute the initial version someplace else.
Our download links could instead provide downloads with the spring-launcher + game dist_cfg included.
Basically, I want to transform the current download links to:
http://tzaeru.com:4445/download?path=Balanced-Annihilation-Linux.zip
http://tzaeru.com:4445/download?path=Balanced-Annihilation-Windows.zip
http://tzaeru.com:4445/download?path=SpringBoard-Linux.zip
http://tzaeru.com:4445/download?path=SpringBoard-Windows.zip
This would be done by taking the launcher (frozen) + dist_cfg files from a game (and using those files to overwrite the launcher), and providing this in one download
This would allow us to use the backend for initial downloads - which are always at the latest state of the launcher+game (don't need self-updates)
It also makes it easier for game devs, as they don't have to host their own initial files (i've been using google docs now and it feels like a bad idea, especially as it's not automated)
PS: Once this issue is done I would start using SpringBoard with this launcher-backend system completely.
Also, the executable file (launcher.exe or launcher) should be renamed to game-appropriate name, e.g. SpringBoard.exe, SpringBoard, Balanced-Annihilation.exe, Balanced-Annihilation, depending on the game (human) name and platform
| gharchive/issue | 2018-06-04T17:09:15 | 2025-04-01T06:46:06.116383 | {
"authors": [
"gajop"
],
"repo": "tzaeru/chobby-launcher-backend",
"url": "https://github.com/tzaeru/chobby-launcher-backend/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1416202791 | 日志行数超过200就开始报错了,但有时候光一个ls就会弄好多行
日志行数超过200就开始报错了,但有时候光一个ls就会弄好多行
他就会报满屏的错误
Originally posted by @18030705033 in https://github.com/tzfun/vue-web-terminal/issues/2#issuecomment-1285146051
对于“一个ls就会弄出好多行”请描述更具体一点,另外提供一下报错信息
比如我一次发送500条数据,200行以内正常展示,另外会发送300条的提示消息 ,内容就是超过200条就会影响浏览器性能,按理来说如果想提示,就应该在200条的时候结束发送,然后给一条提示就可以了,没必要来个300条。还有一个问题,做拖动不好做到header上面啊,因为时动态的,做到全局,滚动条又没办法使用了,这是个问题。还有就是滚动条按理来说时做给主体,header应该是固定的,目前都是自己些css,还是比较麻烦
然后按理说提示的消息完了以后,你让用户使用clear来清空,那也应该能让用户输入啊,现在是提示了光标也消失了。没有重开一行让用户输入。就只能启用新的,历史数据就没有了啊
警告提示信息目前确实会存在这个问题,下次版本更新会修复,目前你可以先扩大warnLogCountLimit值临时解决;
目前没有考虑做拖动效果,因此header没有提前设计为拖动区域,目前的解决办法只有先隐藏header自己在terminal外部包装实现一个边框;
header之所以设计为悬浮的是为了实现隐藏效果,后续看一下能不能优化这个痛点。
小宾。
@.***
其实大佬你可以考虑做拖动和固定表头,固定表头的好处挺多的,比如我现在很多命令行了,关闭还好,直接用命令,如果我突然要最小化和最大化,难道还拖拖动条到初始位置吗,这个在设计上不太合理,拖动的话其实很正常,毕竟终端的宽高不可能太小,很容易挡住一些用户需要看到的东西,所以拖动在所难免。还有就是ls这个命令的处理。每一个文件如果单独发送一条信息又太浪费,放在一起又太拥挤,用空格隔开又不能判定单词来换行。关键使用终端ls这个命令还是最常用的
------------------ 原始邮件 ------------------
发件人: "tzfun/vue-web-terminal" @.>;
发送时间: 2022年10月20日(星期四) 下午4:58
@.>;
@.@.>;
主题: Re: [tzfun/vue-web-terminal] 日志行数超过200就开始报错了,但有时候光一个ls就会弄好多行 (Issue #7)
然后按理说提示的消息完了以后,你让用户使用clear来清空,那也应该能让用户输入啊,现在是提示了光标也消失了。没有重开一行让用户输入。就只能启用新的,历史数据就没有了啊
这个问题我这边测试没有遇到,光标仍然处于激活状态,可以键入
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you were mentioned.Message ID: @.***>
警告提示确实存在Bug,我会尽快修复并发布新的版本。对于表头是否固定问题不同的应用场景有不同的需求,有些开发者在使用时提出想表头因此才将其设计为悬浮式的,针对你这边的场景可以考虑下使用fullscreen接口直接全屏,对于header的设计我再考虑下是否有更好的解决方案,感谢你的建议
还有就是ls这个命令的处理。每一个文件如果单独发送一条信息又太浪费,放在一起又太拥挤,用空格隔开又不能判定单词来换行。关键使用终端ls这个命令还是最常用的
对于提到的这个问题你可以通过api来实现,我在文档中也有提到:
onExecCmd(key, command, success, failed) {
if (key === 'loop') {
let loop = parseInt(command.split(" ")[1])
for (let i = 0; i < loop; i++) {
Terminal.$api.pushMessage("my-terminal", {
type: "normal",
content: "loop => " + i
})
}
success()
}
}
实现效果
现在遇到的场景是,如果一个ls,返回值是200个数据,那么就相当于要发200条数据,其实不方便查看,但是如果发一条数据的话,目前换行并不是以单词为单位来换行](word-wrap: break-word;)就会造成一个单词割裂。所以建议大佬将ls这个命令单独弄一下
| gharchive/issue | 2022-10-20T08:36:13 | 2025-04-01T06:46:06.135146 | {
"authors": [
"18030705033",
"tzfun"
],
"repo": "tzfun/vue-web-terminal",
"url": "https://github.com/tzfun/vue-web-terminal/issues/7",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
809913825 | Is it Hardcoded to have maximum around 1K connections
I have stress tested it and Results are somewhat good. It is definitely faster than nodeJS (connected in 0-1ms where node taked 2-3 ms). Apart from that NodeJS WS library is able to sustain 20K connections easily. Where uWS is only able to keep 1014 connections open. It seems like some hard-coded limit.
Also when I press Ctrl+C on Nodejs terminal to stop the client, the uWS server suddenly connects to about 500 more. However, one thing to note is that these same numbers can be seen every time. Its not like in 2nd or third test i get different numbers!, Its always same around 1K connections and then the server stalls FOREVER and then suddenly jumps to 1500 when I stop the nodejs client(1K and 500, every single time).
This behaviour can be reproduced -
uWS server code -
/* We simply call the root header file "App.h", giving you uWS::App and uWS::SSLApp */
#include "App.h"
#include <iostream>
using namespace std;
/* This is a simple WebSocket echo server example.
* You may compile it with "WITH_OPENSSL=1 make" or with "make" */
int main() {
/* ws->getUserData returns one of these */
struct PerSocketData {
/* Fill with user data */
int id;
};
int conns = 0;
/* Keep in mind that uWS::SSLApp({options}) is the same as uWS::App() when compiled without SSL support.
* You may swap to using uWS:App() if you don't need SSL */
uWS::App().ws<PerSocketData>("/*", {
/* Settings */
.compression = uWS::SHARED_COMPRESSOR,
.maxPayloadLength = 16 * 1024,
.idleTimeout = 16,
.maxBackpressure = 1 * 1024 * 1024,
.closeOnBackpressureLimit = false,
.resetIdleTimeoutOnSend = false,
.sendPingsAutomatically = true,
/* Handlers */
.upgrade = nullptr,
.open = [&conns](auto *ws) {
/* Open event here, you may access ws->getUserData() which points to a PerSocketData struct */
struct PerSocketData *data = (struct PerSocketData*) ws->getUserData();
data->id = ++conns;
cout << "Connected #" << (data->id) << endl;
/* For Ending Connection - ws->end(); */
},
.message = [](auto *ws, std::string_view message, uWS::OpCode opCode) {
ws->send(message, opCode, true);
},
.drain = [](auto */*ws*/) {
/* Check ws->getBufferedAmount() here */
},
.ping = [](auto */*ws*/) {
/* Not implemented yet */
},
.pong = [](auto */*ws*/) {
/* Not implemented yet */
},
.close = [](auto */*ws*/, int /*code*/, std::string_view /*message*/) {
/* You may access ws->getUserData() here */
}
}).listen(9001, [](auto *listen_socket) {
if (listen_socket) {
std::cout << "Listening on port " << 9001 << std::endl;
}
}).run();
}
WS client code -
const WebSocket = require('ws');
const host = "127.0.0.1";
const port = 9001;
const count = 20000;
var i=0;
let interval = setInterval(() => {
if(i == count) {
clearInterval(interval);
return;
}
const ws = new WebSocket('ws://' + host + ":" + port);
console.log("Connecting #" + (i+1));
ws.id = i+1
ws.start = (new Date()).getTime();
ws.on('open', function open() {
ws.end = (new Date()).getTime();
var diff = ws.end - ws.start
console.log("Time taken for #" + ws.id + " - " + diff + " ms");
});
ws.on('message', function incoming(data) {
console.log(data);
});
ws.on('error', () => {
console.log("Hung Up - #" + ws.id);
i--;
});
ws.on('close', () => {
console.log("Closed #" + ws.id);
})
i++;
}, 5);
How can I have 20K concurrent connections (which is supported by default on node WS, that too single-threaded).
https://github.com/uNetworking/uWebSockets.js/issues/457
| gharchive/issue | 2021-02-17T06:54:48 | 2025-04-01T06:46:06.240970 | {
"authors": [
"alexhultman",
"thomasb892"
],
"repo": "uNetworking/uWebSockets",
"url": "https://github.com/uNetworking/uWebSockets/issues/1198",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
390918818 | [WIP] Issue #1409: [WIP] support portal.allServerNames property in a safe way
Resolves #1409
Checklist
[x] the individual contributor license agreement is signed
[x] commit message follows commit guidelines
Description of change
I could really use your help on this one @jgribonvald. ;-)
This PR is another take on issue #1409 and pull request #1410.
The key difference in this PR -- and the essence of the issue, I think -- is that I feel we should avoid having the token _CURRENT_SERVER_NAME_ (or any such token) inside a Spring-managed property.
The token replacement tech is 100% custom. Spring doesn't know about it, so it faithfully inserts the original property value (incl. the token) everywhere the property is referenced.
All of the following have multiple uses in the portal:
portal.protocol.server.context
portal.login.url
org.apereo.portal.channels.CLogin.CasLoginUrl
Putting _CURRENT_SERVER_NAME_ into any of them (inside security.properties) will result in the unprocessed token getting used somewhere in the portal.
This PR updates the UrlMultiServerNameCustomizer, attempting to parse the URL being processed into its bits (host, port, path, etc.) and replace some of those bits in the intended way .
It's not entirely working ATM. All the tests (3) are broken (CasLoginRefUrlEncoderTest). One seems to be only an encoding issue; the other two don't get the hostname they expect.
I understand the problem but I'm against providing a default secure value as anybody won't be able to see/obtain a problem when a change will be done, or we should find a way to test it correctly with a CAS server on all side/request it could be needed.
I provided the feature on this way to provide a real view of the use and to be able to check if it work on all change case.
Also I can assume some effort to keep it working, just provide me an issue, the use case, and condition to reproduce the problem - the best would be to provide a unit test to reproduce the problem. And on case that I didn't manage something like OIDC we should talk of what is desired and I can tell how to solve it !
Folks,
In my case, I think I could be a lot more effective contributing to a good solution if I understood the new feature better.
Help me remember/understand more thoroughly: what is the problem the _CURRENT_SERVER_NAME_ change is trying to solve?
We need one portal deployment to support authentication with multiple CAS deployments (I hope it's this one)
We need one CAS deployment to support authentication with multiple portal deployments (probably not this one... this one is easy)
We need one portal deployment to operate as several different hostnames (I hope it's not this one; this one sounds hard)
The portal.server and portal.protocol.server.context properties are used in a lot of places, and some adopters use them in additional ways that aren't currently configured in uPortal-start. I would like to keep these properties simple... 100% managed by Spring, and predictable.
From looking around yesterday, it appears that the important properties for this feature are org.apereo.portal.channels.CLogin.CasLoginUrl and logout.redirect. They both include a query string parameter containing a URL that points back to the portal. They aren't used very often. I think we could decorate these properties with beans that contain logic to rewrite them in custom ways.
Drew the _CURRENT_SERVER_NAME_ permit to manage several public uPortal hostname on the same uPortal instance (don't need more than that to work). And it's working really well on my case. If you need we can talk more about it on hangout.
This permit to set the current url from current user request into CAS serviceValidate URL into the service url param. It's important to get back the user to the good hostname whereas the session cookie won't be on the good hostname...
So this doesn't permit to manage several CAS server in CAS authentications, for that you should have a WAYF like page to ask the user "where to authenticate". For more details we can talk about what I did (also I should work on CAS server for a big upgrade when back from christmas hollydays)
Also this is the proff that I should provide a documentation on how it's working and on how to use it. And surely present the use case !
#2525 fix the last remaining problem from OIDC. Like that the OIDC server can be multi-tenant URL (several server name on tenant)
@jgribonvald should we drop this PR now?
@bjagg yes you can as all was merged
| gharchive/pull-request | 2018-12-14T00:35:33 | 2025-04-01T06:46:06.259281 | {
"authors": [
"bjagg",
"drewwills",
"jgribonvald"
],
"repo": "uPortal-Project/uPortal",
"url": "https://github.com/uPortal-Project/uPortal/pull/1470",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
85590635 | Changed copypaste-socks-chromeapp to listen to all local interfaces.
Useful for making it work in docker images.
:+1:
| gharchive/pull-request | 2015-06-05T17:44:20 | 2025-04-01T06:46:06.261247 | {
"authors": [
"bemasc",
"mollyling"
],
"repo": "uProxy/uproxy-lib",
"url": "https://github.com/uProxy/uproxy-lib/pull/192",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
118476416 | Quiver: Enforce that user names are not empty
The Quiver login UX should enforce that the user's name is not empty. We should make sure this works in all cases:
Login from network selection screen (invite-user.html)
Login from settings drawer
Login from navigating to a Quiver invite URL when not yet logged into Quiver
Fixed in #2098
| gharchive/issue | 2015-11-23T21:28:44 | 2025-04-01T06:46:06.262754 | {
"authors": [
"dborkan"
],
"repo": "uProxy/uproxy",
"url": "https://github.com/uProxy/uproxy/issues/2080",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1909592346 | Adds permafrost and degree day variable examples for ArcticEDS
This PR adds the "CSV preview" data for the ArcticEDS endpoints of permafrost and degree days (heating, freezing_index, and thawing_index).
To test, use a variety of different latitudes and longitudes in the /eds/all endpoint such as:
http://localhost:5000/eds/all/65/-156
You can also test the new endpoints by themselves which are:
Permafrost
http://localhost:5000/eds/permafrost/point/65/-156
Heating Degree Days
http://localhost:5000/eds/degree_days/heating/65/-156/preview
Freezing Index
http://localhost:5000/eds/degree_days/freezing_index/65/-156/preview
Thawing Index
http://localhost:5000/eds/degree_days/thawing_index/65/-156/preview
I tested this branch and all of the ...eds/permafrost... and ...eds/degree days... routes above return the expected data.
I am wondering why the route for permafrost does not include the /preview suffix in the route string?
| gharchive/pull-request | 2023-09-22T22:39:30 | 2025-04-01T06:46:06.266950 | {
"authors": [
"BobTorgerson",
"Joshdpaul"
],
"repo": "ua-snap/data-api",
"url": "https://github.com/ua-snap/data-api/pull/346",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1624628298 | 🛑 glpi is down
In 4037cdb, glpi (https://glpi.sbd.uab.cat/index.php?noAUTO=1) was down:
HTTP code: 0
Response time: 0 ms
Resolved: glpi is back up in f696cd5.
| gharchive/issue | 2023-03-15T02:46:27 | 2025-04-01T06:46:06.269424 | {
"authors": [
"JordiRoman"
],
"repo": "uab-dtic/sbd-uab-monitoring",
"url": "https://github.com/uab-dtic/sbd-uab-monitoring/issues/121",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
91253430 | Question: Support for Swift 2.0
Hello - was just wondering if PageMenu is (or will be) supported for Swift 2.0 (given Xcode 7 beta 2 is out)?
There isn't that much to port to Swift 2.0, I just did it. The larger issue is that it complete crashes Xcode 7 Beta 5 =) Dunno if that's Xcode's problem or not though
+1
It is supported (#84) but it seems that @fahlout didn't released it on CocoaPods. In the meantime, you can install the class file required for PageMenu manually and it'll work with Swift 2.0.
| gharchive/issue | 2015-06-26T13:35:31 | 2025-04-01T06:46:06.271231 | {
"authors": [
"aventurella",
"nitinhayaran",
"oxidec",
"projectwakii"
],
"repo": "uacaps/PageMenu",
"url": "https://github.com/uacaps/PageMenu/issues/66",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
479876792 | CI: Switching Between Github.com and Github Enterprise Usernames
Issue: Switching to the Github Enterprise instances requires a set of usernames that do not work both on Github.com and Github Enterprise.
The usernames have been switched from Github.com to Github Enterprise usernames.
Possible Solution: Depending on the API format configured in the .env of the GH_API property, we can switch between sets of usernames that work on Github.com and Github Enterprise respectively.
This possible solution must work for both TestHarness.ts and GithubActions.ts files:
UPDATE: More username references that are Github.com specific broke autotest application tests on Github Enterprise.
See diff between master from 076cfc5aa764a2572103417d2a1e363c102de6de for these files:
packages/autotest/test/githubEvents/comment_master_bot_one-deliv.json
packages/autotest/test/githubEvents/comment_master_bot_two-deliv.json
packages/autotest/test/githubEvents/comment_other-branch_bot_one-deliv.json
packages/autotest/test/ClassPortalSpec.ts
Update: More hardwired usesrnames exist in autotest. See diff and master on 076cfc5aa764a2572103417d2a1e363c102de6de for differences under files:
packages/autotest/test/GitHubEventSpec.ts,
packages/autotest/test/GitHubServiceSpec.ts
Solution: Github usernames were re-used between Github.com and Github Enterprise instances. As we control the LDAP username integrations, we created test accounts and re-created these usernames on Github.com.
The application can switch between Github integrations without having to update the Github usernames. One may update the Github usernames if they wish in the .env file, but this requires quite a lot of manual work.
Released on branch: feature/circle-ci-github-dev-make-script-new-env
| gharchive/issue | 2019-08-12T22:22:35 | 2025-04-01T06:46:06.295677 | {
"authors": [
"andrewstec"
],
"repo": "ubccpsctech/classy",
"url": "https://github.com/ubccpsctech/classy/issues/35",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
177462372 | depreciating authors.rst in favour of contributors.json
In order to track and acknowledge contributions to em.geosci.xyz, we have set up the contributors page (http://em.geosci.xyz/contributors.html). This is auto-built from the contributors.json file in the root of the em directory. This will allow us to acknowledge contributions throughout the site and have it properly linked back to information on you!
The structure is
{
"lheagy":{
"name":"Lindsey J. Heagy",
"affiliation":"University of British Columbia",
"location":"Vancouver, BC",
"email":"lheagy@eos.ubc.ca",
"url":"http://lindseyjh.ca",
"avatar":"https://avatars.githubusercontent.com/u/6361812?v=3",
"ORCID":"0000-0002-1551-5926"
}
}
A couple things to note:
the url and ORCID (http://orcid.org/) are optional.
the avatar can be a weblink to a photo or you can include a photo locally and put the file path there. The photo needs to be square, so please crop it. Please start an "images" directory in the root em if you would like to do this. Using a file path has not been rigorously tested, so let me know if you run into issues.
I would like to depreciate the AUTHORS.rst by the end of next week (Sept 23), so please add yourself to the contributors.json! If you run into any issues, or are unclear on how to do this, feel free to send me your info and I am happy to help out.
[x] @dougoldenburg
[x] @micmitch
[ ] @mikemcm
[x] @sgkang
[x] @fourndo
[ ] @dccowan
[ ] @lacmajedrez
[ ] @prhjiajie
[ ] @krisdavis
[ ] @DanielBild-Enkin
[ ] @dccowan
[x] @Pbellive
[ ] @sannatti
[ ] @grosenkj
[ ] @rowanc1
also @thast : you should add a picture :) (could use: http://simpeg.xyz/s/img/people/thibaut.jpg)
@ubcgif/developers : I will be removing the "AUTHORS.rst" at the end of the week - please add yourselves to "contributors.json"
| gharchive/issue | 2016-09-16T16:04:41 | 2025-04-01T06:46:06.304207 | {
"authors": [
"lheagy"
],
"repo": "ubcgif/em",
"url": "https://github.com/ubcgif/em/issues/302",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2079970929 | Unreachable when writing a square bracket and a list
In text mode, if a list is written after a square bracket without escaping, an unreachable error is displayed during parsing.
The list does not have to be immediately after the square bracket, and it seems to reproduce even when placed across paragraphs and apart.
Example code:
[foo] bar
baz
- foo
Result:
unreachable src/scanner.c:237
(source_file [0, 0] - [5, 0]
(text [0, 0] - [0, 1])
(text [0, 1] - [0, 4])
(text [0, 4] - [0, 5])
(text [0, 6] - [0, 9])
(parbreak [0, 9] - [2, 0])
(text [2, 0] - [2, 3])
(parbreak [2, 3] - [4, 0])
(item [4, 0] - [4, 5]
(text [4, 2] - [4, 5])))
Since installing tree-sitter-typst, my Typst environment has become much more comfortable. Thank you so much!
I found the bug and fixed it.
Explanation:
Plain text square brackets, as they have an effect on the semantic:
#strong[Hello [] World]
#strong[Hello ] World]
They are considered as containers, and as they do not effect the indentation, no indentation level is pushed on the indentation stack. But when the ] was encountered, the indentation stack was pop of one level. I removed this pop.
I tried ecf8596 it in my environment and confirmed that the problem has been resolved. Thank you for your quick response and fix!
| gharchive/issue | 2024-01-13T01:40:55 | 2025-04-01T06:46:06.309804 | {
"authors": [
"monaqa",
"uben0"
],
"repo": "uben0/tree-sitter-typst",
"url": "https://github.com/uben0/tree-sitter-typst/issues/18",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
138861817 | update math docs
@ibgreen @mikolalysenko
Vast majority of these get rid of the LumaGL.* syntax in favor of accessing the naked classes.
Couple of minor wording improvements and removed PhiloGL release notes.
Wonderful. We're making progress!
| gharchive/pull-request | 2016-03-07T01:43:09 | 2025-04-01T06:46:06.311130 | {
"authors": [
"ibgreen",
"wwwtyro"
],
"repo": "uber-common/luma.gl",
"url": "https://github.com/uber-common/luma.gl/pull/52",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
517356429 | Duplicated decision due to same chilldWorkflowID
Cadence client use workflowID as decisionID, so decision was rejected because of duplication. This is mostly OK but in case people want to rerun childWF with same ID they will hit an error of decisionTaskFailed.
Solution is to use alternatives for decisionID for childWorkflow, like incremental ID with randomUUID in Java client.
This is not a valid use case: workflow code can take use of ChildWorkflow future to make sure that they don't need to start same child workflow if the previous one is still running.
| gharchive/issue | 2019-11-04T19:34:40 | 2025-04-01T06:46:06.312586 | {
"authors": [
"longquanzheng"
],
"repo": "uber-go/cadence-client",
"url": "https://github.com/uber-go/cadence-client/issues/873",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
447920371 | PR title syncing
For public → private PR syncing, including the actual PR title (for example, Add foobar (via public/repo#123)) rather than just Sync public/repo#123 would make things more understandable I think. It's hard to remember PR numbers.
Will be fixed after #10 merges
| gharchive/issue | 2019-05-23T23:52:02 | 2025-04-01T06:46:06.313712 | {
"authors": [
"chrisdothtml",
"rtsao"
],
"repo": "uber-workflow/probot-app-monorepo-sync",
"url": "https://github.com/uber-workflow/probot-app-monorepo-sync/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
437440510 | Sharding
implement sharding
add kafka sink
Codecov Report
Merging #146 into master will decrease coverage by 0.12%.
The diff coverage is 66.93%.
@@ Coverage Diff @@
## master #146 +/- ##
==========================================
- Coverage 73.21% 73.08% -0.13%
==========================================
Files 106 108 +2
Lines 16623 16820 +197
==========================================
+ Hits 12170 12293 +123
- Misses 3529 3591 +62
- Partials 924 936 +12
Impacted Files
Coverage Δ
client/connector.go
71.16% <0%> (-1.7%)
:arrow_down:
subscriber/common/job/controller.go
43.88% <100%> (+0.71%)
:arrow_up:
subscriber/common/job/driver.go
66.19% <100%> (+0.12%)
:arrow_up:
subscriber/config/service_config.go
83.33% <100%> (+1.51%)
:arrow_up:
subscriber/common/job/retry_failure_handler.go
82.5% <100%> (ø)
:arrow_up:
memstore/live_vector_party.go
70.05% <33.33%> (ø)
:arrow_up:
subscriber/common/message/message_parser.go
50.52% <35.71%> (-16.69%)
:arrow_down:
subscriber/common/rules/job_config.go
70.86% <60%> (-2.48%)
:arrow_down:
subscriber/common/sink/ares_database.go
56.75% <60%> (ø)
subscriber/common/sink/sink.go
74.35% <74.35%> (ø)
... and 5 more
Continue to review full report at Codecov.
Legend - Click here to learn more
Δ = absolute <relative> (impact), ø = not affected, ? = missing data
Powered by Codecov. Last update 1839c86...6e72d4a. Read the comment docs.
| gharchive/pull-request | 2019-04-25T23:41:00 | 2025-04-01T06:46:06.328255 | {
"authors": [
"codecov-io",
"lxning"
],
"repo": "uber/aresdb",
"url": "https://github.com/uber/aresdb/pull/146",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2330406841 | Add typescript types tests runner
Add jest-runner-tsd for testing types.
The library is using https://github.com/mrazauskas/tsd-lite internally which is lately be archived.
The owner moved to build a newer API(TSTyche) that has its v2 in beta release, so we can keep an eye on that repo and migrate to it once becoming stable.
@Assem-Hafez Glad to see you are interested in TSTyche. I am its author. Just wanted to let you know that stable version 2 is already released. It is definitely worth using TSTyche instead of tsd-lite. Please, ping me if you have any questions.
| gharchive/pull-request | 2024-06-03T07:40:44 | 2025-04-01T06:46:06.330823 | {
"authors": [
"Assem-Hafez",
"mrazauskas"
],
"repo": "uber/cadence-web",
"url": "https://github.com/uber/cadence-web/pull/587",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
804492875 | Drop pointer for task id fields
What changed?
Why?
How did you test it?
Potential risks
Coverage increased (+0.05%) to 64.657% when pulling 8a6b4451553e48d17e75032426e3181ec1a6cb6f on vytautas-karpavicius:drop-ptr-task-id into 5ffd942d386e80e9e26820149914353e181303a0 on uber:master.
| gharchive/pull-request | 2021-02-09T12:08:19 | 2025-04-01T06:46:06.333606 | {
"authors": [
"coveralls",
"vytautas-karpavicius"
],
"repo": "uber/cadence",
"url": "https://github.com/uber/cadence/pull/3969",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
427935202 | Fix layer demos not update parameters
Background
Embed demos do not pick up the parameters from user because of a recent change.
https://github.com/uber/deck.gl/pull/2814/
Change List
remove unapplied H3 links
fix parameters not updated in embed demos
Coverage remained the same at 61.433% when pulling 4ec3118fbacc879cde222b3399a204b7d4d589ba on xx/website-demos into 8f8d5e60bb375ae2d05502736945c418b96f18a8 on master.
Coverage remained the same at 61.433% when pulling 4ec3118fbacc879cde222b3399a204b7d4d589ba on xx/website-demos into 8f8d5e60bb375ae2d05502736945c418b96f18a8 on master.
| gharchive/pull-request | 2019-04-01T22:14:00 | 2025-04-01T06:46:06.337395 | {
"authors": [
"coveralls",
"xintongxia"
],
"repo": "uber/deck.gl",
"url": "https://github.com/uber/deck.gl/pull/2877",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
177313009 | Upgrade to the new K/V Span logging API
https://github.com/opentracing/opentracing-go/pull/108
TODO:
[x] restore skipped test
[x] add more tests for new methods
[x] finalize conversion of log tags into zipkin annotation
[x] wait for official API release
Coverage decreased (-1.6%) to 74.79% when pulling 9a1588121652ed1503b38241a96cd59eeae17f72 on spanlog into fd2740f0c618951e96940386b2fcc4ed84bfd941 on master.
Coverage decreased (-1.5%) to 74.871% when pulling b0fdc9c41532047aae50c324095c83b9216c1ca7 on spanlog into fd2740f0c618951e96940386b2fcc4ed84bfd941 on master.
Coverage decreased (-1.5%) to 74.871% when pulling b0fdc9c41532047aae50c324095c83b9216c1ca7 on spanlog into fd2740f0c618951e96940386b2fcc4ed84bfd941 on master.
Coverage decreased (-1.5%) to 74.871% when pulling ba42bb6a6e41da82b3bd62e4627daee2fa64bc60 on spanlog into fd2740f0c618951e96940386b2fcc4ed84bfd941 on master.
we should probably add to the zipkin README a big caution on encoding json, as it will screw up other tooling, such as datastore-layer queries which are used in the zipkin ecosystem. Here's a note about this https://github.com/openzipkin/openzipkin.github.io/issues/52#issuecomment-249749327
Another option is to open a hook for a formatter, similar to most logging libraries, such that people can avoid this problem.
Coverage decreased (-1.5%) to 74.918% when pulling ac445299da1997b41123248ed6da9e33426b4b00 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage decreased (-1.5%) to 74.918% when pulling ac445299da1997b41123248ed6da9e33426b4b00 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
@adriancole the primary reason we're doing JSON for logs is because our new data model actually supports KV logs natively and we have a reverse transform zipkin->jaeger for the new UI and want that transform to be lossless.
Coverage decreased (-0.6%) to 75.867% when pulling 4dbaa2efe24535eae112769169dd887f18ddd8ec on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage decreased (-0.6%) to 75.867% when pulling 4dbaa2efe24535eae112769169dd887f18ddd8ec on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage increased (+0.05%) to 76.476% when pulling 9f3d30764b11843b840125a1ed28913db6eb44a8 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage increased (+0.05%) to 76.476% when pulling 9f3d30764b11843b840125a1ed28913db6eb44a8 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage increased (+0.05%) to 76.476% when pulling 8f5a3b30ac4fe7998fc4be69fc0af0e2a8901a46 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage increased (+0.05%) to 76.476% when pulling 8f5a3b30ac4fe7998fc4be69fc0af0e2a8901a46 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage increased (+0.05%) to 76.476% when pulling 253eb68c13527e177556285a8de1690b172665e5 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
Coverage increased (+0.05%) to 76.476% when pulling 253eb68c13527e177556285a8de1690b172665e5 on spanlog into b45a5ba5a3fcbb67aa996cd9b10c11580a7f5479 on master.
| gharchive/pull-request | 2016-09-15T23:15:56 | 2025-04-01T06:46:06.356194 | {
"authors": [
"adriancole",
"coveralls",
"yurishkuro"
],
"repo": "uber/jaeger-client-go",
"url": "https://github.com/uber/jaeger-client-go/pull/42",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
423118504 | Is there a way to force 24 hours time format in time scale?
I'm using XYPlot with xType={'time'} and Russian locale as described in documentation, is there a way to pass 24h time format to d3-scale?
Instead of "03 AM" I need something like "03:00".
I am also interested in the answer to this
I'd recomment using MomentJS
or just Date().parse() https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Date/parse
I deal with the same issue. The change of the locale does not switch to 24-hour format. Also the usage of tickFormat doesn't give me the flexible time display on XAxis according to number of values.
I gave up on react-vis. It is too buggy, not enough features. I ended up going to recharts but now I'm thinking about chartJS.
| gharchive/issue | 2019-03-20T08:10:58 | 2025-04-01T06:46:06.369481 | {
"authors": [
"Falieson",
"hodn",
"paddotk",
"yojeek"
],
"repo": "uber/react-vis",
"url": "https://github.com/uber/react-vis/issues/1129",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
232448191 | Cannot Cheapest Product
I'm using your uber SDK on iOS app. Last days, I can get cheapest product with ridesClient.fetchCheapestProduct() function. But today, that function always return nil value.
Any problem with your server or sdk. Please help me about this issue
(I'm in Vietnam and my client_id still works fine for Android)
Thanks
Hey @nlhdung, I'm unable to reproduce this bug. Can you please give me more details about how you're using the SDK?
Are you getting an error in the response object? If so, are you adding your Server Token properly?
See here: https://github.com/uber/rides-ios-sdk#sdk-configuration
Since this is question is from over a month ago, I'm going to close this until further response :)
| gharchive/issue | 2017-05-31T04:10:23 | 2025-04-01T06:46:06.372374 | {
"authors": [
"edjiang",
"nlhdung"
],
"repo": "uber/rides-ios-sdk",
"url": "https://github.com/uber/rides-ios-sdk/issues/164",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
286849797 | Multiple DC replication strategy
We are trying to run ureplicator across 4 dc's and currently neither mirrormaker nor ureplicator afaik supports replicating everywhere in a way that we don't have duplicate message or infinite chain.
In order to support such a functionality, we want to add a custom consumer, which reads into the message, finds a "dc" field and if it's different from one in the config of ureplicator then it will replicate.
Do you think this is something that we can contribute back to ureplicator or how have you solved this issue ?
cc @drdee to get your opinions on this
Hi, @yagnik , we are also facing the situation which required replication filtering and we found a simple solution to filter this in producer (https://github.com/uber/uReplicator/pull/132/files). It's quite difficult to filter the message on the consumer side which caused a lot of memory and cpu overhead during our test. Lemme know if you have any thoughts or suggestion.
| gharchive/issue | 2018-01-08T18:58:19 | 2025-04-01T06:46:06.374593 | {
"authors": [
"xhl1988",
"yagnik"
],
"repo": "uber/uReplicator",
"url": "https://github.com/uber/uReplicator/issues/105",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
449417556 | Change use of 'Frame' to 'Message' in XVIZ
The term 'Frame' is overloaded, and in nearly all case
in XVIZ we mean an XVIZ Message. We have created a distinction
where if we use 'Frame' it is not an individual message but the
set of data that constitutes a visual 'Frame' which could come from
multiple messages. Therefore, 'Frame' is still used in @xviz/parse
This change includes the following:
Changes to @xviz/builder, specifically getFrame() to getMessage()
getFrame() is marked as deprecated in the source code and removed from docs
@xviz/io changes all names using Frame to use Message
@xviz/io shares json binary packing with @xviz/builder
Documentation has been updated
Examples conversions have been changed from convertFrame() to convertMessage()
example/server has not been changed
@xviz/parser has not been changed
Coverage decreased (-0.1%) to 75.911% when pulling 4f63fa590772d53aad2d7d5d4e7bb69b88d7b1f0 on tpw/frame-to-message into 51e6c9f9bef08ef613b2359e4f67d8c6fd8f6186 on master.
| gharchive/pull-request | 2019-05-28T18:36:39 | 2025-04-01T06:46:06.379168 | {
"authors": [
"coveralls",
"twojtasz"
],
"repo": "uber/xviz",
"url": "https://github.com/uber/xviz/pull/467",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
120173306 | Item (Magento\Catalog\Model\Product\Interceptor) with the same id "231" already exist
Hi guys,
i am using the new version Magento2 Data Migration version 0.1.21. After the successfull import i get an error in my product management. See attached. Is there any Idea why.
thanks for help
https://github.com/ubertheme/magento2_data_migration/issues/13
@derdeakon:
I need a closer look to detect the root of issue as you said.
Can you share me a back-up of your Magento 1 database?
Regards,
quynhvv
Yes i can share you my database.
but i can not share my database in this ticket. It is a Zip an doesn't work here
@derdeakon
Use dropbox to share ;)
https://www.dropbox.com/s/phyaiumju50x2cb/database.sql?dl=0
@derdeakon: What version of Magento 1 for the database you are using for that database?
@derdeakon :
Item (Magento\Catalog\Model\Product\Interceptor) with the same id "231" already exist
Thank for the database you have shared.
I have detected the root of issue as you said in step 5. There is a issue with cataloginventory_stock_item table after data migrated in case we have multiple websites.
To solve the issue you can do:
Open the file at /protected/controllers/MigrateController.php
and find to the block code:
//cataloginventory_stock_item
$stock_item = Mage1StockItem::model()->find("product_id = {$model->product_id} AND stock_id = {$model->stock_id}");
if ($stock_item){
$stock_item2 = new Mage2StockItem();
foreach ($stock_item2->attributes as $key => $value){
if ($key != 'item_id' && isset($stock_item->$key)){
$stock_item2->$key = $stock_item->$key;
}
}
$stock_item2->website_id = $website_id;
$stock_item2->save();
}
and replace it by the block code:
//cataloginventory_stock_item
$stock_item2 = Mage2StockItem::model()->find("product_id = {$model->product_id} AND stock_id = {$model->stock_id}");
if (!$stock_item2){
$stock_item = Mage1StockItem::model()->find("product_id = {$model->product_id} AND stock_id = {$model->stock_id}");
if ($stock_item){
$stock_item2 = new Mage2StockItem();
foreach ($stock_item2->attributes as $key => $value){
if ($key != 'item_id' && isset($stock_item->$key)){
$stock_item2->$key = $stock_item->$key;
}
}
$stock_item2->website_id = $website_id;
$stock_item2->save();
}
}
and make reset the step 5
and re-start with data migration on step 5.
Let's do it to solve the issue and tell me know how it goes.
Regards,
quynhvv.
Hi quynhvv,
it works fine. Thanks you. After reindex and cache my products will shown in the backend and in the correct category in the frontend. :)
@derdeakon :
Yes, and please note that:
We have just released one Magento 2 modules for this at:
https://github.com/ubertheme/module-ubdatamigration
Regards,
quynhvv
| gharchive/issue | 2015-12-03T13:40:14 | 2025-04-01T06:46:06.392425 | {
"authors": [
"derdeakon",
"quienti",
"quynhvv"
],
"repo": "ubertheme/magento2_data_migration",
"url": "https://github.com/ubertheme/magento2_data_migration/issues/15",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
51074889 | Multiple markers?
Is it possible to display multiple markers on a single map? In this case, I'm not talking about a form page or autocompletion, just showing a map on a page and sending in lat/lon coordinates from the view layer or json (I'm using Rails).
<div class="map_canvas_form"></div>
<span class="hidden latitude"><%= job.latitude %></span>
<span class="hidden longitude"><%= job.longitude %></span>
setPanelLocation = if !!$('.latitude').html() then [$('.latitude').html(), $('.longitude').html()] else ""
$("div.map_canvas").geocomplete
location: setPanelLocation
map: "div.map_canvas"
Can I pass an array or JSON object of lat/lon coordinates to location?
Hello, does somebody found solution for this problem?
I also need to use multiple markers on the geocomplete map.
Thanks
| gharchive/issue | 2014-12-05T08:37:34 | 2025-04-01T06:46:06.394705 | {
"authors": [
"Ganasist",
"hivearts"
],
"repo": "ubilabs/geocomplete",
"url": "https://github.com/ubilabs/geocomplete/issues/176",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
214208349 | Converting input to an suggest object on button press
I have an app that allows people to begin typing in an address to start looking for locations. Currently if you start typing and click one of the suggestions or press enter, it performs a search and
the onSuggestSelect function fires off which returns and object that has the location parameter which contains the lat/lng. This works great.
The issue I'm running into is that I also have a search button that the user can click on. If the user begins typing and doesn't hit enter or choose a suggestion, than the address isn't parsed and an object with location information isn't returned. I'm not sure why it would work by submitting the form with enter vs pressing a submit button.
I tried getting around by setting the autoActivateFirstSuggest to true and grabbing the first suggestion based on the input into the textbox, but unforunately all this returns is a label and not the location object (lat/lng).
In short:
Type in a zip code
Don't choose a selection or hit enter
Use a submit button to submit the form
An object isn't returned, but the plain text in the box.
The goal is to allow users to quickly type in their zip codes and hit the submit button, but that doesn't work at the moment.
What could a solution to this be? We can’t hook into the submit of a parent form…
Closing this as there is no feedback. Reopen if needed.
| gharchive/issue | 2017-03-14T21:13:07 | 2025-04-01T06:46:06.397905 | {
"authors": [
"mskwarcan",
"ro-ka"
],
"repo": "ubilabs/react-geosuggest",
"url": "https://github.com/ubilabs/react-geosuggest/issues/284",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2303429178 | Other refresh rates no longer available in game mode on GPD Win 4
Describe the bug
60 Hz is the only refresh rate available to select.
What did you expect to happen?
Should have the full range of 40 - 60 Hz refresh rates available, as it had been before.
Output of rpm-ostree status
State: idle
Deployments:
● ostree-image-signed:docker://ghcr.io/ublue-os/bazzite-deck:stable
Digest: sha256:f15f2dfeb61a526aab8a3e7c6072699e84e2eaca5c2932fb725ba778a243daef
Version: 40.20240517.0 (2024-05-17T17:08:59Z)
LocalPackages: lact-0.5.4-0.x86_64
ostree-image-signed:docker://ghcr.io/ublue-os/bazzite-deck:stable
Digest: sha256:04a49156c4a01e10f35b9833aae13caaa50dd9c40493ab1cefd8325d82b5ceb6
Version: 40.20240514.0 (2024-05-14T17:07:51Z)
LocalPackages: lact-0.5.4-0.x86_64
Hardware
G1618-04
GPD Win 4 6800U version
Extra information or context
After some update, not sure how long ago, the toggle to separate frame rate limits and refresh rate control disappeared, but I didn't realize until just recently when I tried selecting the option to disable the frame rate that the slider changes to refresh rate control instead. Then I realized that 60 Hz was the only refresh rate available.
Checking the device quirks script, it still has the line to set the environment variable to make 40 - 60 Hz be available, so not sure what's broken specifically to cause this. It seemed to coincide with a Steam client update in some recent week, but I'm not certain.
I read elsewhere that this happened due to a gamescope update, and that version 40-20240427 was the last one before that update.
| gharchive/issue | 2024-05-17T19:27:06 | 2025-04-01T06:46:06.403272 | {
"authors": [
"ShadowFlare"
],
"repo": "ublue-os/bazzite",
"url": "https://github.com/ublue-os/bazzite/issues/1144",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2399712225 | Reinstall Bazzite leads to error during « 'ostree container image deploy »
Describe the bug
Hi !
I’m trying to reinstall Bazzite on the same hardware and re-installation always gives me this error
The command 'ostree container image deploy --sysroot=/mnt/sysimage --image=frun/install/repo/bazzite-deck-stable --transport=oci--no-signature-verification' exited with the code 1,
I’ve tried multiple things without success
Followed the troubleshooting guide for code 1 https://universal-blue.discourse.group/docs?topic=2495
Install Kinoite and rebase also gives me an error
Swap LAN adapter
Disable secure boot
Use another usb port
completely wipe SSD from Hiren’s boot cd
boot from Fedora 40 and remove partitions
I’m running out of ideas 🤪.
thanks for helping
What did you expect to happen?
Bazzite installs successfully
Output of rpm-ostree status
No response
Hardware
CPU : Intel 13600K
GPU : AMD 7900 XT
DDR5 : 64 Go
Extra information or context
No response
Ok slept on it and have the idea of resetting all BIOS settings to defaut.
That solved the problem.
Here are the settings that was reverted to default (in case someone faces the same issue)
Secure Boot from Disabled to Enabled
iGPU Disabled to Enabled
Memory XMP profile DDR5-6000 to DDR5-4800
I will try to put again DDR5@6000 and iGPU Off to see what happens.
Ok definately a RAM issue, putting back DDR5@6000 makes the system weird/unstable.
I'm marking this issue as closed.
For those having similar issue, please mind to start with stock settings :)
| gharchive/issue | 2024-07-10T04:21:37 | 2025-04-01T06:46:06.409718 | {
"authors": [
"T13nou"
],
"repo": "ublue-os/bazzite",
"url": "https://github.com/ublue-os/bazzite/issues/1346",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2437796722 | DeckHQ screen orientation
Describe the bug
I installed Bazzite on my Steam Deck, which was modified with a DeckHQ screen. When I boot up in the deck image, the installer boots up in portrait mode, 90 degrees anticlockwise (footer on the right). After completing the installation, KDE boots with this same misconfiguration.
What did you expect to happen?
I'd expect the installer and the installation to be in landscape mode, not like this:
Output of rpm-ostree status
State: idle
Deployments:
● ostree-image-signed:docker://ghcr.io/ublue-os/bazzite-deck:stable
Digest: sha256:12d9af08ffa86b928c7e3f2435fe639bbe14ff0e75cfac6343419c931cdb0fe6
Version: 40.20240728.0 (2024-07-28T21:46:28Z)
LayeredPackages: htop ncdu
ostree-image-signed:docker://ghcr.io/ublue-os/bazzite-deck:stable
Digest: sha256:f4e22243c01f2d2c7448223ed1a81c4671a2a8bcadd6573f959f03044613d6dd
Version: 40.20240728.0 (2024-07-28T20:11:49Z)
LayeredPackages: htop ncdu
Hardware
Jupiter
Extra information or context
I'm now in the state where the boot screen starts up in portrait mode, but I've changed my screen orientation in KDE (which carries through to Steam) to correct the problem. Is there a way to change the screen orientation during the boot process so that I don't need to make the correction in KDE?
Thanks for Bazzite, and keep up the great work!
Thanks for reminding me of this. I'll post the command to rotate the panel by default shortly, but I'll also push a patch upstream to also include DeckHD as a rotated panel by default, so no command will be needed once merged.
Great stuff, I look forward to that command to set things right.
Patch is here: https://patchwork.freedesktop.org/patch/606792/?series=136749&rev=5
I'll get the command to you shortly.
Completely forgpt about this, sorry.
Here it is: sudo rpm-ostree kargs --append-if-missing=video=eDP-1:panel_orientation=right_side_up
Great, thanks for getting back to me on this 🙏
| gharchive/issue | 2024-07-30T13:05:33 | 2025-04-01T06:46:06.415060 | {
"authors": [
"RodoMa92",
"marc0der"
],
"repo": "ublue-os/bazzite",
"url": "https://github.com/ublue-os/bazzite/issues/1433",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1744908279 | feat(dx): move languages off the image
And then replace them with the minimum bare set of deps for enabling homebrew, part of #257
I don't think it's worth trying to figure out how to bring clan back to F37, so I think we should just drop dx 37
Going to hold on this until homebrew lands.
Ok between all the other stuff we have we can go ahead and remove these.
| gharchive/pull-request | 2023-06-07T02:13:34 | 2025-04-01T06:46:06.416623 | {
"authors": [
"castrojo"
],
"repo": "ublue-os/bluefin",
"url": "https://github.com/ublue-os/bluefin/pull/261",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2005309403 | github action.
Does GitHub action work for you for unit test so far? @reshnashrestha
Yes @uarif It is working for my unit test.
We should also lint calculus.py
It is working for us.
okay, thank you
Good luck, let me know if you need help
Thank you
| gharchive/issue | 2023-11-21T23:26:59 | 2025-04-01T06:46:06.454971 | {
"authors": [
"AhmedCode99",
"Pranjal-Srivastava-2023",
"reshnashrestha",
"uarif"
],
"repo": "ubsuny/23-Homework6G3",
"url": "https://github.com/ubsuny/23-Homework6G3/issues/54",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2021522960 | Adding a friend list to user profiles where each friend links to the friend's profile page
This PR adds a button to a user's own profile page that allows them to access a list of their friends. The list of friends contains a list of friend emails, each of which links to a friend's profile page.
Closing comment:
This PR handles the friend list button functionality on a user's profile page. With the Friend List button, users are able to see a list of their friends displayed using a FriendTable. Each friend's email links to that friend's profile page. All assignees worked together on this functionality.
Issue Score: Excellent
Comments:
Great work!
| gharchive/pull-request | 2023-12-01T20:13:16 | 2025-04-01T06:46:06.489102 | {
"authors": [
"24raniwalar",
"elizabethli31"
],
"repo": "uchicago-cs/chigame",
"url": "https://github.com/uchicago-cs/chigame/pull/357",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1929685012 | fix margins for newbuildings energy graph labels
close #621
I am not sure but I think for me the issue is still there?
I didn't manage to perfectly fix it, but it's much better. Maybe you can have another go @johannahaha
| gharchive/pull-request | 2023-10-06T08:30:35 | 2025-04-01T06:46:06.494740 | {
"authors": [
"aeschi",
"johannahaha"
],
"repo": "uclab-potsdam/klimataz",
"url": "https://github.com/uclab-potsdam/klimataz/pull/622",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
220402250 | error in Sentiment Classification Projects.ipynb
Hi, I'm working on Sentiment_Classification_Projects.ipynb.
And I am confused with following code at the end of project 3:
mlp = SentimentNetwork(reviews[:-1000],labels[:-1000], learning_rate=0.1)
It will preprocess the given reviews to fill word2index which is a dict, but when I execute next line:
mlp.test(reviews[-1000:],labels[-1000:])
It throws KeyError: 'reprogram'.
It seems that there are some words only existing in last 1000 reviews, and I think first line should be:
mlp = SentimentNetwork(reviews, labels, learning_rate=0.1)
I'm not sure about it, did I miss something?
That error happens because a test review contains a word that is not in training reviews so that word is not saved into word2index set. You can ignore those unknown words by not setting input vector whenever these words are encountered.
This should probably be fixed in the programming instructions or else it makes it really confusing to debug
For future reference something like this will work :
for word in review.split(' '):
try:
self.layer_0[0,self.word2index[word]] += 1
except:
pass
The following worked for me
indices = set()
for word in review.split(' '):
if word in self.word2index.keys():
indices.add(self.word2index[word])
We will make appropriate changes in the refresh. Thanks :)
| gharchive/issue | 2017-04-08T14:13:34 | 2025-04-01T06:46:06.642553 | {
"authors": [
"CihanSoylu",
"CtheSky",
"LivingProgram",
"abhiojha8",
"nghiattran"
],
"repo": "udacity/deep-learning",
"url": "https://github.com/udacity/deep-learning/issues/66",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1096514914 | Allow multiple-dirs and outputfile as arguments
Allow to add more arguments to the command:
# Set output file
php ./vendor/uderline/openapi-php-attributes/opag app/ public/openapai.json
# Scan multiple folders
php ./vendor/uderline/openapi-php-attributes/opag app/ src/ src2/
# combined
php ./vendor/uderline/openapi-php-attributes/opag app/ src/ src2/ public/openapai.json
Great ! Thanks
| gharchive/pull-request | 2022-01-07T17:23:25 | 2025-04-01T06:46:06.649349 | {
"authors": [
"puggan",
"uderline"
],
"repo": "uderline/openapi-php-attributes",
"url": "https://github.com/uderline/openapi-php-attributes/pull/11",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
830329091 | Ajustar la opción de carga de elementos en registrar acta para que no quede bloqueada después de cada carga
[x] En registro de acta recibido la opción de carga de elementos no debe quedar bloqueada después de que se hace la primera carga.
PR #586
| gharchive/issue | 2021-03-12T17:47:04 | 2025-04-01T06:46:06.650623 | {
"authors": [
"AlexFBP",
"juandavidvelezaalvarez"
],
"repo": "udistrital/arka_cliente",
"url": "https://github.com/udistrital/arka_cliente/issues/564",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
854874720 | mock-ups traslados
Tomar como referencia el sistema de solicitudes que está en Bodega de Consumo (solicitud/consulta/respuesta)
[ ] mock-ups
[ ] revisar posibles ajustes a modelos de datos
Queda pendiente colocar (enlaces a) la documentación derivada de este Issue
https://app.diagrams.net/#G1mwRoyOnt_49JAnvNHg8mHN_-2goQv1Fy
| gharchive/issue | 2021-04-09T21:50:02 | 2025-04-01T06:46:06.652392 | {
"authors": [
"AlexFBP",
"juandavidvelezaalvarez"
],
"repo": "udistrital/arka_cliente",
"url": "https://github.com/udistrital/arka_cliente/issues/604",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2265987393 | Ajuste de nombre de rutas en micro clientes (refactoring) para el cumplimiento de lineamientos
Se requiere realizar el Ajuste de nombre de rutas en micro clientes (refactoring) para el cumplimiento de lineamientos, la issue debe cumplir con los siguientes requerimientos:
Sub Tareas
[x] Realizar el ajuste de los nombres de los path de los micro clientes (componentes) que no cumplan con los lineamientos.
[x] Crear árbol de configuraciones de los nuevos path ajustados en el api de configuraciones en el ambiente de pruebas https://pruebasconfiguracion.portaloas.udistrital.edu.co/ con el usuario de utest01 y asignarlo al rol de ADMIN_SGA
[x] Ajustar la consulta del cliente con arquitectura de micro cliente con el nuevo menú definido
Criterios de aceptación
[x] Ajustar los path del cliente
[x] Crear el árbol de menú
[x] Ajustar el consumo de nuevo menú
Requerimientos
No aplica
Definition of Ready - DoR
[x] Está refinada y estimada en puntos de historia por el equipo.
[x] Incluye la descripción y criterios de aceptación, con el detalle funcional y especificaciones técnicas, de forma entendible por cualquier miembro del equipo.
[x] No tiene bloqueos que impidan su ejecución.
[x] Las dependencias están identificadas y resueltas.
[x] Puede ser probada dentro del Sprint.
Definition of Done - DoD - Desarrollo
[x] Desarrollo en local.
[x] Push en Feature.
[x] Pruebas locales (funcionales).
[x] PR a Develop.
[ ] Criterios de aceptación cumplidos.
[ ] Documentación de issue realizada.
[ ] Aprobada por SM/Líder técnico.
Realizar el ajuste de los nombres de los path de los micro clientes
[ ] sga_cliente_calendario_academico_mf
[ ] sga_cliente_proyecto_curricular_mf
[ ] sga_cliente_inscripcion_mf CORS⚠️
[ ] sga_cliente_admisiones_mf CORS⚠️ cambio de path ANTIGUO: admision NUEVO: admisiones
[ ] sga_cliente_derechos_pecuniarios_mf
[ ] sga_cliente_practicas_academicas_mf
[ ] sga_cliente_datos_mf
[ ] sga_cliente_espacios_academicos_mf CORS⚠️
[ ] sga_cliente_administracion_mf CORS⚠️
[ ] sga_cliente_notas_mf CORS⚠️
[ ] sga_cliente_plan_docente_mf CORS⚠️
[ ] sga_cliente_gestion_horarios_mf CORS⚠️ cambio de path: ANTIGUO: gestion-horarios NUEVO: horarios
PENDIENTE
[ ] sga_cliente_plan_estudios_mf PRESENTA CONFLICTOS
TERMINADO
se revisa la issue se da por finalizado y se despliega en pruebas
| gharchive/issue | 2024-04-26T15:07:51 | 2025-04-01T06:46:06.674774 | {
"authors": [
"fabianbarreto02",
"norbeydanilo",
"redflox"
],
"repo": "udistrital/sga_documentacion",
"url": "https://github.com/udistrital/sga_documentacion/issues/196",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
856353837 | Crear vista de boletin diario
En base a los mockups crear la vista de boletín diario
Se hace merge con develop
| gharchive/issue | 2021-04-12T20:55:32 | 2025-04-01T06:46:06.675751 | {
"authors": [
"Miranda13"
],
"repo": "udistrital/tesoreria_cliente",
"url": "https://github.com/udistrital/tesoreria_cliente/issues/88",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2713470929 | 🛑 Udit Mittal is down
In a4fe4fd, Udit Mittal (https://uditmittal.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Udit Mittal is back up in 03df117 after 4 minutes.
| gharchive/issue | 2024-12-02T21:51:43 | 2025-04-01T06:46:06.678143 | {
"authors": [
"udit-001"
],
"repo": "udit-001/portfolio-status",
"url": "https://github.com/udit-001/portfolio-status/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
270720810 | is Guardian.Plug.current_resource the best place to put it ?
a small design question
In my mind plug should be functions that take a Plug.Conn and return a Plug.Conn so I wondered why did you decide to put current_resource/1 in Guardian.Plug.
Disclaimer : I don't expect a change since it would be breaking for most users but assuming your more experimented developpers than I am, I wanted to ask this question :)
Best Regards
Julien
@julien-leclercq that is interesting but I would say that if you think from domain of the application.
Plug is a module that deals with Plug.Conn. Don't forget that this is functional programming, modules are just for encapsulation of the functions togethers.
By design, I would say that is all about grouping functionalities rather than OOP where you group by objects.
But that is my opinion.
There are many examples of functions in the Plug.Conn module that do not take and return a conn. read_body, get_session, get_req_header, send_chunked.
assign and put_private do return a conn struct but to access the data on them you access via assigns or private attributes. Guardian stores it's data in the connection in the private attribute, namespaced so as not to collide (as suggested by core members). It also provides accessors for how to get at this data. I don't believe this is an unusual pattern and tbh I'm not sure how else you'd do it.
@julien-leclercq it sounds like the module name might be confusing? The Guardian.Plug module isn't a plug, it's a collection of helpers for working with a connection.
yep I got it now thanks for your answers mates !
| gharchive/issue | 2017-11-02T16:31:14 | 2025-04-01T06:46:06.695394 | {
"authors": [
"doomspork",
"hassox",
"julien-leclercq",
"yordis"
],
"repo": "ueberauth/guardian",
"url": "https://github.com/ueberauth/guardian/issues/416",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
298798457 | Is possible a Expo Support?
I believe that all dependencies you use are compatible. Expo is way faster to work than RN.
Cheers.
Hi Michel
Expo actually does not support react-native-navigation.
Care to explain what libraries we include are supported by Expo?
We know Expo is easy, but we feel it's not something you would want to publish.
Agree with @birkir here, expo is far from suitable for a polished performant production product, not to mention the bundle size as well. It's great though to prototype I guess.
@birkir @Salakar Sorry to hijack the thread but I believe it's pretty related: do you believe react-native-starter to be approachable to a React dev with ample experience developing web apps with Redux and MobX, as a first professional RN build?
I am very interested in the looks of this project as it seems incredibly sensible, but as a starter kit it's relatively under the radar and I have the feeling that once I fork this thing and start to build something of real complexity, there's going to be a huge gap in knowledge when something breaks.
I have spun up a few Expo apps and have come to the same conclusion that they're nowhere near what one would expect from a production app.
Do you characterise this repo as approachable to new RN devs coming from a React background, or something which would require previous RN experience to handle?
First of all, the React-Native (Facebook) team itself is recommending first-time use of Expo by focusing on "create-react-native-app" read in "https://facebook.github.io/react -native / docs / getting-started.html "
I believe it's a good entry aproach.
Expo does not support non-native modules. JS modules only. I did not know that the react-native-navigation had native needs. I thought it was pure JS. Anyway, but https://reactnavigation.org/ works perfectly. And it's simple to use.
All JS modules are automatically compatible. Lodash, Mobx (all you use), react-native-blur is supported internally by the EXPO. Firebase is also supported.
What do you mean "we do not want you to publish"? why?
@MichelDiz co-author of react-native-firebase here - expo is most definitely not supported 😛 the Web SDK "works" in EXPO sure, but has a very limited feature set:
@MichelDiz Actually they never state that create-react-native-app is recommended. But they do however say it's the easiest.
You could be right that react-native-navigation works perfectly, maybe for you. For others, it doesn't. There is so much gap between those two libraries in terms of performance and consistency. You should really try RNN out to see the difference.
We know that all JS modules work with Expo, and that is not the issue. The issue is native modules like @Salakar was kind enough to explain to you.
It's a great discussion for pointers where we can update the documentation for concerns like this.
@isaachinman this project is actually a really good start and covers the most common headaches you'd find on your own
The main thing about developing React Native apps without expo is dealing with xcode and android studio - they can be a headache and sometimes cause random issues.
Generally though you only have to deal with the native side complications in 3 places:
Beginning of your project - setting up / adding native modules etc
Upgrading/Removing native dependancies
At the end - Publish/Release cycles
So it's not all the time, just in the seemingly annoying places like the start and end :) but with persistence those easily become second nature - don't be afraid to get stuck in, you get a full picture of the whole process.
If you end up using firebase services from this starter come join our discord - more than happy to discuss / help.
@birkir it's a constant battle for us this, welcome to the show 😄 for example
@birkir , my interpretation is that it is a recommendation. If it were not they would only "quote" the Expo. And they would not create a tool that few understand that this is Expo by under the hood.
birkir I think it you get confused, I said in the name of "React Navigation" that has the "support" open the React-Native team. And it's actually easy for anyone, just evaluate for yourself if I'm wrong. It is much easier than "react-native-navigation" because you do not even need complex native configurations.
Okay, are you going to limit your starter kit in favor for just one module? As Isaachinman said, what is your target audience? yourself, your colleagues or newbies?
Maybe someone who has no interest in Firebase, just want something with "graphql" not to use your kit. So I would have thought that a kit aimed at beginners had to be simple or with a CLI for optional configurations like the Salakar RNF.
Anyway, it was just a question. If the answer is categorically a "no". There is nothing to discuss. Case closed.
And @Salakar , I know your project. I already studied it. But I gave it up for the simple fact that I did not have Expo support. And I know it's a choice of the Expo team, as it would increase the bundle size of the Expo.
My case back there with Firebase was that I wanted only one or two Firebase features and not a whole complex features out of the box. But this is irrelevant. In my opinion StarterKits should be "starter" for people to absorb in a smooth learning curve. Otherwise isn't a "Starter Kit" is a "middle advanced kit".
Few things to point out.
create-react-native-app is not created or maintained by Facebook/React Native.
You are confusing easier with better. We never mention that anything we do here is easier. We don't do things the way they are because it's easier.
We will limit the starter kit to anything that is previously not good enough.
This starter kit is aimed at developers who want to step out of create-react-native-app and into something a little bit more concrete.
"Starter Kit" does not mean it is for beginners. It means "get started quickly" with some specific set of tools.
"has the "support" open the React-Native team" isn't "created or maintained by" oO It is misinterpretation.
And What is the big deal with the Expo? What makes you guys argue so passionately? For me it's a simple and easy way to start a study. And what I see is that isn't a big deal at all because you have an official "support" from RNA team...
Anyway, whatever xD Cheers guys! I appreciate your time.
| gharchive/issue | 2018-02-21T00:19:42 | 2025-04-01T06:46:06.714267 | {
"authors": [
"MichelDiz",
"Salakar",
"birkir",
"isaachinman"
],
"repo": "ueno-llc/react-native-starter",
"url": "https://github.com/ueno-llc/react-native-starter/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1744861658 | isSafeDevice logic is confusing
hi, i just find out that the logic for isSafeDevice is absolutely confusing
return isJailBroken || !isRealDevice || canMockLocation;
the docs indicates that :
// Check if device violates any of the above
but the variable name is isSafeDevice
I suggest to swap either the variable name or the return logic @ufukhawk
safe_device: ^1.1.5
| gharchive/issue | 2023-06-07T01:21:26 | 2025-04-01T06:46:06.829520 | {
"authors": [
"icgoogo",
"ufukhawk"
],
"repo": "ufukhawk/safe_device",
"url": "https://github.com/ufukhawk/safe_device/issues/28",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1931085546 | [Feature Request] I want you to bundle makePositionParams to this.
I want you to bundle makePositionParams to this.
It is internal function of ddu-source-lsp.
https://github.com/uga-rosa/ddu-source-lsp/blob/f1d15a479c3a12fba4e9ec7b046657cb7218b761/denops/ddu_source_lsp/params.ts#L15-L26
Still now, import from GitHub is available in deno, but I thought it is versatile function and I want to use it from stable library.
Added in v0.6.0
| gharchive/issue | 2023-10-07T00:44:33 | 2025-04-01T06:46:06.831220 | {
"authors": [
"kuuote",
"uga-rosa"
],
"repo": "uga-rosa/deno-denops-lsputil",
"url": "https://github.com/uga-rosa/deno-denops-lsputil/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1160370429 | Execution problem
Hi,
I use currently rsync-time-backup, but rsync-time-browse do not work for me, and for sure I am missing something needed; someone can point me to the problem please?
:~$ /home/xxx/rsync-time-browse/tmbrowse.py -h
File "/home/xxx/rsync-time-browse/tmbrowse.py", line 26
while chunk := f.read(8192):
^
SyntaxError: invalid syntax
while
:~$ /home/xxx/rsync-time-backup/rsync_tmbackup.sh -h
Usage: rsync_tmbackup.sh [OPTION]... <[USER@HOST:]SOURCE>
<[USER@HOST:]DESTINATION> [exclude-pattern-file]
Options
-p, --port SSH port.
...
This is my environment,
Linux 5.10.0-11-amd64 #1 SMP Debian 5.10.92-1 (2022-01-18) x86_64 GNU/Linux
debian_version 11.2
python --version Python 3.7.3
The walrus operator on line 26 is only available after Python 3.8. I removed it. Clone the repo again and retry.
Coincidentally I just installed a fresh Debian install and it came with Python 3.9.2
xxx@debian:~/code/rsync-time-browse$ uname -a
Linux debian 5.10.0-11-amd64 #1 SMP Debian 5.10.92-1 (2022-01-18) x86_64 GNU/Linux
cooper@debian:~/code/rsync-time-browse$ python3 --version
Python 3.9.2
Now it works, thanks, but output is just this.
Tested on past environment and another one; same result.
/home/pi/tmbrowse.py ./backup_Xdati/latest/bil/5p.ods
/media/pi/backup_X/latest/bil/5p.ods
24 backups at root: /media/pi/backup_Xdati
I think I got it. Give the current repo a try and let me know if it works.
Explanation: The 'latest' folder is a symlink and it gets expanded on osx automatically but not on linux. os.path.realpth() replaces it with the original.
Now it works perfectly. :-)
Just another hint, the "latest" folder looks like the most "natural" folder where choose the file argument. Maybe I'm wrong. What is your normal choice for this?
Many, many thanks
It can be any folder. If the file you want was recently deleted it won't necessarily be in latest.
Off topic but I just realized you can also use wildcards...
tmbrowse.py ./latest/foo/*.txt
Will report on every text file in folder foo.
I'm thinking of expanding this wildcard capabilities. So you could use a wildcard to find across the whole repo not just a single folder.
| gharchive/issue | 2022-03-05T14:50:54 | 2025-04-01T06:46:06.841081 | {
"authors": [
"calamarim",
"uglygus"
],
"repo": "uglygus/rsync-time-browse",
"url": "https://github.com/uglygus/rsync-time-browse/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
218208432 | Fix CONTRIBUTING.md typos
Fix the directory name of cloned repository ui-router/ng2.
Merged manually
| gharchive/pull-request | 2017-03-30T14:04:36 | 2025-04-01T06:46:06.851791 | {
"authors": [
"christopherthielen",
"gcca"
],
"repo": "ui-router/ng2",
"url": "https://github.com/ui-router/ng2/pull/80",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
175549104 | Added new filters to the listings page
Fixed forms for boolean fields
Added new filters to the listings page
Fixed forms for boolean fields
add full width background to filters
Fixed the filters for Booleans
tweaks
Fixed up verbose names, and fixed issue with string based filters
Will add unit tests and resubmit PR
| gharchive/pull-request | 2016-09-07T16:31:24 | 2025-04-01T06:46:06.900652 | {
"authors": [
"TheUKDave"
],
"repo": "uktrade/navigator",
"url": "https://github.com/uktrade/navigator/pull/41",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
492506542 | Fix unicode errors on develop
Explanation
Made relevant fixes to get yield check to work. Cases I considered
yield is the start of a variable name, like yield_test_file = ... (allow these, modified regex)
yield is inside a multi-line string (don't look at these during linting), note that this can look like
"""
yield(1, 2)
"""
or
export_node("""
yield(1, 2)""")
so looking at start and end of each line is not very useful, since there's no guarantee the multi-line string behaves that way.
yield is simply 'yield' (allow these, modified regex)
Checklist
[ ] The PR title starts with "Fix #bugnum: ", followed by a short, clear summary of the changes. (If this PR fixes part of an issue, prefix the title with "Fix part of #bugnum: ...".)
[ ] The PR explanation includes the words "Fixes #bugnum: ..." (or "Fixes part of #bugnum" if the PR only partially fixes an issue).
[ ] The linter/Karma presubmit checks have passed.
These should run automatically, but if not, you can manually trigger them locally using python -m scripts.pre_commit_linter and bash scripts/run_frontend_tests.sh.
[ ] The PR is made from a branch that's not called "develop".
[ ] The PR has an appropriate "PROJECT: ..." label (Please add this label for the first-pass review of the PR).
[ ] The PR has an appropriate "CHANGELOG: ..." label (If you are unsure of which label to add, ask the reviewers for guidance).
[ ] The PR follows the style guide.
[ ] The PR addresses the points mentioned in the codeowner checks for the files/folders changed. (See the codeowner's wiki page.)
[ ] The PR is assigned to an appropriate reviewer.
If you're a new contributor, please ask on Gitter for someone to assign a reviewer and don't tick this checkbox.
If you're not sure who the appropriate reviewer is, please assign to the issue's "owner" -- see the "talk-to" label on the issue. Do not only request the review but also add the reviewer as an assignee.
Ah, seems the merge went through when I resolved merge conflicts. Was hoping you could review it before it published to main Oppia branch. Sorry about that!
| gharchive/pull-request | 2019-09-11T23:03:03 | 2025-04-01T06:46:06.929636 | {
"authors": [
"ulloaluis"
],
"repo": "ulloaluis/oppia",
"url": "https://github.com/ulloaluis/oppia/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1488146390 | Is folder based docs structure incompatible with static navigation?
I'm trying to build a mkdocs document with this plugin. Since it contains 12 categories and dozens of pages and images, I choose to use folder base docs structure when working on its i18n.
It works perfectly when I just set the folder. However, when I'm trying to add static navigation, it begins to throw an exception as below:
INFO - Building documentation...
INFO - Setting the default 'theme.language' option to 'zh'
INFO - Cleaning site directory
INFO - Translated default navigation to zh
INFO - Translated navigation to zh
Traceback (most recent call last):
File "C:\Users\night\anaconda3\lib\runpy.py", line 194, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\night\anaconda3\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "C:\Users\night\anaconda3\Scripts\mkdocs.exe\__main__.py", line 7, in <module>
File "C:\Users\night\anaconda3\lib\site-packages\click\core.py", line 1128, in __call__
return self.main(*args, **kwargs)
File "C:\Users\night\anaconda3\lib\site-packages\click\core.py", line 1053, in main
rv = self.invoke(ctx)
File "C:\Users\night\anaconda3\lib\site-packages\click\core.py", line 1659, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "C:\Users\night\anaconda3\lib\site-packages\click\core.py", line 1395, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "C:\Users\night\anaconda3\lib\site-packages\click\core.py", line 754, in invoke
return __callback(*args, **kwargs)
File "C:\Users\night\anaconda3\lib\site-packages\mkdocs\__main__.py", line 234, in serve_command
serve.serve(dev_addr=dev_addr, livereload=livereload, watch=watch, **kwargs)
File "C:\Users\night\anaconda3\lib\site-packages\mkdocs\commands\serve.py", line 83, in serve
builder(config)
File "C:\Users\night\anaconda3\lib\site-packages\mkdocs\commands\serve.py", line 76, in builder
build(config, live_server=live_server, dirty=dirty)
File "C:\Users\night\anaconda3\lib\site-packages\mkdocs\commands\build.py", line 302, in build
nav = config.plugins.run_event('nav', nav, config=config, files=files)
File "C:\Users\night\anaconda3\lib\site-packages\mkdocs\plugins.py", line 520, in run_event
result = method(item, **kwargs)
File "C:\Users\night\anaconda3\lib\site-packages\mkdocs_static_i18n\plugin.py", line 418, in on_nav
nav = folder_structure.on_nav(self, nav, config, files)
File "C:\Users\night\anaconda3\lib\site-packages\mkdocs_static_i18n\folder_structure.py", line 402, in on_nav
raise Exception(
Exception: could not find default homepage Page(url='')
Are they incompatible, or do I miss some setting on MkDocs or this plugin?
Here is my project:
https://github.com/CeVIO-User-Guide-Unofficial/CeVIO-AI/tree/i18n
The package version. Thought Material for MkDocs is 9.0.0b3 here, I have met this problem in ver 8.5 before:
mkdocs 1.4.2
mkdocs-material 9.0.0b3
mkdocs-material-extensions 1.1
mkdocs-static-i18n 0.53
I think I have reproduced the problem with the given demo pages. In mkdocs.yml, if you put the index.md under a top level like this:
nav:
- Home:
- index.md
It will throw this exception.
But only folder mode occurs this problem, it works fine in suffix mode.
To be honest, I can avoid this problem if I put index.md in a separate top level, but I really do not want to use the 13th category in the document. And for some reason in this document, I had to put some MarkDown files together with index.md. Would you please fix it?
On the off-chance of sounding dense: I am not sure whether I understand your problem correctly to be honest.
so this config snippet:
nav:
- Home:
- index.md
in conjunction with docs_structure: folder will result in the exception being thrown, correct?
What do you mean with this?:
I can avoid this problem if I put index.md in a separate top level,
and this:
I really do not want to use the 13th category in the document
How would the desired outcome look like exactly?
Also:
And for some reason in this document, I had to put some MarkDown files together with index.md
Which files exactly are you refering to here?
Hello @NachtgeistW
Just like @adamkusmirek I installed your documentation but I find any problem with it and your static navigation.
Is this issue still accurate?
Hello @NachtgeistW
Just like @adamkusmirek I installed your documentation but I find any problem with it and your static navigation.
Is this issue still accurate?
Yes, it is. I will reply to adamkusmirek soon in detail.
@adamkusmirek Thanks for your reply! Because it's too long, I use @ to reply to you.
so this config snippet:
nav:
- Home:
- index.md
in conjunction with docs_structure: folder will result in the exception being thrown, correct?
Yes. According to my research, if index.md isn't put in the root folder, for example, putting it in a subfolder named Home, the plugin will throw this exception.
I can avoid this problem if I put index.md in a separate top level,
means "A simple way to avoid this is to move index.md out of the subfolder." Apologize for my poor English😢
I really do not want to use the 13th category in the document
Please ignore it. This is not related to the exception. Apologize for reporting unrelated content.
BTW: just tried to run your code and everything seemed to go just fine
I think I have pushed the right codes on that branch after reporting this bug by accident. Now I push it again, please use those codes:
https://github.com/CeVIO-User-Guide-Unofficial/CeVIO-AI/tree/i18n_plugin_testing
@NachtgeistW did you make progress on this issue?
You might be interested in #175 and the testing branch of #187
will land in 0.55
| gharchive/issue | 2022-12-10T07:28:42 | 2025-04-01T06:46:06.943413 | {
"authors": [
"NachtgeistW",
"adamkusmirek",
"ultrabug"
],
"repo": "ultrabug/mkdocs-static-i18n",
"url": "https://github.com/ultrabug/mkdocs-static-i18n/issues/168",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2241210654 | Fix image preview placement when messages are preceded by a date in the timeline
Currently image previews are printed one line too high if the message had a date inserted into it indicating a change of date between the current and preceding message. With this PR we conditionally check if a date was inserted and adjust the y offset used to place the rendered the image.
Tested in: XTerm(374), wezterm 20240203-110809-5046fc22, and konsole 24.02.1.
Before change:
After change:
Merged! Thank you for fixing this!
| gharchive/pull-request | 2024-04-13T01:44:37 | 2025-04-01T06:46:07.064872 | {
"authors": [
"reynoldsme",
"ulyssa"
],
"repo": "ulyssa/iamb",
"url": "https://github.com/ulyssa/iamb/pull/257",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
654694686 | gulp script to attempt to keep item templates in sync. #6
the gulp script will find the folders in the ItemTemplate folder and then attempt to copy matching files from the PackageTemplate back into the item template.
It makes sense to do it this way i think because you can actually work on the item template, and confirm it works, vs working on item templates which are not part of a project, so you don't know if the code compiles etc.
n.b this script isn't automatically ran - for now its there as an aid to keep things in sync, and if you add new files to an item template you still need to update the template.json, because it has an array of 'primaryOutputs' which are things that get added to the solution.
| gharchive/pull-request | 2020-07-10T10:50:43 | 2025-04-01T06:46:07.068412 | {
"authors": [
"KevinJump"
],
"repo": "umbraco/Package.Templates",
"url": "https://github.com/umbraco/Package.Templates/pull/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1955472752 | Add profile input validation and resolve build error with temporary s…
…olution
Regarding the Audio DB issues: as mentioned in the chat, can you try the following:
Because I renamed the table for the DB, you should do the following: 1. Stop the device, 2. Wipe all data from the emulated device, 3. Start the device again, 4. Then clean / run the app (flutter clean; flutter pub get; flutter run;)
If that works, then there is no need to comment out any of the Audios stuff (although that is not hooked up to anything important just yet ... that PR is in project).
Thanks!
| gharchive/pull-request | 2023-10-21T12:32:46 | 2025-04-01T06:46:07.109694 | {
"authors": [
"zach-bowman",
"zcappella"
],
"repo": "umgc/fall2023",
"url": "https://github.com/umgc/fall2023/pull/32",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1707313395 | calendar日历 模式为range 范围大约3个月 ,并且设置默认展示日期为当前时间,展开的时候滚动到位置不准确
版本
2.0.36
转载链接
www.uviewui.com
重现步骤
minDate: ${year-2}-${month}-${date},
maxDate: ${year}-${month}-${date},
期望的结果是什么?
希望可以修复,我目前
// 这里需要另一个延时,因为获取宽度后,会进行月份数据渲染,只有渲染完成之后,才有真正的高度
// 因为nvue下,$nextTick并不是100%可靠的
uni.$u.sleep(1500).then(() => {
this.getWrapperWidth()
this.getMonthRect()
})
实际的结果是什么?
实际滚动不到指定位置
我也遇到了这个问题,不过我采取了如下方式进行解决。
我用你这个解决了 但是我是写小程序啊 没写app不用nvue 为什么也生效了
这个就不清楚了,不知道小程序的渲染机制是怎样的。
| gharchive/issue | 2023-05-12T09:47:07 | 2025-04-01T06:46:07.113602 | {
"authors": [
"HateCodeFrank",
"douwzy",
"jiahui336"
],
"repo": "umicro/uView2.0",
"url": "https://github.com/umicro/uView2.0/issues/923",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1407867560 | pull request to merge fork_agrouaze/all_in_1_xarray into origin/pre-release-v0.9
the merge & push command are protected on origin so I have to create a new PR
go for merging!
| gharchive/pull-request | 2022-10-13T14:04:52 | 2025-04-01T06:46:07.142049 | {
"authors": [
"agrouaze"
],
"repo": "umr-lops/xsar",
"url": "https://github.com/umr-lops/xsar/pull/116",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
716571005 | Add LE 07 (2020 Winter)
Add the assignment autograder to the Gradescope and the files to this repo.
❗ LE 07 includes additional files that must be added to this repo along with the README, template, and solution files: airports.csv, airports_ranked_correct.csv.
Gradescope
Create programming assignment
Name = Problem Set XX (2020 Winter) / Lab Exercise XX (2020 Winter)
Assign points
4 Release date: 5 Oct 2020 11:59 PM
End date: 28 May 2021 11:59 PM
Once programming assignment is created, click on three dots (far right of assignment) and edit the settings. Increase CPU to 4 CPUs and save.
Local code
Do a git pull on SI 506 master
Review SI 506 2020win/problem_set_XX or lab_exercise_XX and confirm that the README adheres to the style guide.
Test tests_output.py locally and confirm that solution file passes all tests. Do not forget to delete temporary _solution suffix from import name before generating autograder zip file.
If errors are encounter, fix and issue a PR.
Generate autograder zip file
Upload autograder zip to Gradescope (make sure you are uploading to the correct assignment)
Once the Docker container is built, submit the solutions file to test. If successful, note that test passed in a comment in the issue.
SI506-practice code
Fork and clone this repo
Create a the appropriate directory to store the files
/lab_exercises/le_xx-2020Winter
/problem_sets/ps_xx-2020Winter
Copy the template *.py, solution *.py and README.md and any other required files from the SI506 repo to the SI506-practice repo.
Do a git add, commit -m "Added XX", git push and open a PR.
Feel free to work in concert with other team members.
Test Passed
| gharchive/issue | 2020-10-07T14:15:19 | 2025-04-01T06:46:07.148506 | {
"authors": [
"arwhyte",
"jadeharr"
],
"repo": "umsi-arwhyte/SI506-practice",
"url": "https://github.com/umsi-arwhyte/SI506-practice/issues/11",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
68875961 | python-slugify==1.0.1 disappeared from PyPi
Version 1.0.1 of this package has disappeared from PyPi repository and this broke our pip requirement python-slugify==1.0.1.
Please never remove old versions from PyPy repository.
You are absolutely right. I goofed around and there was no undo button. Please update to 1.0.2.
| gharchive/issue | 2015-04-16T08:49:43 | 2025-04-01T06:46:07.151341 | {
"authors": [
"kurevin",
"un33k"
],
"repo": "un33k/python-slugify",
"url": "https://github.com/un33k/python-slugify/issues/12",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
361249534 | Reason to use 54 dimensions
Hi Una-dinosauria,
Can you please explain me the reason to use only 54 dimensions out of 99 dimensions. I tried to find it in the paper and the code. But I was unable to find. Your response is highly appreciated.
Thank you,
Kavindu
Hi,
Could you please tell me in what part of the code do you see the 54 dimensions?
In general, not all dimensions vary. If they don't then there is no need to model them.
Some other dimensions vary very little (eg < 1e-4). In that case, the normalization process, which subtracts the mean and divides by the standard deviation migth run into numerical issues. Therefore we sometimes also ignore dimensions with very small stdevs.
Hi Una-dinosauria,
I have a conflict in two places.
self.HUMAN_SIZE = 54 in the seq2seq_model.py file
dimensions_to_use.extend( list(np.where(data_std >= 1e-4)[0]) ) in data_utils.py file
In the first one, 54 is defined and the second one the size of dimensions_to_use is 54. Therefore my question of is the 54 defined in first one according to the size of the array dimensions_to_use?
Thank you,
Kavindu
I think that is the case.
Thank you for the confirmation.
| gharchive/issue | 2018-09-18T11:01:19 | 2025-04-01T06:46:07.156166 | {
"authors": [
"KavinduZoysa",
"una-dinosauria"
],
"repo": "una-dinosauria/human-motion-prediction",
"url": "https://github.com/una-dinosauria/human-motion-prediction/issues/37",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
182742417 | Read more
How about an animated read more toggle?
WIP: http://codepen.io/BitsPls/pen/bwKjBQ?editors=1100
I'd probably do something like this
https://jsfiddle.net/link2twenty/mdx7x37r/
I love the way you made it look/work! 👍 Also, good points, i guess those are the limitations of css. 😐 I discussed this website with some community members and one of them remarked that most of these methods either use a "hack", aren't supported well or depend on the DOM content. Ofcourse I pointed out that there is a statement about browser support and testing it before using it in production at the top of the website and I still think its a cool showcase of the possibilities of css.
Is this much different than the accordion?
@una not really, it's the same concept but using checkboxes rather than radio buttons (to have multiple open), it does show how you can change label text based on check state, however. Which might be nice to include in an example somewhere (even if it's just part of accordion or something).
Without knowing the size it is changing too before hand it's impossible to animate it, meaning you can animate it getting smaller but not getting larger.
You can animate max-height to some big value. And toggle overflow to auto for edge-cases.
| gharchive/issue | 2016-10-13T09:42:16 | 2025-04-01T06:46:07.160397 | {
"authors": [
"DevItWithDavid",
"Link2Twenty",
"jj199611",
"marsjaninzmarsa",
"una"
],
"repo": "una/YouMightNotNeedJS",
"url": "https://github.com/una/YouMightNotNeedJS/issues/53",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
298832045 | initial push
inverse lookup using NOT instead of NOT IN subquery .
inverse_lookup/lookups.py
current lookup_names are
neexact
neiexact
necontains
neicontains
nestartswith
neistartswith
neendswith
neiendswith
inverse_lookup is django app, import lookup and Field.register_lookup when app is ready.
test sqlite, mysql, postgresql on trvis-ci, sqlite on local (not passed enviroment variable).
tests/test_settings.py
tests/test_inverse_lookups.py
structure based on cookiecutter-pypackage
リポジトリ名をdjango-lookup-extensionsに変えるので一旦in progressへ
renamed repository and package name.
mainly see below.
lookup_extensions/lookups.py
tests/test_settings.py
tests/test_lookup_extensions.py
| gharchive/pull-request | 2018-02-21T03:42:29 | 2025-04-01T06:46:07.186456 | {
"authors": [
"tsuyukimakoto"
],
"repo": "uncovertruth/django-inverse-lookup",
"url": "https://github.com/uncovertruth/django-inverse-lookup/pull/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
577572868 | Invalid function: edebug-after 0 1
I would like to use undercover in my project https://gitlab.com/tseenshe/haskell-tng.el
However, when I follow the instructions and run
UNDERCOVER_FORCE=true cask exec ert-runner
I get
Invalid function: (edebug-after 0 1 haskell-tng--layout-at-point)
and no further output.
I can only minimise to running the test containing that function
UNDERCOVER_FORCE=true cask exec ert-runner test/haskell-tng-layout-test.el
I followed these instructions, which I found when reading the undercover source code
(message "UNDERCOVER: 2. Run M-: `%s';" "(require 'edebug)")
(message "UNDERCOVER: 3. Run M-x `edebug-all-defs';")
(message "UNDERCOVER: 4. Run M-x `toggle-debug-on-error'.")
(message "UNDERCOVER: 5. Run M-x `eval-buffer';")
but I do not get any error when I eval-buffer on the file containing that function. So I do not know how to proceed.
Can you test if this change fixes the bug?
Please leave a comment if this problem is still reproducible.
Please leave a comment if this problem is still reproducible.
| gharchive/issue | 2020-03-08T21:56:48 | 2025-04-01T06:46:07.195413 | {
"authors": [
"CyberShadow",
"doublep",
"tseenshe"
],
"repo": "undercover-el/undercover.el",
"url": "https://github.com/undercover-el/undercover.el/issues/59",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1960372633 | 🛑 Piped API is down
In caf791f, Piped API (https://pipedapi.unforest.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Piped API is back up in ff6504b after 19 minutes.
| gharchive/issue | 2023-10-25T02:06:10 | 2025-04-01T06:46:07.285109 | {
"authors": [
"unforest"
],
"repo": "unforest/uptime",
"url": "https://github.com/unforest/uptime/issues/2419",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1964676870 | 🛑 Rimgo is down
In bbe9941, Rimgo (https://rimgo.unforest.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Rimgo is back up in b2c5139 after 13 minutes.
| gharchive/issue | 2023-10-27T02:59:55 | 2025-04-01T06:46:07.287515 | {
"authors": [
"unforest"
],
"repo": "unforest/uptime",
"url": "https://github.com/unforest/uptime/issues/2869",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1981165066 | 🛑 Piped API is down
In 13afd6f, Piped API (https://pipedapi.unforest.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Piped API is back up in 1acd07f after 32 minutes.
| gharchive/issue | 2023-11-07T11:45:25 | 2025-04-01T06:46:07.290102 | {
"authors": [
"unforest"
],
"repo": "unforest/uptime",
"url": "https://github.com/unforest/uptime/issues/5289",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1986787389 | 🛑 Piped is down
In 93a061f, Piped (https://piped.unforest.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Piped is back up in 2f843e6 after 1 hour, 11 minutes.
| gharchive/issue | 2023-11-10T03:29:11 | 2025-04-01T06:46:07.292447 | {
"authors": [
"unforest"
],
"repo": "unforest/uptime",
"url": "https://github.com/unforest/uptime/issues/5834",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1995779165 | 🛑 Rimgo is down
In fa1003c, Rimgo (https://rimgo.unforest.net) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Rimgo is back up in d52d8a7 after 43 minutes.
| gharchive/issue | 2023-11-15T23:56:36 | 2025-04-01T06:46:07.294786 | {
"authors": [
"unforest"
],
"repo": "unforest/uptime",
"url": "https://github.com/unforest/uptime/issues/7066",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
482046404 | Unable to parse reddit atom feed
Try to parse
http://www.reddit.com/.rss
rss.Read returns empty struct with nil error.
Can this library parse atom feeds?
How to reproduce:
v , err:= rss.Read("http://www.reddit.com/.rss")
Environment:
go version go1.12.8 linux/amd64
https://www.reddit.com/r/redditdev/comments/5w60r1/error_429_too_many_requests_i_havent_made_many/de7ngx6/
Adding --user-agent "curl:myappname:v0.0 (by /u/ocelost)" to the curl request ensures that the reddit returns content. I will implement it.
@SuddenGunter Could you test?
@030 tried using cod based on this https://github.com/ungerik/go-rss/blob/2f6706b577373dacdb1d6819334e9c8d89932ac4/example/main.go - works good, thank you.
Although one thing I've noticed: not sure reddit is the correct name for variable at Read function - well, it's very specific imho. Mb should call it something like sendUserAgent
| gharchive/issue | 2019-08-18T22:41:09 | 2025-04-01T06:46:07.298445 | {
"authors": [
"030",
"SuddenGunter"
],
"repo": "ungerik/go-rss",
"url": "https://github.com/ungerik/go-rss/issues/18",
"license": "Unlicense",
"license_type": "permissive",
"license_source": "github-api"
} |
2189341123 | Add autoplay blocking patch/functionality
Description
Add autoplay blocking patch/functionality
Who's implementing?
[ ] I'm willing to implement this feature myself
The problem
Autoplay can be incredibly annoying on sites, and can use additional bandwith.
Possible solutions
Add a patch (and feature flag) blocking autoplay functionality such as Brave or Thorium
Alternatives
No response
Additional context
I have zero C++ knowledge, but can experiment with creating a patch.
This has pretty much been proved to be too hard to implement, see #1565 for example
| gharchive/issue | 2024-03-15T19:33:01 | 2025-04-01T06:46:07.301134 | {
"authors": [
"ajleitzke",
"networkException"
],
"repo": "ungoogled-software/ungoogled-chromium",
"url": "https://github.com/ungoogled-software/ungoogled-chromium/issues/2759",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1359807047 | Feat: initial implementation for single artifact release
This PR introduces an initial working implementation for a Keygen.sh Release action.
Closed in favor of identical PR https://github.com/unhack/keygen-action/pull/11 (to have it in a feature branch)
| gharchive/pull-request | 2022-09-02T07:23:17 | 2025-04-01T06:46:07.302747 | {
"authors": [
"hfend"
],
"repo": "unhack/keygen-action",
"url": "https://github.com/unhack/keygen-action/pull/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
197330459 | Alpha blend two images
Two new PixelImageOperations were introduced that blends two images together according to a mask.
The mask describes the blending of the foreground onto the background. Either two PixelImage[RGB] are passed with a separate mask as PixelImage[Double] or the foreground contains an alpha channel that is used as mask.
Looks convenient, thank you. Can you add a one-line comment to indicate what alphaBlending does and what mask is?
| gharchive/pull-request | 2016-12-23T08:30:33 | 2025-04-01T06:46:07.341323 | {
"authors": [
"Andreas-Forster",
"sschoenborn"
],
"repo": "unibas-gravis/scalismo-faces",
"url": "https://github.com/unibas-gravis/scalismo-faces/pull/5",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1404975561 | Argument Parser Improvements
The argument parser can be improved. Currently, treatments must be made for the following cases. The adoption of the suggestions can exclude this necessity.
Instead of providing three different options for the automation level, a single option (-i) can be offered and the user can provide the desired value (choices=[1, 2, 3]) [ref].
Instead of allowing the user to input invalid inputs (e.g. set the model as complete and incomplete), mutually exclusive groups [ref] can be created.
Implemented.
| gharchive/issue | 2022-10-11T17:14:43 | 2025-04-01T06:46:07.343653 | {
"authors": [
"pedropaulofb"
],
"repo": "unibz-core/OntCatOWL",
"url": "https://github.com/unibz-core/OntCatOWL/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
53391859 | Make logfile and number of line selection bookmarkable in the log file viewer
The logfile viewer allows selecting a logfile and a number of files to view. If the list of logfiles is very long, re-selecting the suitable logfile can be cumbersome. Also, it is not possible to share the selection at the moment. It would thus be a great improvement to be able to bookmark this selection along with the number of lines specified.
Hello Olaf,
Did you already add this enhancement into 3.5.0?
I cannot find it there (log file viewer + OSGi configuration).
| gharchive/issue | 2015-01-05T12:34:12 | 2025-04-01T06:46:07.345142 | {
"authors": [
"michal-ociepka",
"olaf-otto"
],
"repo": "unic/neba",
"url": "https://github.com/unic/neba/issues/24",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
375921248 | No test matches the given testcase filter " not able run the tests
[10/31/2018 3:18:49 PM Warning] No test matches the given testcase filter " not able run the tests
I am using 2017 Visual Studio
TechTalk.SpecFlow version 2.4.0
Unickq.SpecFlow.Selenium.SpecFlowPlugin version 1.3.0.0
WebDriver version is 3.14.0.0
When trying run in debug mode getting below exception
'testhost.x86.exe' (CLR v4.0.30319: Dependency finder domain): Unloaded 'C:\PROGRAM FILES (X86)\MICROSOFT VISUAL STUDIO\2017\ENTERPRISE\COMMON7\IDE\EXTENSIONS\TESTPLATFORM\Microsoft.VisualStudio.TestPlatform.ObjectModel.dll'
The thread 0x34c0 has exited with code 0 (0x0).
The thread 0x18d4 has exited with code 0 (0x0).
The program '[16712] testhost.x86.exe: Program Trace' has exited with code 0 (0x0).
The program '[16712] testhost.x86.exe' has exited with code 0 (0x0)
What's your NUnit version, @PoojaGugu ?
NUnit.Framework version 3.10.1.0
Upgraded version 3.11.0.0 still not working
Didn't get the problem you have.
You installed:
Visual Studio 2017
TechTalk.SpecFlow 2.4.0
Unickq.SpecFlow.Selenium.SpecFlowPlugin 1.3.0.0 with dependencies
Selenium 3.14
NUnit 3.11
What do you do next? You generated code from feature files and run it with VS test runner?
when trying to run i am getting above error
Step 1 :Create the feature file
Step 2 :Generated the Step definitions and updated the function
Test explore run the test getting below message
\source\repos\SampleTestOnemoreTime\SampleTestOnemoreTime\bin\Debug\SampleTestOnemoreTime.dll
[11/2/2018 11:23:32 AM Informational] ========== Run test finished: 0 run (0:00:00.5394029) ==========
[11/2/2018 11:24:31 AM Informational] ------ Run test started ------
[11/2/2018 11:24:32 AM Warning] No test is available in C:\Users\venkatasiva.b\source\repos\SampleTestOnemoreTime\SampleTestOnemoreTime\bin\Debug\SampleTestOnemoreTime.dll. Make sure that test discoverer & executors are registered and platform & framework version settings are appropriate and try again.
[11/2/2018 11:24:32 AM Informational] ========== Run test finished: 0 run (0:00:00.6144884) ==========
Do you have NUnit3TestAdapter installed?
Not installed . i will try now with installing the NUnit3TestAdapter
Hi unickq
Its working now thanks for your help
Do you have NUnit3TestAdapter installed?
Thanks, I had the same issue, fixed after installa NUnit3TestAdapter.
| gharchive/issue | 2018-10-31T11:40:18 | 2025-04-01T06:46:07.354312 | {
"authors": [
"HelibertoArias",
"PoojaGugu",
"unickq"
],
"repo": "unickq/SpecFlow.Selenium.Plugin",
"url": "https://github.com/unickq/SpecFlow.Selenium.Plugin/issues/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
375926313 | Fix for coq/coq#7925.
Not for merge yet, will signal when this is ready upstream.
Coq's PR merged. Can you merge this? Thanks!
@mattam82 ping.
| gharchive/pull-request | 2018-10-31T11:54:54 | 2025-04-01T06:46:07.380404 | {
"authors": [
"maximedenes",
"ppedrot"
],
"repo": "unicoq/unicoq",
"url": "https://github.com/unicoq/unicoq/pull/15",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
490986043 | Documentation url paths
Could be great to update current documentation URL paths, because they're not very friendly :
For exemples :
https://docs.uns.network/uns-usethenetwork --> https://docs.uns.network/use
https://docs.uns.network/uns-introduction --> https://docs.uns.network/introduction
We could simply, remove uns prefix in all top routes.
Cc @laurentlourenco
done with branch and pull request https://github.com/unik-name/docs.uns.network/pull/65
| gharchive/issue | 2019-09-09T09:32:24 | 2025-04-01T06:46:07.407142 | {
"authors": [
"Nigui",
"dlecan",
"laurentlourenco"
],
"repo": "unik-name/docs.uns.network",
"url": "https://github.com/unik-name/docs.uns.network/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
822959885 | add scirate
I haven't selected any bounties
There certainly are a few large features that could be treated that way; not sure which is best. I don't have a good sense for how much time estimate an issue should have to receive a bounty; no matter how I think about it the bounty seems like it would be very high to cover dev costs :)
Happy to help discuss/rubber duck the bounty stuff, but I'll merge this in for now. I'll be on the UF discord tomorrow at 8am PDT (15 Mar) if you want to talk synchronously, else I'll send you an email :)
This is great, thanks @silky!
| gharchive/pull-request | 2021-03-05T10:33:54 | 2025-04-01T06:46:07.444489 | {
"authors": [
"crazy4pi314",
"nathanshammah",
"silky"
],
"repo": "unitaryfund/unitaryhack",
"url": "https://github.com/unitaryfund/unitaryhack/pull/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2308572027 | Unable to access lidar data
In the examples section there is no mention on how to use the lidar data. Is there any way to access it using only python sdk without ros ?
it would be great to have this feature
| gharchive/issue | 2024-05-21T15:29:47 | 2025-04-01T06:46:07.445389 | {
"authors": [
"Harshvgupta",
"quantingxie"
],
"repo": "unitreerobotics/unitree_sdk2_python",
"url": "https://github.com/unitreerobotics/unitree_sdk2_python/issues/2",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2394799198 | Remove generated code from repo and generate it compile time
PR Checklist
[x] A description of the changes is added to the description of this PR.
[x] If there is a related issue, make sure it is linked to this PR.
[x] If you've fixed a bug or added code that should be tested, add tests!
[x] If you've added or modified a feature, documentation in docs is updated
Description of changes
Remove generated code from the repo and rather generate it compile time
@vikrantpuppala
lgtm. please address conflict.
I poked around at this a little bit. I think to fix the tab ball creation all you have to do is to change projectJarFiles in Tarball.scala to
val projectJarFiles = Seq(
(LocalProject("serverModels") / Compile / packageBin).value.getAbsoluteFile,
(LocalProject("server") / Compile / packageBin).value.getAbsoluteFile,
(LocalProject("cli") / Compile / packageBin).value.getAbsoluteFile,
(LocalProject("client") / Compile / packageBin).value.getAbsoluteFile,
)
that is, just move "serverModels" first.
Thanks @creechy, I tried that (and a bunch of other things) but still couldn't get this to work. It always fails in a clean env. Once models are generated, it is able to resolve dependencies correctly (i.e. sbt compile|generate createTarball works or even running createTarball two times).. hence, feel like there is something i'm missing in the sbt build.
@rameshchandra/@haogang this is ready to be merged.
| gharchive/pull-request | 2024-07-08T06:53:35 | 2025-04-01T06:46:07.449510 | {
"authors": [
"creechy",
"vikrantpuppala"
],
"repo": "unitycatalog/unitycatalog",
"url": "https://github.com/unitycatalog/unitycatalog/pull/169",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2515210586 | [SPARK] Workaround for read_write credential
at this time, we don't know if the table will be read or written. For now we always equest READ credentials as the server doesn't distinguish between READ and READ_WRITE credentials as of today. When loading a table, Spark should tell if it's for read or write, we can request the proper credential after fixing Spark.
@cloud-fan
| gharchive/pull-request | 2024-09-10T01:29:43 | 2025-04-01T06:46:07.450808 | {
"authors": [
"amaliujia"
],
"repo": "unitycatalog/unitycatalog",
"url": "https://github.com/unitycatalog/unitycatalog/pull/453",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
845557911 | Failed for image
CF bypass is successful, but the image content is broken or corrupted, how to fix this?
Could you be a bit more descriptive than just corrupted image?
Could you give me an example so I can try to replicate the bug?
Here is my script
const puppeteer = require('puppeteer-extra');
const StealthPlugin = require('puppeteer-extra-plugin-stealth');
puppeteer.use(StealthPlugin());
const Koa = require('koa');
const bodyParser = require('koa-bodyparser');
const app = new Koa();
app.use(bodyParser());
const jsesc = require('jsesc');
const headersToRemove = [
"host", "user-agent", "accept", "accept-encoding", "content-length",
"forwarded", "x-forwarded-proto", "x-forwarded-for", "x-cloud-trace-context"
];
const responseHeadersToRemove = ["Accept-Ranges", "Content-Length", "Keep-Alive", "Connection", "content-encoding", "set-cookie"];
(async () => {
let options = {
headless: true,
args: ['--no-sandbox', '--disable-setuid-sandbox']
};
if (process.env.PUPPETEER_SKIP_CHROMIUM_DOWNLOAD)
options.executablePath = '/usr/bin/chromium-browser';
if (process.env.PUPPETEER_HEADFUL)
options.headless = false;
if (process.env.PUPPETEER_PROXY)
options.args.push(`--proxy-server=${process.env.PUPPETEER_PROXY}`);
const browser = await puppeteer.launch(options);
app.use(async ctx => {
if (ctx.query.url) {
const url = ctx.url.replace("/?url=", "");
let responseBody;
let responseHeaders;
const page = await browser.newPage();
if (ctx.method == "POST") {
await page.removeAllListeners('request');
await page.setRequestInterception(true);
page.on('request', interceptedRequest => {
var data = {
'method': 'POST',
'postData': ctx.request.rawBody
};
interceptedRequest.continue(data);
});
}
const client = await page.target().createCDPSession();
await client.send('Network.setRequestInterception', {
patterns: [{
urlPattern: '*',
resourceType: 'Document',
interceptionStage: 'HeadersReceived'
}],
});
await client.on('Network.requestIntercepted', async e => {
let obj = { interceptionId: e.interceptionId };
if (e.isDownload) {
await client.send('Network.getResponseBodyForInterception', {
interceptionId: e.interceptionId
}).then((result) => {
if (result.base64Encoded) {
responseBody = Buffer.from(result.body, 'base64');
}
});
obj['errorReason'] = 'BlockedByClient';
responseHeaders = e.responseHeaders;
}
await client.send('Network.continueInterceptedRequest', obj);
if (e.isDownload)
await page.close();
});
let headers = ctx.headers;
headersToRemove.forEach(header => {
delete headers[header];
});
await page.setExtraHTTPHeaders(headers);
try {
let response;
let tryCount = 0;
response = await page.goto(url, { timeout: 30000, waitUntil: 'domcontentloaded' });
responseBody = await response.text();
while (responseBody.includes("cf-browser-verification") && tryCount <= 10) {
response = await page.waitForNavigation({ timeout: 30000, waitUntil: 'domcontentloaded' });
responseBody = await response.text();
tryCount++;
}
responseHeaders = response.headers();
const cookies = await page.cookies();
if (cookies)
cookies.forEach(cookie => {
const { name, value, secure, expires, domain, ...options } = cookie;
ctx.cookies.set(cookie.name, cookie.value, options);
});
await page.close();
} catch (error) {
if (!error.toString().includes("ERR_BLOCKED_BY_CLIENT")) {
ctx.status = 500;
ctx.body = error;
}
}
// responseHeadersToRemove.forEach(header => delete responseHeaders[header]);
Object.keys(responseHeaders).forEach(header => ctx.set(header, jsesc(responseHeaders[header])));
ctx.body = responseBody;
}
else {
ctx.body = "Please specify the URL in the 'url' query string.";
}
});
app.listen(3000);
})();
Actualy I used original code from this repo but I've comment responseHeadersToRemove.forEach
// responseHeadersToRemove.forEach(header => delete responseHeaders[header]);
The reason is the variable header is not defined.
everything running fine for html, then I've try to get the image from this url
http://localhost:3000/?url=https://cdn1.webtoon.xyz/manga_5ee580463415d/chapter-1/001.jpg
There is a response image received, but the image is broken or corrupted.
I encountered the same error. file tells me that Images are being returned as RIFF container.
I fixed this using response.buffer()
diff --git a/index.js b/index.js
index 9671b72..1a1169c 100644
--- a/index.js
+++ b/index.js
@@ -29,6 +29,7 @@ const responseHeadersToRemove = [ "Accept-Ranges", "Content-Length", "Keep-Alive
if ( ctx.query.url ) {
const url = ctx.url.replace( "/?url=", "" );
let responseBody;
+ let responseData;
let responseHeaders;
const page = await browser.newPage();
if ( ctx.method == "POST" ) {
@@ -60,7 +61,7 @@ const responseHeadersToRemove = [ "Accept-Ranges", "Content-Length", "Keep-Alive
interceptionId: e.interceptionId
} ).then( ( result ) => {
if ( result.base64Encoded ) {
- responseBody = Buffer.from( result.body, 'base64' );
+ responseData = Buffer.from( result.body, 'base64' );
}
} );
obj[ 'errorReason' ] = 'BlockedByClient';
@@ -83,12 +84,14 @@ const responseHeadersToRemove = [ "Accept-Ranges", "Content-Length", "Keep-Alive
waitUntil: 'domcontentloaded'
} );
responseBody = await response.text();
+ responseData = await response.buffer();
while ( responseBody.includes( "cf-browser-verification" ) && tryCount <= 10 ) {
response = await page.waitForNavigation( {
timeout: 30000,
waitUntil: 'domcontentloaded'
} );
responseBody = await response.text();
+ responseData = await response.buffer();
tryCount++;
}
responseHeaders = response.headers();
@@ -105,16 +108,19 @@ const responseHeadersToRemove = [ "Accept-Ranges", "Content-Length", "Keep-Alive
} = cookie;
ctx.cookies.set( cookie.name, cookie.value, options );
} );
- await page.close();
} catch ( error ) {
if ( !error.toString().includes( "ERR_BLOCKED_BY_CLIENT" ) ) {
ctx.status = 500;
ctx.body = error;
}
}
- responseHeadersToRemove.forEach( header => delete responseHeaders[ header ] );
+ await page.close();
+ responseHeadersToRemove.forEach( header => {
+ if ( responseHeaders && header in responseHeaders )
+ delete responseHeaders[ header ];
+ } );
Object.keys( responseHeaders ).forEach( header => ctx.set( header, jsesc( responseHeaders[ header ] ) ) );
- ctx.body = responseBody;
+ ctx.body = responseData;
} else {
ctx.body = "Please specify the URL in the 'url' query string.";
}
patch.gz
@ahmubashshir feel free to open a pull request :)
I fear that my local repo will break existing scripts, because I patched it like cors-anywhere.
don't be afraid of opening one. you could even use the github editor if you don't feel comfortable doing it with the git command directly.
I'm comfortable with git cli :3, ok... cherry-picking now :3
| gharchive/issue | 2021-03-31T00:50:16 | 2025-04-01T06:46:07.484346 | {
"authors": [
"aalfiann",
"ahmubashshir",
"unixfox"
],
"repo": "unixfox/pupflare",
"url": "https://github.com/unixfox/pupflare/issues/16",
"license": "WTFPL",
"license_type": "permissive",
"license_source": "github-api"
} |
1838767614 | 增加功能,方便线上使用。
增加功能:1、支持IP CIDR形式配置 ; 2、保存在内存的数据换成到redis共享存储,便于扩容。
作者已经不维护了,参与这个项目吧,功能更多。https://github.com/bukaleyang/zhongkui-waf
| gharchive/pull-request | 2023-08-07T06:25:56 | 2025-04-01T06:46:07.486504 | {
"authors": [
"Lorryrui",
"bukaleyang"
],
"repo": "unixhot/waf",
"url": "https://github.com/unixhot/waf/pull/45",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1017156536 | What API endpoints are you using?
I'm trying to make my own implementation in Python, but I'm having trouble finding the individual endpoints you're using as I don't really understand Go. If you wouldn't mind, could you give a quick explanation about the endpoints to toggle a light and change the color temperature?
As his README states, it's not just an HTTP API to hit their servers. The pieces that are HTTP based (login, device listing) are in a const block in login.go. What you're asking for is the meat of this project which is a (proprietary?) socket based protocol. His README shows the URL he's sending packets to, but you're going to have to inspect how he constructs the packets in packet.go for each operation. If you want to build this in python, you need to be familiar with bit manipulation as those packets are constructed manually.
| gharchive/issue | 2021-10-06T00:54:31 | 2025-04-01T06:46:07.488085 | {
"authors": [
"KTibow",
"dfrazeeii"
],
"repo": "unixpickle/cbyge",
"url": "https://github.com/unixpickle/cbyge/issues/9",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
838581278 | adds shell.nix
@buffet i was asked to ask you to check if it's alright
It also requires nixpkgs-unstable :/
Yes you do need rustc, in case you have another version installed outside the nix-shell. As far as I am aware it also doesn't really add more dependencies, as cargo and rustc already point to the same version.
And sqlx-cli isn't necessarily needed for building, only for changing the DB. So I'm unsure where to put that.
Ok cool! Thank you so much!
Epic, thanks! Also thanks to @buffet for the review
And @fortuneteller2k.
| gharchive/pull-request | 2021-03-23T10:35:41 | 2025-04-01T06:46:07.490159 | {
"authors": [
"elkowar",
"legendofmiracles"
],
"repo": "unixporn/trup-rs",
"url": "https://github.com/unixporn/trup-rs/pull/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.