id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
904233833 | [Salvador/BA | Remoto] Front-end Vue.js Developer @ nTopus
Descrição da vaga
Construir e otimizar a experiência do usuário e nossas funcionalidades para ajudar a entregar simplicidade e um universo de novas possibilidades aos nossos clientes. Irá atuar em conjunto com os outros profissionais do time, implantando novos recursos, migrando código legado, escrevendo testes e demais atividades relacionadas.
Local
Remoto ou Escritório, Salvador-BA
Benefícios
Plano de saúde e odontológico;
Auxílio home-office para aquisição de acessórios;
R$ 1.200,00 a R$ 2.000,00 em benefícios flexíveis;
Livros e Cursos.
Diferenciais
Home office mesmo após a pandemia;
Escritório físico para os que desejarem;
Aula de Inglês;
Sem dress-code;
Day off;
Happy Hour.
Requisitos
Obrigatórios:
Conhecimentos em Javascript (ES6);
Conhecimentos em HTML5, CSS3;
Conhecimentos em estruturas conteinerizadas (Docker);
Conhecimentos em testes de software;
Conhecimentos em sistemas de versionamento de código (git, Hg (Mercurio), SVN, etc).
Desejáveis:
Ter trabalhado com APIs REST;
Ter trabalhado com frameworks e libs de terceiros;
Ter trabalhado com estrutura de micro-serviços;
Saber trabalhar com Websocket;
Conhecimentos sobre kubernetes;
Linux;
Boas práticas de Log/auditoria;
Métricas na aplicação;
Conhecimentos em Typescript;
Conhecimento em VueJs.
Contratação
CLT R$ 4.000,00 a 6.000,00, dependendo do nível do candidato.
Nossa empresa
A nTopus é uma startup de tecnologia que desenvolve soluções que melhoram a segurança das pessoas e aumentam a eficiência dos processos em tempo real a partir da transmissão de dados sobre múltiplos meios de comunicação, principalmente os rádios de comunicação de voz, que são obrigatórios em qualquer ambiente de missão crítica. Nossas soluções de software realizam a integração com dispositivos de tecnologias distintas a exemplo de tag’s de IoT, rádios digitais, smartphones, câmeras com vídeo analítico e outros sensores e atuadores industriais.
Aqui usamos:
Vue.js
Typescript
Go
Python
Docker
Kubernetes
e coisas que você nunca viu...
Como se candidatar
Por favor, Envie o seu currículo para comercial@ntopus.com.br com o assunto: Vaga Frontend
Labels
Alocação
Flexível
Presencial
Remoto
(sim, todas são válidas e permanecerão após a pandemia)
Experiência
Júnior
Pleno
Regime
CLT
PJ
possibilidade de PJ?
CV enviado!
| gharchive/issue | 2021-05-27T21:48:10 | 2025-04-01T06:46:11.618599 | {
"authors": [
"Cosmess",
"aurium",
"hallss93"
],
"repo": "vuejs-br/vagas",
"url": "https://github.com/vuejs-br/vagas/issues/296",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
269483793 | fix Es6 code issue.
This file is not transpiled, so we should stick to ES5
const is supported in node since v4 and above, so i don't see how is this a problem?
@anubhav7495 I found this program in iOS9 device, so this code is run in browser, not run in node
the old mobile device is not supported const
@zhenhappy I don't think you spin up a dev-server from an iPhone
you can see this link https://github.com/vuejs-templates/webpack/commit/993717604d15b240e07d5da8d55b815452d11581#diff-cac48dfa87b0049be0ff4be8f3d9f6dc
Looks good. would test it locally once before merging. Thanks @zhenhappy
(Note: I merged the fix on the webpack template)
This file is added to the bundle that is served to the client by the dev-server - and since it's not transpiled, will break on any brower or device that doesn'T support ES6.
So changing it to var is the correct fix here.
@LinusBorg But any such browser will also not support service-workers and developing on such a browser for a pwa is quite odd, do you think?
| gharchive/pull-request | 2017-10-30T05:13:13 | 2025-04-01T06:46:11.626431 | {
"authors": [
"LinusBorg",
"Naramsim",
"anubhav7495",
"zhenhappy"
],
"repo": "vuejs-templates/pwa",
"url": "https://github.com/vuejs-templates/pwa/pull/105",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
276289609 | Mtue
ok
This seems to be an accidental PR, closing....
| gharchive/pull-request | 2017-11-23T08:01:32 | 2025-04-01T06:46:11.627277 | {
"authors": [
"LinusBorg",
"valarsu"
],
"repo": "vuejs-templates/webpack",
"url": "https://github.com/vuejs-templates/webpack/pull/1100",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
456718127 | find out a mistake
find out a mistake
原文没有错误,谢谢。
| gharchive/pull-request | 2019-06-17T02:44:59 | 2025-04-01T06:46:11.628958 | {
"authors": [
"ALi1216",
"Justineo"
],
"repo": "vuejs/cn.vuejs.org",
"url": "https://github.com/vuejs/cn.vuejs.org/pull/941",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1203420593 | Typescript 4058 error for ref sugar
Version
3.2.32
Reproduction link
sfc.vuejs.org/
Steps to reproduce
i want to export a $ref variable,like this:
const num = $ref(0);
export default num;
What is expected?
no error
What is actually happening?
i got a ts 4058 error :
Exported variable 'num' has or is using name 'RefType' from external module "path/to/vue@3.2.32/node_modules/vue/macros" but cannot be named.
How can I fix this problem? thanks
tsconfig.json close composite to false can fix it
I am version 3.2.33, encountered the same problem as you, and then I got the latest version number, 3.2.37, after the upgrade can be
I am version 3.2.33, encountered the same problem as you, and then I got the latest version number, 3.2.37, after the upgrade can be
两个中国人在这说英语,难为你了哈哈哈哈
I am version 3.2.33, encountered the same problem as you, and then I got the latest version number, 3.2.37, after the upgrade can be
两个中国人在这说英语,难为你了哈哈哈哈
兄弟,我还有个问题
我生成.d.ts文件的时候全红,咋解决
I am version 3.2.33, encountered the same problem as you, and then I got the latest version number, 3.2.37, after the upgrade can be
两个中国人在这说英语,难为你了哈哈哈哈
兄弟,我还有个问题 我生成.d.ts文件的时候全红,咋解决
https://staging-cn.vuejs.org/guide/extras/reactivity-transform.html#explicit-opt-in
看文档就行了,我webstorm和vscode都能正常识别,或者你vue版本低了?
| gharchive/issue | 2022-04-13T14:56:21 | 2025-04-01T06:46:11.636811 | {
"authors": [
"kikyouink",
"q2578443177",
"zz541843"
],
"repo": "vuejs/core",
"url": "https://github.com/vuejs/core/issues/5712",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1939345196 | Types: type narrowing not work in parent component when child component use union type to defineProps
Vue version
3.3.4
Link to minimal reproduction
https://play.vuejs.org/#eNqNUk1v1DAQ/SuWLwWpGx/KKYRSQD2UA1SUG+YQ4snWxV+yx9tFZf87Y0e7DemHeoniN/PevPm44x9CaDYZeMu7NEQdkCXAHE6lk07b4COyT96yMdJH8kbQoxIkLxnSDd4lZFv2TvKkxqQmvBOTWJXpEGwwPQK9uiLG8E+AQgADA0rONr3JBaBfUZLEgcGPOSaqMep1c5O8I5930jGyMpARbSB+DajJg+Qtq5ES643xt58rhjHD8R4frmH4/Qh+k7YFk/wyQoK4oe4OMezjGshkCZ9ffYEt/R+C1qtsKPuZ4DdI3uTicUr7mJ0i27O86vaizlq79fd0vkVwad9UMVoydzVfcpp92cFTrd/bPWlOKk+6HU1xv7flopnp3ZpGjyRDo9cOIY79AOzChYyX0Yc0iZedvW/ZkS740dsC1bW1LGEk34RQqXv+VV3uQoD409IfEfjxc5JQMGoHldjNTPydK56+ek3Jz17a8oymY17ZPiwOaQr8P8V64GVb14ghtUIMyhFNgdGb2DhA4YIVZ5QmYnaoLayUt2c08uaNUDrhHG4g2dWv6G/ptEhktvnCr/cWVxGo8Vju4mVlF7R56UXoQfnDVez+AfyVXwE=
Steps to reproduce
If type prop is 'select', value should just a 'String[]', but it allow 'String | String[]'.
In React, it is expected.
What is expected?
Type narrowing should work
What is actually happening?
Type narrowing not work
System Info
No response
Any additional comments?
No response
I believe this is caused by the language service not by vue core
/cc @johnsoncodehk
| gharchive/issue | 2023-10-12T07:07:56 | 2025-04-01T06:46:11.640610 | {
"authors": [
"pikax",
"zjtakato"
],
"repo": "vuejs/core",
"url": "https://github.com/vuejs/core/issues/9381",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1329596858 | TypeError: Cannot read properties of null (reading 'instanceMap')
Version
6.2.1
Browser and OS info
Google Chrome 104.0.5112.81
Steps to reproduce
Just open console
What is expected?
No error
What is actually happening?
TypeError: Cannot read properties of null (reading 'instanceMap')
Duplicate of #1892
| gharchive/issue | 2022-08-05T08:05:15 | 2025-04-01T06:46:11.654334 | {
"authors": [
"Akryum",
"shengwei-sam"
],
"repo": "vuejs/devtools",
"url": "https://github.com/vuejs/devtools/issues/1928",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1712721183 | Volar does not takeover when type-checking .js
As you know, in TypeScript, you can define a type-checked file in a few ways:
With a .ts extension & tsconfig.json file for compiler settings
With a .js extension & tsconfig.json with allowJs: true and checkJs: true
With a .js extension & jsconfig.json with checkJs: true.
With a .js extension and // @ts-check at the top of the file
This implies that a .vue file should be interpreted in "TypeScript-mode" if any of the above are true.
However, it appears that Volar / Vue language services only initiate takeover mode if a Vue tag has <script lang="ts"> or <script setup lang="ts"> IMO this is incorrect. For instance, in my setup, jsconfig.json is defined, and set to type-check all JS. I also tried adding "include": ["./lib/**/*.vue", "./lib/**/*.js"] but this had no effect.
I've defined a computed prop with JSDoc like:
computed: {
/** @returns {[string, any][]} */
dynamicSlots () {
return Object.entries(this.$slots).filter(([key]) => key !== 'drag-image' && key !== 'default');
}
}
However, in the <template> block, it reports that dynamicSlots has a type of any. If I change the script setup to <script type="ts">, it works, but I don't think this should be necessary?
Or, is there another way to enable Vue TypeScript language services for a type-checked JS document?
This! Takeover mode not working in JSDoc comments not only results in this doing nothing, but it also results in an error being displayed in my editor.
| gharchive/issue | 2023-05-16T20:48:39 | 2025-04-01T06:46:11.670226 | {
"authors": [
"GrantGryczan",
"matthew-dean"
],
"repo": "vuejs/language-tools",
"url": "https://github.com/vuejs/language-tools/issues/3192",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2170443135 | 代码能够运行,编辑器爆红
使用slot-scope时 编辑器爆红 该怎么调
After updated vue-tsc to 2.0.5, all pinia function get errors: ts-plugin(2339), ts(2339), ts(2769)!
将 vue-tsc 更新到 2.0.5 后,所有 pinia 函数都会出现错误:ts-plugin(2339)、ts(2339)、ts(2769)!
我应该将Vue - Official切换到哪个版本
将 vue-tsc 更新到 2.0.5 后,所有 pinia 函数都会出现错误:ts-plugin(2339)、ts(2339)、ts(2769)!
我应该将Vue - Official切换到哪个版本
切换到1.8.7可以正常使用
将 vue-tsc 更新到 2.0.5 后,所有 pinia 函数都会出现错误:ts-plugin(2339)、ts(2339)、ts(2769)!
我应该将Vue - Official切换到哪个版本
切换到1.8.7可以正常使用
感谢
同样的问题,看来目前只能先降级了
相同的问题,目前看来只能先降级了
没有1.8.7版本了
Please provide a reproduction
@Hao1617 1.8.27
Closing this one as many things have changed since 2.x. Feel free to create a new issue with a minimal reproduction if you're still having issues, thanks!
| gharchive/issue | 2024-03-06T01:19:24 | 2025-04-01T06:46:11.675972 | {
"authors": [
"Hao1617",
"MarkSky",
"Ttou",
"davidmatter",
"qhl123456",
"so1ve"
],
"repo": "vuejs/language-tools",
"url": "https://github.com/vuejs/language-tools/issues/4000",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
330497727 | build a library or web component throw error when enable pwa
Version
3.0.0-beta.15
Reproduction link
https://github.com/anhulife/vue-cli-issue-reproduction
Steps to reproduce
# install dependencies
npm install
# build a web component will throw error
./node_modules/.bin/vue-cli-service build --target wc --name hello-world src/components/HelloWorld.vue
What is expected?
vue-cli-service will produce a single JavaScript file (and its minified version) with everything inlined.
What is actually happening?
this command exit with a error.
Building for production as web component... ERROR TypeError: Cannot read property 'tapAsync' of undefined
TypeError: Cannot read property 'tapAsync' of undefined
at compiler.hooks.compilation.tap.compilation (/Users/******/Sources/vue-cli-demo/node_modules/@vue/cli-plugin-pwa/lib/HtmlPwaPlugin.js:18:63)
at SyncHook.eval [as call] (eval at create (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/HookCodeFactory.js:17:12), <anonymous>:15:1)
at SyncHook.lazyCompileHook [as _call] (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/Hook.js:35:21)
at Compiler.newCompilation (/Users/******/Sources/vue-cli-demo/node_modules/webpack/lib/Compiler.js:451:26)
at hooks.beforeCompile.callAsync.err (/Users/******/Sources/vue-cli-demo/node_modules/webpack/lib/Compiler.js:487:29)
at AsyncSeriesHook.eval [as callAsync] (eval at create (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/HookCodeFactory.js:24:12), <anonymous>:6:1)
at AsyncSeriesHook.lazyCompileHook [as _callAsync] (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/Hook.js:35:21)
at Compiler.compile (/Users/******/Sources/vue-cli-demo/node_modules/webpack/lib/Compiler.js:482:28)
at readRecords.err (/Users/******/Sources/vue-cli-demo/node_modules/webpack/lib/Compiler.js:221:11)
at Compiler.readRecords (/Users/******/Sources/vue-cli-demo/node_modules/webpack/lib/Compiler.js:349:11)
at hooks.run.callAsync.err (/Users/******/Sources/vue-cli-demo/node_modules/webpack/lib/Compiler.js:218:10)
at AsyncSeriesHook.eval [as callAsync] (eval at create (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/HookCodeFactory.js:24:12), <anonymous>:6:1)
at AsyncSeriesHook.lazyCompileHook [as _callAsync] (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/Hook.js:35:21)
at hooks.beforeRun.callAsync.err (/Users/******/Sources/vue-cli-demo/node_modules/webpack/lib/Compiler.js:215:19)
at AsyncSeriesHook.eval [as callAsync] (eval at create (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/HookCodeFactory.js:24:12), <anonymous>:15:1)
at AsyncSeriesHook.lazyCompileHook [as _callAsync] (/Users/******/Sources/vue-cli-demo/node_modules/tapable/lib/Hook.js:35:21)
In case you didn't notice... it was fixed by the commit https://github.com/vuejs/vue-cli/commit/85e6e5ea6c34ac56ab54127d72b86d25271a31fa - which is why your previous issue was closed.
| gharchive/issue | 2018-06-08T02:27:11 | 2025-04-01T06:46:11.683030 | {
"authors": [
"anhulife",
"yyx990803"
],
"repo": "vuejs/vue-cli",
"url": "https://github.com/vuejs/vue-cli/issues/1515",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
346123382 | Debugging tests failed
Version
3.0.0-rc.10
Node and OS info
yarn 1.6.0 / Windows 7
Steps to reproduce
to reproduce this error you create empty project (like this vue create hello-world).
What is expected?
debugging tests
What is actually happening?
you receive the following error:
$ node --inspect-brk ./node_modules/.bin/vue-cli-service test:unit
Debugger listening on ws://127.0.0.1:9229/170adada-3120-4d04-97e9-2881670ea58e
For help see https://nodejs.org/en/docs/inspector
Debugger attached.
D:\Projects\test2\node_modules\.bin\vue-cli-service:2
basedir=$(dirname "$(echo "$0" | sed -e 's,\\,/,g')")
^^^^^^^
SyntaxError: missing ) after argument list
at new Script (vm.js:51:7)
at createScript (vm.js:136:10)
at Object.runInThisContext (vm.js:197:10)
at Module._compile (module.js:613:28)
at Object.Module._extensions..js (module.js:660:10)
at Module.load (module.js:561:32)
at tryModuleLoad (module.js:501:12)
at Function.Module._load (module.js:493:3)
at Function.Module.runMain (module.js:690:10)
at startup (bootstrap_node.js:194:16)
Waiting for the debugger to disconnect...
What options did you chose while creating the project?
For windows the command should be:
node --inspect-brk ./node_modules/@vue/cli-service/bin/vue-cli-service.js test:unit
I'll update the docs.
| gharchive/issue | 2018-07-31T10:15:45 | 2025-04-01T06:46:11.686109 | {
"authors": [
"Akryum",
"fourteenmeister",
"yyx990803"
],
"repo": "vuejs/vue-cli",
"url": "https://github.com/vuejs/vue-cli/issues/2005",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
893376223 | Components non-direct children of Teleport do not trigger vnode-unmounted
Version
3.0.11
Reproduction link
Vue SFC Playground
Steps to reproduce
Place a component with a onUnmounted hook and / or vnode-unmounted event inside a node inside a teleport component.
What is expected?
When the component is unmounted, hook and / or event must be triggered
What is actually happening?
Nothings
I'm not sure this is a bug or by design: only direct-children (e.g. the div that is wrapping Comp) trigger the hook
duplicate of #3623
duplicate of #3623
| gharchive/issue | 2021-05-17T14:12:53 | 2025-04-01T06:46:11.697917 | {
"authors": [
"HcySunYang",
"edison1105",
"germsb",
"posva"
],
"repo": "vuejs/vue-next",
"url": "https://github.com/vuejs/vue-next/issues/3785",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
196967667 | Props on router-view and beforeRouteEnter
Hello,
I figured out that we can pass data to a view component using props on RouterView.
<router-view :foo="bar"></router-view>
I dont know if it s supposed to works like this it s working.
I would like to be able to use those data in beforeRouteEnter to prefetch some data related to this props.
But it dont seams to be possible to use the props data in beforeRouteEnter.
Maybe you can add this feature ?
You already can:
beforeRouteEnter (to, from, next) {
next(vm => {
// access to component instance via `vm`
})
}
(http://router.vuejs.org/en/advanced/navigation-guards.html#in-component-guards)
You can fetch something else with the hook.
You can also use a global state management lib like vuex
If you got more questions, please, ask them on the forum or StackOverflow
I dont want to do it in the hook because i want do it before the route enter
| gharchive/issue | 2016-12-21T15:57:57 | 2025-04-01T06:46:11.701741 | {
"authors": [
"cgarnier",
"posva"
],
"repo": "vuejs/vue-router",
"url": "https://github.com/vuejs/vue-router/issues/1034",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
580620420 | Add aria-current="page" to the active link
What problem does this feature solve?
Users that can see the page can have a nice experience of knowing which link is currently active just by looking at a styled active link with the activeClass.
But visually impaired users, those who rely on screen readers, currently have no way of knowing that the currently selected link is the active link.
To fix this, aria-current="page" adds an extra information to the screen reader that the currently focused link is the same as the current page, so if we add that together with activeClass we'll be aiding both sighted and blind users.
What does the proposed API look like?
Basically, when we add activeClass also pass aria-current="page" to the generated anchor.
Duplicate of https://github.com/vuejs/vue-router/issues/2116
| gharchive/issue | 2020-03-13T14:01:52 | 2025-04-01T06:46:11.704389 | {
"authors": [
"phiter",
"posva"
],
"repo": "vuejs/vue-router",
"url": "https://github.com/vuejs/vue-router/issues/3144",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
614827402 | Displaying a random number when i used a wrong Directive syntax
Version
2.6.11
Reproduction link
https://jsfiddle.net/2zq0tabm/
Steps to reproduce
it is displaying random numbers on screen when i used a wrong syntax for directive
it shows warning i have used a wrong directive syntax but it is also displaying a number on screen.
the number keeps changing when i enter something to input field
I have used the wrong syntax for click event in Vue. But why is a random number displaying for absolute no reason?
Please use the code below to reproduce.
<div id="app">
<input type="text" placeholder="enter here" v-model="name">
<h1>{{name}}</h1>
<button v-click="coun()">Click</button>
<p>{{counter}}</p>
</div>
<script>
var app = new Vue({
el: '#app',
data: {
name: 'some',
counter: 2
},
methods: {
coun(){
this.counter
}
}
})
</script>
What is expected?
Only The warning message in the console.
showing a Number on screen is not accepted
What is actually happening?
it is displaying random numbers on screen when i used a wrong syntax for directive
it shows warning i have used a wrong directive syntax but it is also displaying a number on screen.
the number keeps changing when i enter something to input field
was just learning VueJs Reacting to Changes
The number isn't random, your directive has a function call that gets executed every time it renders, incrementing the counter. But since the counter is also displayed in the template, it renders again and increments again. Vue defends itself by limiting to ~100 calls, that's where the 100 is coming from
| gharchive/issue | 2020-05-08T15:52:08 | 2025-04-01T06:46:11.709274 | {
"authors": [
"posva",
"vinayjoseph705"
],
"repo": "vuejs/vue",
"url": "https://github.com/vuejs/vue/issues/11384",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
300571738 | when use v-model with Array prop, update prop unexpected by someway.
Version
2.5.13
Reproduction link
https://jsfiddle.net/5uLrp7w6/
Steps to reproduce
step1. open
step2. click tab replied.
The detail I wrote in comments.
What is expected?
component update it's share data between parent component should be more strict.
I wish statement list[0].active = true should work.
What is actually happening?
statement list[0].active = true work unexpected, and it modified share data between it's parent component.
This is inherent to javascript nature: objects are passed by reference so all components end up modifying the same object.
BTW, you should declare active in the initial data.
I'm not sure why you are hardcoding values in your example but maybe this is what you want to do: https://jsfiddle.net/441w01qv/
Please, next time consider using the forum, the Discord server or StackOverflow for questions first. But feel free to come back and open an issue if it turns out to be a bug 🙂
| gharchive/issue | 2018-02-27T10:25:27 | 2025-04-01T06:46:11.713999 | {
"authors": [
"Rouberg",
"posva"
],
"repo": "vuejs/vue",
"url": "https://github.com/vuejs/vue/issues/7718",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
384865380 | New gold OpenCollective sponsor doesn't show up in README.md
Version
2.5.17
Reproduction link
https://jsfiddle.net/yect7ohk/ <- actually on this link it works, it only doesn't work on actual repository README
Steps to reproduce
Go to https://github.com/vuejs/vue
Scroll down to Gold OpenCollective Sponsors
Notice that there is no blue bird icon (Getform.io logo)
If you click on Become a Sponsor image, you get redirected to Getform.io website
What is expected?
There should be a Getform.io logo between CodeFirst and Become a Sponsor in the Gold OpenCollective sponsors section.
What is actually happening?
Instead of Getform logo, it's a Become a Sponsor button that redirects to Getform website.
I've been notified on chat.vuejs.org that the logo of the new gold sponsor from OpenCollective is missing from README.md. I don't see it either, but I checked the source code and it is there:
<a href="https://opencollective.com/vuejs/tiers/gold-sponsors/2/website" target="_blank" rel="noopener noreferrer"><img src="https://opencollective.com/vuejs/tiers/gold-sponsors/2/avatar.svg" height="60px"></a>
When I copy the same exact code over to the test environment, the logo shows up properly:
https://jsfiddle.net/yect7ohk/
It may be some weird caching issue related to the fact that the code itself wasn't changed.
It started to show up a moment ago. Only the Become a Sponsor button is missing now.
Seems to be all working now.
| gharchive/issue | 2018-11-27T15:59:44 | 2025-04-01T06:46:11.720996 | {
"authors": [
"gustojs",
"yyx990803"
],
"repo": "vuejs/vue",
"url": "https://github.com/vuejs/vue/issues/9115",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
412666372 | make serverPrefetch() rejection trappable
What problem does this feature solve?
Currently, rejections in serverPrefetch cannot be handled
eg.
serverPrefetch() {
return Promise.reject('myError')
}
What does the proposed API look like?
Maybe send serverPrefetch rejections as renderStream error event ?
eg.
const renderStream = renderer.renderToStream(context)
renderStream.on('error', err => { ... })
Is there necessary to stop the renderer and throw an error for renderStream?
In Vue 2.6.x, I thinks this means "only this component has error, not the hold app",so it choose to stop render this component only, and the warning will be report when hydrating in browser.
I agree that stopping the rendering is too severe for a rejection, but we need at least a mean to accurately detect that something is going wrong during a serverPrefetch.
Just came across this. This issue actually prevents me from implementing renderToStream in my app at all because there's just no way of recovering from that state and for example refresh the page using renderToString. Really hope this could be fixed in Vue 2.7 and we'd get an error event if serverPrefetch fails.
| gharchive/issue | 2019-02-20T22:50:56 | 2025-04-01T06:46:11.724726 | {
"authors": [
"CyberAP",
"FranckFreiburger",
"skyline0705"
],
"repo": "vuejs/vue",
"url": "https://github.com/vuejs/vue/issues/9539",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
649603378 | fix(keep-alive): solve memory leak probelm
What kind of change does this PR introduce? (check at least one)
[x] Bugfix
[ ] Feature
[ ] Code style update
[ ] Refactor
[ ] Build-related changes
[ ] Other, please describe:
Does this PR introduce a breaking change? (check one)
[ ] Yes
[x] No
If yes, please describe the impact and migration path for existing applications:
The PR fulfills these requirements:
[x] It's submitted to the dev branch for v2.x (or to a previous version branch), not the master branch
[ ] When resolving a specific issue, it's referenced in the PR's title (e.g. fix #xxx[,#xxx], where "xxx" is the issue number)
[x] All tests are passing: https://github.com/vuejs/vue/blob/dev/.github/CONTRIBUTING.md#development-setup
[ ] New/updated tests are included
If adding a new feature, the PR's description includes:
[ ] A convincing reason for adding this feature (to avoid wasting your time, it's best to open a suggestion issue first and wait for approval before working on it)
Other information:
Are you fixing a specific issue? https://github.com/vuejs/vue/issues?q=is%3Aissue+is%3Aopen+keep+alive Which one?
If not, can you provide a reproduction?
Version
2.6.11
Reproduction link
https://codesandbox.io/s/zealous-minsky-tv83s?file=/src/App.vue
Steps to reproduce
<template>
<div id="app">
<div class="nav">
<router-link class="nav-item" to="/home">Home</router-link>
<router-link class="nav-item" to="/about">About</router-link>
</div>
<keep-alive include="Home">
<router-view/>
</keep-alive>
</div>
</template>
1.First in home route, click "Collect garbage" and "Take heap snapshot"(snapshot 1)
2.Click About link, then click Home link, then click "Collect garbage" and "Take heap snapshot"(snapshot 2)
3.Make comparison, search detached html
snapshot1
snapshot2
comparison
I use keep-alive to cache home page, when i click to about page then click back to home page, i found that the div of about page was still reachable, but that page should be destroyed.
Then i found that the cached home vnode was not latest, we can find div of about page from this vnode.
It will cause memory leak, so i modified to fix this.
The detached number seems to be constant no matter how many times we navigate between home and about. There is no memory leak
Thank you for your PR but the detached number seems to be constant no matter how many times we navigate between home and about. There is no memory leak:
Thank you for your PR but the detached number seems to be constant no matter how many times we navigate between home and about. There is no memory leak:
This is just a demo to describe the problem, about page only has one dom node, but if about page has so many dom nodes, it will cause memory leak. In my project, it actually happened.
I hope you can check it again. Thanks a lot.
@JonathanChan94 Excuse me, have you solved this problem ? I am confused now.
| gharchive/pull-request | 2020-07-02T03:41:05 | 2025-04-01T06:46:11.737621 | {
"authors": [
"JonathanChan94",
"posva",
"qtttttttttting"
],
"repo": "vuejs/vue",
"url": "https://github.com/vuejs/vue/pull/11484",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
436700476 | Support remapping types for polymorphic entities
An API may return a polymorphic type that the frontend may want to remap to a differently named vuex-orm entity. For example, if morphTo uses record_type to map to entities and record_type has 'team_members', remapping this to a model of entity = 'users' may be preferred, especially if the API cannot be modified.
@Dishwasha Hi! Thanks for the report. There should be a way to map the data structure. This is definitely good enhancement to have. 👍
When data is normalized from a one-to-many source, the _type and _id coming from the backend should probably just be ignored. I'm new to vuex-orm though, so I could just be missing the proper way to do it. Here is my current workaround:
...
static insert ({data}) {
if (Array.isArray(data)){
for (var i in data){
this.remap(data[i]);
}
} else if (data) {
this.remap(data);
}
return super.insert({data});
};
static remap(team) {
for (var key in team){
if (team[key] === null){
delete team[key];
}
}
if (team.links){
for (var i in team.links){
delete team.links[i].linking_type;
delete team.links[i].linking_id;
}
}
};
...
| gharchive/issue | 2019-04-24T13:34:09 | 2025-04-01T06:46:11.809121 | {
"authors": [
"Dishwasha",
"bovesan",
"kiaking"
],
"repo": "vuex-orm/vuex-orm",
"url": "https://github.com/vuex-orm/vuex-orm/issues/346",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
449371051 | How does this handle nested data/relationships when defining a model?
Just came across this plugin and it seems awesome and just the thing I've been looking for.
Scanning briefly over the documentation I can see that I can define the fields & relationships of a model as:
return {
id: this.attr(null),
user_id: this.attr(null),
age: this.attr(''),
sex: this.attr(''),
user: this.belongsTo(User, 'user_id')
}
However, my server returns the data nested as follows:
{
"data": {
"id": "...",
"type": "post",
"attributes": {
"title": "...",
"body": "..."
},
"relationships": {
"user": {
"id": "...",
"type": "user",
"attributes": {
"name": "TestUser"
},
"relationships": []
}
}
}
}
How would I define the relationships for the model using this response since my server doesn't actually return the user_id field in the response but rather the entire user object?
My current idea is to change my server response to actually return the foreign keys under the 'relationships' key, and move the actual returned model to new 'related' key like:
{
"data": {
"id": "...",
"type": "post",
"attributes": {
"title": "...",
"body": "..."
},
"relationships": {
"user_id": "..."
},
"related": {
"user": {
"id": "...",
"type": "user",
"attributes": {
"name": "TestUser"
},
"relationships": []
}
}
}
}
If I implemented this change, could I then define the relationships as (using a dot notation) like below:
user: this.belongsTo(User, 'relationships.user_id')
Or if I don't change the response, could I do it as:
user: this.belongsTo(User, 'relationships.user.id')
@Alig96 Hi! Thank you so much for the feedback! I'm sorry to tell but currently there're no way to map custom structure, so you can't insert data as is. There's an open issue #326 about this one and if we can implement that, then this becomes possible.
I have something similar in my app too, the server returns the models and their relations (other models).
My solution was to make a helper function to parse the API respose recursively and add relation keys if needed.
In your example, i would check if relationship is empty, if not, for each model, get the id and add it to the post object.
Then continue with inserting process.
@Alig96 I'll be closing the issue because I think we can follow this feature at #326. Please feel free to reopen this if there's anything that is not covered by #326 and needs discussion.
| gharchive/issue | 2019-05-28T16:40:10 | 2025-04-01T06:46:11.814478 | {
"authors": [
"Alig96",
"kiaking",
"mr-chick"
],
"repo": "vuex-orm/vuex-orm",
"url": "https://github.com/vuex-orm/vuex-orm/issues/381",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1224831967 | Update some mailgun deps to remove dependency on mgo
github.com/mailgun/timetools has a dependency on mgo's BSON package.
mgo hasn't been updated in ~4+ years and is unsupported.
Additionally github.com/mailgun/holster has replaced
github.com/mailgun/timetools and github.com/mailgun/ttlmap.
While here, bump minimally supported version of go to go 1.17.
https://github.com/vulcand/oxy/issues/196#issuecomment-622613895
Yes, the problems are still here. The holder has too many unrelated dependencies.
Holster was put together with the hope that one day graph pruning would be implemented in go. Since it now supports this, I don't see an issue with including it in the code base. The dependencies introduced with this PR only includes what is needed by the code.
PS: Until moving to go 1.18 you'll have to run go mod tidy -compat=1.17 to get the full pruning.
Maybe I was not clear: the holder has too many unrelated dependencies, it's not a graph problem but a transitivity problem.
For example: google.golang.org/grpc, go.opentelemetry.io/otel, ...
@ldez We are talking about the same thing here. I too am talking about transitive dependencies. Where are they in the transitive package graph of this pr? For example:
$ go mod why go.opentelemetry.io/otel/...
go: downloading go.opentelemetry.io/otel/exporters/jaeger v1.4.1
go: downloading github.com/go-logr/logr v1.2.2
go: downloading github.com/go-logr/stdr v1.2.2
go: downloading github.com/stretchr/objx v0.1.0
# go.opentelemetry.io/otel
(main module does not need package go.opentelemetry.io/otel)
# go.opentelemetry.io/otel/attribute
(main module does not need package go.opentelemetry.io/otel/attribute)
# go.opentelemetry.io/otel/baggage
(main module does not need package go.opentelemetry.io/otel/baggage)
# go.opentelemetry.io/otel/codes
(main module does not need package go.opentelemetry.io/otel/codes)
# go.opentelemetry.io/otel/internal
(main module does not need package go.opentelemetry.io/otel/internal)
# go.opentelemetry.io/otel/internal/baggage
(main module does not need package go.opentelemetry.io/otel/internal/baggage)
# go.opentelemetry.io/otel/internal/global
(main module does not need package go.opentelemetry.io/otel/internal/global)
# go.opentelemetry.io/otel/internal/internaltest
(main module does not need package go.opentelemetry.io/otel/internal/internaltest)
# go.opentelemetry.io/otel/internal/matchers
(main module does not need package go.opentelemetry.io/otel/internal/matchers)
# go.opentelemetry.io/otel/internal/trace/noop
(main module does not need package go.opentelemetry.io/otel/internal/trace/noop)
# go.opentelemetry.io/otel/propagation
(main module does not need package go.opentelemetry.io/otel/propagation)
# go.opentelemetry.io/otel/semconv/v1.4.0
(main module does not need package go.opentelemetry.io/otel/semconv/v1.4.0)
# go.opentelemetry.io/otel/semconv/v1.5.0
(main module does not need package go.opentelemetry.io/otel/semconv/v1.5.0)
# go.opentelemetry.io/otel/semconv/v1.6.1
(main module does not need package go.opentelemetry.io/otel/semconv/v1.6.1)
# go.opentelemetry.io/otel/semconv/v1.7.0
(main module does not need package go.opentelemetry.io/otel/semconv/v1.7.0)
# go.opentelemetry.io/otel/exporters/jaeger
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger)
# go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/agent
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/agent)
# go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/agent/agent-remote
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/agent/agent-remote)
# go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/jaeger
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/jaeger)
# go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/jaeger/collector-remote
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/jaeger/collector-remote)
# go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/zipkincore
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/zipkincore)
# go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/zipkincore/zipkin_collector-remote
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger/internal/gen-go/zipkincore/zipkin_collector-remote)
# go.opentelemetry.io/otel/exporters/jaeger/internal/third_party/thrift/lib/go/thrift
(main module does not need package go.opentelemetry.io/otel/exporters/jaeger/internal/third_party/thrift/lib/go/thrift)
# go.opentelemetry.io/otel/sdk/instrumentation
(main module does not need package go.opentelemetry.io/otel/sdk/instrumentation)
# go.opentelemetry.io/otel/sdk/internal
(main module does not need package go.opentelemetry.io/otel/sdk/internal)
# go.opentelemetry.io/otel/sdk/internal/env
(main module does not need package go.opentelemetry.io/otel/sdk/internal/env)
# go.opentelemetry.io/otel/sdk/resource
(main module does not need package go.opentelemetry.io/otel/sdk/resource)
# go.opentelemetry.io/otel/sdk/trace
(main module does not need package go.opentelemetry.io/otel/sdk/trace)
# go.opentelemetry.io/otel/sdk/trace/tracetest
(main module does not need package go.opentelemetry.io/otel/sdk/trace/tracetest)
# go.opentelemetry.io/otel/trace
(main module does not need package go.opentelemetry.io/otel/trace)
^ (a pruned transitive dependency)
compared to...
$ go mod why -m github.com/pkg/errors
# github.com/pkg/errors
github.com/vulcand/oxy/cbreaker
github.com/mailgun/holster/v4/clock
github.com/pkg/errors
(an unpruned transitive dependency)
I don't think we are talking about the same thing: you are talking about oxy-holster relations, I'm talking about project-oxy-holster relations.
Offtopic: In all cases, I'm not able to merge a PR on oxy, take a look here https://github.com/vulcand/oxy/issues/208#issuecomment-916320800
How does what I said not apply to the project-oxy-holster relations?
@ldez Does Traefik have some reason not to upgrade to 1.7 or 1.8 to get the graph pruning benefit? This would only effect projects that include oxy yet are NOT using 1.7 or 1.8.
Does Traefik have some reason not to upgrade to 1.7 or 1.8 to get the graph pruning benefit?
Yes, mainly because some tools like golangci-lint (because a problem with SSA, the Go team is working on)
https://github.com/golangci/golangci-lint/issues/2649
That is 1.18 specific. My bump was to a min version of 1.17, specifically to avoid stuff like that and to gain pruning. Given the state of things like you point out (PS: Thanks golangci-lint).
Holster is also go1.17 atm and I would expect a bump to go1.18 now there is also out of the question for the same reasons.
See also: https://github.com/freeformz/oxy-deps-test/blob/main/go.mod for the transitive project dependencies.
Also golangci-lint-1.45.2 has no problem running on my oxy branch with go1.17.
FWIW: With go1.16 holster won't even vendor because that has been upgraded to 1.17 as the minimally supported version.
$ go mod vendor
internal error: failed to find embedded files of github.com/mailgun/holster/v4/clock: //go:build comment without // +build comment
@freeformz I will spend some time on your PR tomorrow (its 11pm for me)
FYI: I will fix the CI in a few minutes
Can you rebase your PR?
(It took me longer than expected, now I'm going to sleep :smile:)
@ldez done. Now I'm going back to bed myself.
@ldez I can revert those. They are type aliases, so it's really the same type under the covers. I was getting kind of overly "complete" in doing that change.
I don't why but I was thinking that clock.Duration was an interface.
Could you don't squash commits?
When I will merge I will squash the commits of the PR (GitHub has a button for that)
Okay. I won't.
Just a tip: don't use git push --force but git push --force-with-lease
https://git-scm.com/docs/git-push#Documentation/git-push.txt---force-with-leaseltrefnamegt
You dropped my commit.
The problem I expected occurs: when I update the dependency in Traefik the module google.golang.org/grpc is updated (from v1.38.0 to v1.40.0)
Which dependencies of the Holster (as a library) are problems from my point of view:
github.com/Shopify/toxiproxy -> github.com/prometheus/client_golang
github.com/ahmetb/go-linq
github.com/golang/protobuf
github.com/hashicorp/consul
go.etcd.io/etcd
google.golang.org/grpc
go.opentelemetry.io/otel
Why they are problems: because oxy just needs a fake clock and a TTLmap implementation.
The influence on Prometheus, protobuf, consul, etcd, grpc, and opentelemetry dependencies is a problem because those dependencies are not related to oxy features.
And those dependencies can have a bad impact on clients of oxy.
I'm not a huge fan of utils/helper libraries or packages, but for me, a utils/helper library must be lean and it must be focused on one topic to avoid unnecessary dependencies.
I really want to drop deprecated dependencies, but I don't want to introduce side effects.
Maybe the Holster must be a multi-module package or maybe we have to carry the code, for now, I don't know what is the best solution.
I was surprised by this, but the realized that Traefik has a direct dependency on google.golang.org/grpc and so does holster, so that dependency won't be filtered out by graph pruning, even if the packages imported from holster don't use the grpc package. Maybe there is some room for improvement in the graph pruning implementation. Anyway.
FWIW: If those dependencies are really that sensitive to Traefik then they should/could be replaced in Traefik's go.mod file to ensure they are locked to a specific version, no matter what dependencies say.
Another workaround would be to create an internal vendor a copy of holster's clock package (or the older timetools stuff; personally I like clock better - but this isn't my repo) until the overarching problem is addressed by go and/or holster.
Which course would you like me to pursue ?
Let's go for an internal package with a copy of holster clock and TTLmap.
@ldez done, let me know if anything else needs to be changed.
| gharchive/pull-request | 2022-05-04T01:07:10 | 2025-04-01T06:46:11.839803 | {
"authors": [
"freeformz",
"ldez",
"thrawn01"
],
"repo": "vulcand/oxy",
"url": "https://github.com/vulcand/oxy/pull/223",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
840950824 | Make device::Features::none() a const fn
Issue
Currently Features::none() is not const which means you cannot use it in const expressions like the example below, this would be really handy for setting program wide sets of features.
const MINIMUM_FEATURES: Features = Features {
geometry_shader: true,
..Features::none()
};
@XAMPPRocky I think that would be a good idea to migrate more functions to const fn whenever applicable. Would you like to lead this migration? We already had a partial migration here, I believe it could be extended.
Also to note, @Rua have plans to migrate to Bitflags. As such a part of the current Sets interfaces will deprecate eventually.
It will be a while, though. I want to get the Ash migration done first, and Ash in turn has its own Bitflags migration in the cards. So it's better for Vulkano to wait for that.
I think that would be a good idea to migrate more functions to const fn whenever applicable. Would you like to lead this migration? We already had a partial migration here, I believe it could be extended.
I would have just made a PR, however the code behind none would need to change a lot as right now, more than I'm comfortable. As it relies on using Default::default with a complex macro, and the use of Default::default would need to be removed entirely to support const.
This is now implemented, so this issue can be closed.
| gharchive/issue | 2021-03-25T13:49:25 | 2025-04-01T06:46:11.844788 | {
"authors": [
"Eliah-Lakhin",
"Rua",
"XAMPPRocky"
],
"repo": "vulkano-rs/vulkano",
"url": "https://github.com/vulkano-rs/vulkano/issues/1514",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1958323455 | Feature: refactor as the below architecture
architecture.pdf
Change both name and architecture. Move to qsee repository
| gharchive/issue | 2023-10-24T01:55:07 | 2025-04-01T06:46:11.850851 | {
"authors": [
"vutuanhai237"
],
"repo": "vutuanhai237/qtm",
"url": "https://github.com/vutuanhai237/qtm/issues/5",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
348442910 | Между столбцами в таблице большие промежутки
Между "value" и "type" очень большой промежуток, хотя в верстке они прижаты к правому и левому краю. Это промежуток между любыми столбцами? Он крадет кучу места. Его можно сделать в раза 4-5 меньше.
@vvzvlad
Уменьшил паддинги с 24px с обоих сторон до 4px. Залил в https://github.com/vvzvlad/glue_web_app/pull/95
Ещё у последних двух столбцов (update time, actions) сделал ширину статичной вместо динамической - на больших экранах должно красть меньше места.
работает.
| gharchive/issue | 2018-08-07T18:47:21 | 2025-04-01T06:46:11.888089 | {
"authors": [
"nismangulov",
"vvzvlad"
],
"repo": "vvzvlad/glue_web_app",
"url": "https://github.com/vvzvlad/glue_web_app/issues/90",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
769811586 | RuntimeError('Please set API_URL environment variable')
Traceback (most recent call last):
File "/app/.heroku/python/lib/python3.8/site-packages/bottle.py", line 868, in _handle
return route.call(**args)
File "/app/.heroku/python/lib/python3.8/site-packages/bottle.py", line 1748, in wrapper
rv = callback(*a, **ka)
File "/app/.heroku/python/lib/python3.8/site-packages/bottle.py", line 3650, in wrapper
result = func(*args, **kwargs)
File "/app/checker/main.py", line 51, in index
url=api_url(), json={'code': code}
File "/app/checker/main.py", line 17, in api_url
raise RuntimeError('Please set API_URL environment variable')
RuntimeError: Please set API_URL environment variable
Hey, this error was raised due to the fact that AWS_ENDPOINT environment variable is not set. You have to obtain it and set
before the execution e.g:
export AWS_ENDPOINT=https://xxxxxx.execute-api.us-east-2.amazonaws.com/v1
Close the issues as per comment above
| gharchive/issue | 2020-12-17T10:11:07 | 2025-04-01T06:46:11.893466 | {
"authors": [
"kakamband",
"vyahello"
],
"repo": "vyahello/pep8-checker",
"url": "https://github.com/vyahello/pep8-checker/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
180719511 | Add comment headers for every source file
#5
Added author + description comment headers for every source file. Used git blame to find the major author for each file.
Nice. Thanks again! :+1: That saves me some time.
| gharchive/pull-request | 2016-10-03T19:03:54 | 2025-04-01T06:46:11.935252 | {
"authors": [
"prashcr",
"w0rp"
],
"repo": "w0rp/ale",
"url": "https://github.com/w0rp/ale/pull/33",
"license": "bsd-2-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2137742459 | feat: call to action section
Feature Request
Feature Proposal
Create the "Call to action" as the seventh section of the landing page following the Figma design:
I want to take on this task
I want to take on this task
hey yo @Hardsys-Andre ! Assigned to you 🚀
hey @heronlancellot, I did the PR for issue #20, waiting for a response, follow the link below:
https://github.com/w3b3d3v/web3task-lp/pull/28
hey @heronlancellot, I did the PR for issue #20, waiting for a response, follow the link below: #28
hey yo @Hardsys-Andre ! Perfect
task under review, will be updated during the week
Closed by #28
| gharchive/issue | 2024-02-16T02:01:03 | 2025-04-01T06:46:11.946160 | {
"authors": [
"0xneves",
"Guboehm",
"Hardsys-Andre",
"heronlancellot"
],
"repo": "w3b3d3v/web3task-lp",
"url": "https://github.com/w3b3d3v/web3task-lp/issues/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
160686969 | idlharness and unions
https://bugzilla.mozilla.org/show_bug.cgi?id=1280275
This issue has been moved to https://github.com/w3c/web-platform-tests/issues/7173; please continue all discussion there.
| gharchive/issue | 2016-06-16T15:13:49 | 2025-04-01T06:46:12.131340 | {
"authors": [
"Ms2ger",
"wpt-issue-mover"
],
"repo": "w3c/testharness.js",
"url": "https://github.com/w3c/testharness.js/issues/199",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1851122364 | Add delightfuldot application
Project Abstract
delightfuldot is a lightweight alternative solution for @polkadot/api that aims to address its high memory consumption issue in helping dapps to connect to a large number of Substrate-based networks effectively and efficiently.
Grant level
[ ] Level 1: Up to $10,000, 2 approvals
[x] Level 2: Up to $30,000, 3 approvals
[ ] Level 3: Unlimited, 5 approvals (for >$100k: Web3 Foundation Council approval)
Application Checklist
[x] The application template has been copied and aptly renamed (project_name.md).
[x] I have read the application guidelines.
[x] Payment details have been provided (bank details via email or BTC, Ethereum (USDC/DAI) or Polkadot/Kusama (USDT) address in the application).
[x] The software delivered for this grant will be released under an open-source license specified in the application.
[x] The initial PR contains only one commit (squash and force-push if needed).
[x] The grant will only be announced once the first milestone has been accepted (see the announcement guidelines).
[ ] I prefer the discussion of this application to take place in a private Element/Matrix channel. My username is: @____:matrix.org (change the homeserver if you use a different one)
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Thanks for the application @sinzii very interesting. I think it would be nice to have an alternative to the Polkadot.js API.
Can you check out this other wrapper project sub-API by KodaDot and briefly compare to it?
I'm not seeing that a lot of people have adopted the Coong Wallet so far; do you plan on continuing to maintain that as well?
Considering the above how do you plan to market the API to get devs to switch to it?
Thanks!
@keeganquigley thank you for taking the time to review the proposal and for the questions
Can you check out this other wrapper project sub-API by KodaDot and briefly compare to it?
As far as from what I've taken a look, sub-API is a helper to help manage connections to multiple Substrate-based networks, this is done via a mapping between a network endpoint/prefix and a @polkadot/api ApiPromise instance. This is helpful when you need to manage a lot of ApiPromise instances since you can easily set them up and quickly/lazily get an instance by providing the network endpoint or prefix.
Most of the dapps that need to connect to multiple networks would also need this API management, for example, SubWallet is also using a similar approach to manage their connections to networks.
But under the hood, they are still using @polkadot/api as a mean of connection so they are still being impacted by the aforementioned high memory consumption & slow startup time issues.
delightfuldot is a replacement for @polkadot/api, where we offer similar functionalities but without those issues and dapps can choose to connect to networks via @polkadot/api or delightfuldot. We could probably redefine again the mapping as below:
type ApiType = ApiPromise | DelightfulApi;
const connectionMap: Map<string, ApiType> = new Map<string, ApiType>()
I'm not seeing that a lot of people have adopted the Coong Wallet so far; do you plan on continuing to maintain that as well?
This is understandable and to what we expected since we haven't made any marketing effort for Coong Wallet. Coong Wallet is now still in its early phase of development/incubation, we plan to add more & more useful features for Coong Wallet before we introduce it to the mass, and delightfuldot is a stepping stone for those.
As a website-based wallet, users tend to open multiple tabs of Coong Wallet and we don't want to see each tab consume hundreds of MB or even GB of RAM on users' devices because of the high-memory consumption issue hence the need for a better & lightweight tool such as delightfuldot for connecting a large number of networks efficiency. We also believe a lot of other dapps in the ecosystem are also in our shoes so we decide to go ahead with this proposal so potentially others could benefit from it when the whole idea becomes a reality.
Considering the above how do you plan to market the API to get devs to switch to it?
We haven't put a lot of thought into this, but I think we can start with some of the below ideas:
Prepare good/clear documentation/website.
Reach out to dapps/wallet in the ecosystem that has the issue with @polkadot/api when connecting to a large number of networks to introduce it (potentially SubWallet, Polkadot JS Plus Extension, Kodadot ...)
Start sharing the ideas/work on Polkadot Forum, Twitter, Telegram, or Matrix channels of the ecosystem so people could be aware and start trying around.
...
Thanks @sinzii I had just found it in passing so thanks for looking into it. I appreciate the thorough answers. I will go ahead and mark the application as ready for review and ping the rest of the committee members. In the meantime, it looks like the image links have broken. Can you fix these? Thanks!
it looks like the image links have broken.
Ah, thank you @keeganquigley for the catch, I've just fixed that!
have you talked to any wallet developers or other network intensive dapp developers to see if this is a common concern and that this solution is needed?
Last year, we had a chance to collab with SubWallet team to review the wallet to improve its performance and stability. At that time, the memory consumption of the wallet was quite high (I don't remember the exact number), and the team had made some effort to limit the usage of @polkadot/api and introduce the toggles to turn on/off the connection to networks (so the wallet doesn't need to connect to all of the networks at the same time, but only to those networks that users have assets on) to somehow mitigate this issue.
We haven't had any chance to talk to other wallets or dapps to have their insights but in general, we believe this is an issue that other dapps/wallets would also running into when connecting to a large number of networks via @polkadot/api
Let me know if you have any feedback/thoughts on the application, thanks!
Sorry for the delay, @sinzii. I will review the application and provide some feedback tomorrow.
Hi @Noc2, I totally understand the concern about the coordination, and to some extent absolutely agree with that.
But we would like to start working on this ourselves at least for the first development phase of the project, as the more people involved in a project, the longer it would probably take in making decisions and moving forward as a whole. We expect to make a lot of decisions about the architecture, API, project setup, and structure, especially for an open-source project there would need a lot of considerations in designing the API so that it would be extensible and open for people to contribute later on. We believe we can move faster starting this ourselves.
At some point in the future, we believe there would definitely need to have coordination between wallet/ui providers/projects as there are people/projects starting to use it and run into problems that need to address/standardize so that it would be easier for them to adopt with their use-cases/needs. We do have plans to talk to other wallets in the ecosystem that are currently relying on @polkadot/api (e.g: SubWallet, PolkadotJS-Plus, ...) to start trying this around and gather feedback to further improve the project when it's in a good shape.
As the maintainer of Coong Wallet, we also understand some of the problems that a wallet/project will likely get into when connecting to a large number of networks to design the library so that it would be easy for others to use and adapt.
Let me know how that sounds, thanks!
Thank you everyone for the feedback and suggestions and in response to that, we're currently in discussions with SubWallet team for potential collaboration on this project. We'll update more on the plan this week.
My apologies for the delayed update here, our discussion with SubWallet team has gone a bit more involved with more plans that are out of the scope of this grant. I'll update as soon as we reach an agreement and finalize the whole plan. Thanks!
Sounds good @sinzii thanks for the update, feel free to ping us all once we should take another look.
@keeganquigley Sorry for the delayed reply here, after a prolonged discussion over the past few weeks with several potential approaches, we couldn't really find a good common ground for both teams to move forward in a collaboration unfortunately.
While working in collaboration is a good approach, but for this first iteration, I don't think it works for us for now. So please do let me know if you have any other suggestions to help move this forward, I'm open to reducing the scope of the work to a smaller level to build this up.
With a plan to support 1000 parachains in the future, this further strengthens our belief in a light and fast JS client solution for Polkadot ecosystem.
Thank you!
Thanks @sinzii I appreciate the update. That's a shame, sorry to hear the collaboration didn't work out. I would personally be willing to go forward with a smaller scope for perhaps a level 2 PoC. It's up to you whether you'd want to combine the core functionalities into one milestone or keep it split into two; feel free to refactor the application.
For now I will go ahead and remove the on hold status. Happy to schedule a call as well, if you'd like to discuss further. If so you can book a slot here.
I have read and hereby sign the Contributor License Agreement.
Hi everyone, I've just reduced the proposal scope to a level 2 grant, the work will mostly be focused on building foundational modules and core functionalities. Please take your time to review that application again and let me know if you have any feedback. Thanks!
@keeganquigley Thanks a lot for your support & approval!
@takahser Thank you for taking the time to review the proposal and for the questions:
In M1 you mention "APIs to inspect pallets' events & errors". Does this mean you can only query past events or is it also possible to listen to them in real-time?
By APIs to inspect pallets' events & errors, I meant to get/inspect events and errors for each pallet from the metadata. These are similar to the @polkadot/api APIs to inspect events & errors.
The ability to query events or listen to real-time events will be supported via storage query to system pallet which will be supported in the APIs to execute pallets' storage queries part. For example, to query/subscribe to events for current block, we'll call the API: api.query.system.events, more details about this can be found here
Another thing that is unclear to me is whether there would be any key management functionality i.e. generating new keys, adding existing keys, signing with keys, etc
In the scope of this grant, we will not provide these functionalities, dapp developers can use existing packages such as @polkadot/keyring, @polkadot/ui-keyring for those purposes. We'll mostly focus on the APIs to interact with a Substrate blockchain.
On that note, we also do notice some issues with existing packages for generating & managing keys, so we do have plans to implement our own solution for those functionalities in the future.
@keeganquigley @Noc2 I've just pushed a commit to fix a small typo, so you may need to reapprove the PR. My apologies for the inconvenience.
@sinzii thanks for the prompt reply.
The ability to query past/current events or listen to real-time events will be supported via storage query to system pallet which will be supported in the APIs to execute pallets' storage queries part.
That's great. Could you make this clear in the proposal?
dapp developers can use existing packages such as @polkadot/keyring, @polkadot/ui-keyring for those purposes.
That's fine by me, if these are downstream dependencies for your package. Can you make it clear in the proposal that this will be the case? Just to make sure the integration will be seamless with your package.
After updating on these 2 issues I'll be happy to add my approval as well.
@takahser I've just updated the proposal, let me know if that addresses your concerns, thanks!
Awesome, thank you everyone for your support! 🚀
One of my wishlist items are not downloading whole metadata for parsing simple rpc calls, see if its possible. I had discussion reagarding this with Rob. He said its possible with merkel proof.
@nikw3f I'm very curious about the merkel proof that Rob mentioned, would you mind sharing it with me if he did elaborate on that term and approach to achieve that? Metadata is one thing we need to optimize to achieve a better performance, so aside from the proposed approach, we're also exploring other approaches to further optimize it.
Sure let me connect with him
Hi everyone, please expect the M1 delivery to be delayed a few weeks because we're currently joining the Polkadot Hackathon Vietnam 2023. The project is in a good progress, we're now wrapping up the last pieces of M1, you can check it out now here the repo. Thank you for your understanding!
Hi @sinzii
That's not a problem at all. Could you amend the timeline accordingly?
@takahser Sure, I've just added a PR for the amendment. Thanks!
| gharchive/pull-request | 2023-08-15T09:05:17 | 2025-04-01T06:46:12.386592 | {
"authors": [
"CLAassistant",
"keeganquigley",
"nikw3f",
"semuelle",
"sinzii",
"takahser"
],
"repo": "w3f/Grants-Program",
"url": "https://github.com/w3f/Grants-Program/pull/1907",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
423647798 | Localization
Grant Application
This application is (select one):
[x] Speculative (use this by default)
[ ] an RFP response
This application is (select one):
[ ] Public (fully)
[x] Public with private finances
Abstract
The current proposal is submitted under the section Education and Documentation of the Areas of Interest.
The project is aimed at localizing existing Substrate support documentation (https://docs.substrate.dev/docs) and Substrate Collectibles Workshop (https://shawntabrizi.github.io/substrate-collectables-workshop/#/) into several languages that we believe are crucial to the growth of the number of Substrate and Polkadot supporters, which will help to reach developers in many countries with strong developer communities. It will also assist in setting up localization processes and workflows that can be used for future localization needs.
Checklist
[x] The grants document has been read and understood.
[x] The Google Form will be completed accurately. Note that the Google Form requires the pull request URL.
[x] Abstract (above) is succinct and complete.
[x] The application is being included into the correct directory: either 'targeted' or 'speculative'.
[x] The application includes a project description.
[x] The application includes all names of team members.
[x] The application includes a description of the team's experience.
[x] The application includes all necessary links (e.g. GitHub and LinkedIn)
[x] The "Development Roadmap" section in the application has a timeline of development ("milestones").
[x] The "Development Roadmap" section in the application has an estimate of funds required.
[x] The "Development Roadmap" section gives an indication of the team's long term plans.
[x] The "Development Roadmap" section includes documentation as a deliverable for at least one milestone.
This is a community effort that more closer aligned the ambassador and community program.
| gharchive/pull-request | 2019-03-21T09:59:15 | 2025-04-01T06:46:12.395699 | {
"authors": [
"EdwardAThomson",
"y3v63n"
],
"repo": "w3f/Web3-collaboration",
"url": "https://github.com/w3f/Web3-collaboration/pull/95",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
640640644 | Error when run setup
When I run $ yarn run setup. this error comes
query failed: SELECT "ns"."nspname" AS "table_schema", "t"."relname" AS "table_name", "cnst"."conname" AS "constraint_name", "cnst"."consrc" AS "expression", CASE "cnst"."contype" WHEN 'p' THEN 'PRIMARY' WHEN 'u' THEN 'UNIQUE' WHEN 'c' THEN 'CHECK' END AS "constraint_type", "a"."attname" AS "column_name" FROM "pg_constraint" "cnst" INNER JOIN "pg_class" "t" ON "t"."oid" = "cnst"."conrelid" INNER JOIN "pg_namespace" "ns" ON "ns"."oid" = "cnst"."connamespace" INNER JOIN "pg_attribute" "a" ON "a"."attrelid" = "cnst"."conrelid" AND "a"."attnum" = ANY ("cnst"."conkey") WHERE "t"."relkind" = 'r' AND (("ns"."nspname" = 'public' AND "t"."relname" = 'pet')) error: { error: column cnst.consrc does not exist at Connection.parseE (/home/asus/Documents/projects/efilling/node_modules/pg/lib/connection.js:554:11) at Connection.parseMessage (/home/asus/Documents/projects/efilling/node_modules/pg/lib/connection.js:379:19) at Socket.<anonymous> (/home/asus/Documents/projects/efilling/node_modules/pg/lib/connection.js:119:22) at Socket.emit (events.js:198:13) at Socket.EventEmitter.emit (domain.js:448:20) at addChunk (_stream_readable.js:288:12) at readableAddChunk (_stream_readable.js:269:11) at Socket.Readable.push (_stream_readable.js:224:10) at TCP.onStreamRead [as onread] (internal/stream_base_commons.js:94:17) name: 'error', length: 196, severity: 'ERROR', code: '42703', detail: undefined, hint: 'Perhaps you meant to reference the column "cnst.conkey" or the column "cnst.conbin".', position: '112', internalPosition: undefined, internalQuery: undefined, where: undefined, schema: undefined, table: undefined, column: undefined, dataType: undefined, constraint: undefined, file: 'parse_relation.c', line: '3369', routine: 'errorMissingColumn' }
How can I solve this ?
How can I solve this ?
Try downgrade postgresql to 11.8
How can I solve this ?
Perhaps no longer relevant for you, but for other: update typeorm - https://github.com/typeorm/typeorm/issues/4332#issuecomment-562321524
| gharchive/issue | 2020-06-17T18:06:07 | 2025-04-01T06:46:12.401248 | {
"authors": [
"cladikzone",
"tongrow",
"underscorekadji"
],
"repo": "w3tecch/express-typescript-boilerplate",
"url": "https://github.com/w3tecch/express-typescript-boilerplate/issues/204",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
108740278 | hang on import
I am trying to figure out why an import of a Julia modules works on one machine, but hangs on another. While I hunt down what differences might exist between the machine images -- each provisioned the same way -- I thought I'd post the stack trace to see if it inspired any ideas. Node 0.12.7, julia v0.4-rc2. I start the app, it hangs on an import statement, and after a while I CTRL+C. If I do it quickly, I just get a plain segfault. If I wait a while, I get something like this.
^Cfatal: error thrown and no exception handler available.
InterruptException()
rec_backtrace at /usr/local/lib/julia/libjulia.so (unknown line)
jl_throw at /usr/local/lib/julia/libjulia.so (unknown line)
unknown function (ip: 0x7f32b97d409f)
unknown function (ip: 0x7f32b97d4109)
unknown function (ip: 0x7f32bae13340)
pthread_cond_wait at /lib/x86_64-linux-gnu/libpthread.so.0 (unknown line)
_ZNSt18condition_variable4waitERSt11unique_lockISt5mutexE at /usr/lib/x86_64-linux-gnu/libstdc++.so.6 (unknown line)
_ZN5JMain12syncQueueGetEv at /MyApp/build/node_modules/node-julia/build/Release/nj.node (unknown line)
_Z8doImportRKN2v820FunctionCallbackInfoINS_5ValueEEE at /MyApp/build/node_modules/node-julia/build/Release/nj.node (unknown line)
_ZN2v88internal25FunctionCallbackArguments4CallEPFvRKNS_20FunctionCallbackInfoINS_5ValueEEEE at node (unknown line)
unknown function (ip: 0x7d9af1)
unknown function (ip: 0x3839292060a2)
unknown function (ip: 0x3839295ac819)
unknown function (ip: 0x38392921e8d5)
unknown function (ip: 0x3839295ac3a2)
unknown function (ip: 0x3839295a4c9f)
unknown function (ip: 0x38392958ff71)
unknown function (ip: 0x38392959a379)
unknown function (ip: 0x38392958ff71)
unknown function (ip: 0x3839292bf54c)
unknown function (ip: 0x3839292bd920)
unknown function (ip: 0x3839292bdbb4)
unknown function (ip: 0x3839292bd920)
unknown function (ip: 0x3839292a3b5e)
unknown function (ip: 0x3839292a614b)
unknown function (ip: 0x383929224ac6)
unknown function (ip: 0x3839292a2c46)
unknown function (ip: 0x38392929d12c)
unknown function (ip: 0x3839292998e0)
unknown function (ip: 0x383929290505)
unknown function (ip: 0x38392928fec4)
unknown function (ip: 0x38392926597f)
unknown function (ip: 0x383929264370)
unknown function (ip: 0x38392921ef40)
unknown function (ip: 0x38392921de90)
_ZN2v88internal9Execution4CallEPNS0_7IsolateENS0_6HandleINS0_6ObjectEEES6_iPS6_b at node (unknown line)
_ZN2v88Function4CallENS_6HandleINS_5ValueEEEiPS3_ at node (unknown line)
_ZN4node15LoadEnvironmentEPNS_11EnvironmentE at node (unknown line)
_ZN4node5StartEiPPc at node (unknown line)
__libc_start_main at /lib/x86_64-linux-gnu/libc.so.6 (unknown line)
unknown function (ip: 0x617cdf)
unknown function (ip: (nil))
Strange that this should be caught in jl_throw since the latest fix should prevent julia from catching signals in 0.4; are you using the latest from master or the latest release?
The latest release. I will try latest from master.
I’ve found the difference in provisioning. There was a script I wasn’t running on the putative ‘production’ machines which did another incantation of apt-get update after adding some sources to the list. It was to get a build of chrome for dev machines, but I imagine it must have subtly changed which libraries were being linked. I’ll work to get to the bottom of it.
it would appear I have to build the latest build with JL_OPTIONS_HANDLE_SIGNALS_OFF defined?
My current theory is that the hang is occurring because the module being imported either
Triggers precompilation
Fails to load a module due to a binary dependency not being copied over as part of my deployment
Running with gdb and getting a trace on the error suggests a stack overflow. I bet the error is something on my side -- something not deployed correctly to my target machine. But whatever the problem is isn't shown to me because calls to find the error are somehow infinitely recurring.
#0 0x00007ffff6b9dc87 in _IO_vfprintf_internal (s=s@entry=0x7ffff4d8c6d0, format=<optimized out>,
format@entry=0x7ffff620919c "could not open file %s", ap=ap@entry=0x7ffff4d8c858) at vfprintf.c:1777
#1 0x00007ffff6bc42a3 in _IO_vasprintf (result_ptr=result_ptr@entry=0x7ffff4d8c800, format=format@entry=0x7ffff620919c "could not open file %s",
args=args@entry=0x7ffff4d8c858) at vasprintf.c:62
#2 0x00007ffff588368b in jl_vexceptionf (exception_type=0x7ffded9218d0, fmt=fmt@entry=0x7ffff620919c "could not open file %s",
args=args@entry=0x7ffff4d8c858) at builtins.c:56
#3 0x00007ffff5883b98 in jl_errorf (fmt=fmt@entry=0x7ffff620919c "could not open file %s") at builtins.c:73
#4 0x00007ffff58e4e26 in jl_load (fname=0x7ffdef581cb0 "/MyApp/node_modules/node-julia/lib/nj.jl", len=45) at toplevel.c:612
#5 0x00007fffee6dce80 in julia_include_680 () at boot.jl:261
#6 0x00007ffff587c16b in jl_apply (nargs=1, args=0x7ffff4d8ca90, f=<optimized out>) at julia.h:1328
#7 jl_apply_generic (F=0x7ffdef37d370, args=0x7ffff4d8ca90, nargs=<optimized out>) at gf.c:1684
#8 0x00007ffff58e7520 in jl_apply (nargs=1, args=0x7ffff4d8ca90, f=<optimized out>) at julia.h:1328
#9 jl_call1 (f=0x7ffdef37d370, a=0x7ffdee54e760) at jlapi.c:155
#10 0x00007ffff690349f in nj::Kernel::load() () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#11 0x00007ffff6904585 in nj::Kernel::invoke(std::string const&, _jl_value_t*, _jl_value_t*) ()
from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#12 0x00007ffff69046b0 in nj::Kernel::getError(_jl_value_t*, _jl_value_t*) () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#13 0x00007ffff690f01f in nj::genJuliaError(_jl_value_t*) () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#14 0x00007ffff6911f2e in nj::getJuliaException(_jl_value_t*) () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#15 0x00007ffff6903608 in nj::Kernel::load() () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#16 0x00007ffff6904585 in nj::Kernel::invoke(std::string const&, _jl_value_t*, _jl_value_t*) ()
from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#17 0x00007ffff69046b0 in nj::Kernel::getError(_jl_value_t*, _jl_value_t*) () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#18 0x00007ffff690f01f in nj::genJuliaError(_jl_value_t*) () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#19 0x00007ffff6911f2e in nj::getJuliaException(_jl_value_t*) () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#20 0x00007ffff6903608 in nj::Kernel::load() () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#21 0x00007ffff6904585 in nj::Kernel::invoke(std::string const&, _jl_value_t*, _jl_value_t*) ()
from /MyApp/build/node_modules/node-julia/build/Release/nj.node
The cycle (e.g. #17-#21) repeats as long as I'm willing to page through the results of backtrace
(gdb) thread apply all bt 8
Thread 3 (Thread 0x7ffff4d8a700 (LWP 8421)):
#0 pthread_cond_wait@@GLIBC_2.3.2 () at ../nptl/sysdeps/unix/sysv/linux/x86_64/pthread_cond_wait.S:185
#1 0x00007ffff7704cdc in std::condition_variable::wait(std::unique_lock<std::mutex>&) () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
#2 0x00007ffff68fca1d in JMain::asyncQueueGet() () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#3 0x00007ffff690815f in Trampoline::operator()() () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#4 0x00007ffff7709e40 in ?? () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
#5 0x00007ffff6f1f182 in start_thread (arg=0x7ffff4d8a700) at pthread_create.c:312
#6 0x00007ffff6c4c47d in clone () at ../sysdeps/unix/sysv/linux/x86_64/clone.S:111
Thread 2 (Thread 0x7ffff558b700 (LWP 8420)):
#0 0x00007ffff6b9dc87 in _IO_vfprintf_internal (s=s@entry=0x7ffff4d8c6d0, format=<optimized out>,
format@entry=0x7ffff620919c "could not open file %s", ap=ap@entry=0x7ffff4d8c858) at vfprintf.c:1777
#1 0x00007ffff6bc42a3 in _IO_vasprintf (result_ptr=result_ptr@entry=0x7ffff4d8c800, format=format@entry=0x7ffff620919c "could not open file %s",
args=args@entry=0x7ffff4d8c858) at vasprintf.c:62
#2 0x00007ffff588368b in jl_vexceptionf (exception_type=0x7ffded9218d0, fmt=fmt@entry=0x7ffff620919c "could not open file %s",
args=args@entry=0x7ffff4d8c858) at builtins.c:56
#3 0x00007ffff5883b98 in jl_errorf (fmt=fmt@entry=0x7ffff620919c "could not open file %s") at builtins.c:73
#4 0x00007ffff58e4e26 in jl_load (fname=0x7ffdef581cb0 "/MyApp/node_modules/node-julia/lib/nj.jl", len=45) at toplevel.c:612
#5 0x00007fffee6dce80 in julia_include_680 () at boot.jl:261
#6 0x00007ffff587c16b in jl_apply (nargs=1, args=0x7ffff4d8ca90, f=<optimized out>) at julia.h:1328
#7 jl_apply_generic (F=0x7ffdef37d370, args=0x7ffff4d8ca90, nargs=<optimized out>) at gf.c:1684
(More stack frames follow...)
Thread 1 (Thread 0x7ffff7fea780 (LWP 8416)):
#0 pthread_cond_wait@@GLIBC_2.3.2 () at ../nptl/sysdeps/unix/sysv/linux/x86_64/pthread_cond_wait.S:185
#1 0x00007ffff7704cdc in std::condition_variable::wait(std::unique_lock<std::mutex>&) () from /usr/lib/x86_64-linux-gnu/libstdc++.so.6
#2 0x00007ffff68fbe0d in JMain::syncQueueGet() () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#3 0x00007ffff693ab99 in doImport(v8::FunctionCallbackInfo<v8::Value> const&) () from /MyApp/build/node_modules/node-julia/build/Release/nj.node
#4 0x00000000007b8e62 in v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) ()
#5 0x00000000007d9af1 in ?? ()
#6 0x00002486aa5060a2 in ?? ()
#7 0x00002486aa506001 in ?? ()
(More stack frames follow...)
it would appear I have to build the latest build with JL_OPTIONS_HANDLE_SIGNALS_OFF defined?
That value is defined in julia.h but only in version 0.4+, thus the ifdef.
The infinite recursion usually stems from some error when loading nj.jl since that acts as a generic error processor. If there's an error in nj.jl then it processes an error which loads nj.jl which causes an error, etc. In each of thread 1 and thread 3, it appears the process is blocked in cond_wait and doing nothing except waiting on the result. Thread 2 is where all the action is. if npm install succeeded, then lib/nj.jl should exist but like you said maybe something about this install is messed up?
This is apt-get issue resolved? I've seen lots of problems with ubuntu apt-get install node because it puts it in /usr/bin/node and it's waaaay too old and then both n and (and probably nvm) put it in /usr/local/bin, and there are weird circumstances where both versions of node end up getting used simultaneously (especially by node-gyp).
It's al little hard to figure out what's going on, but it appears on a tightened up environment I get UVError exceptions thrown if when calling import through node-julia, any pre-compilation needs to be carried out. If I arrange for all precompilation to happen out of band, then my app starts up properly.
I assume this is because of a failure to spawn a child julia instance to do the compilation ,but just why, or why the error isn't cleanly conveyed is somewhat beyond me.
I also get a hang when a binary dependency is missing (bad deployment my part), but obviously triggering the stack overflow is problematic; we can't see the underlying error.
I'll close this as I have a workaround and can't contribute enough context to reliably reproduce.
| gharchive/issue | 2015-09-28T20:30:29 | 2025-04-01T06:46:12.413152 | {
"authors": [
"sebastiang",
"waTeim"
],
"repo": "waTeim/node-julia",
"url": "https://github.com/waTeim/node-julia/issues/21",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1728466633 | 🛑 Packages Repository is down
In aeeb9f4, Packages Repository (https://repo.wabarc.eu.org/) was down:
HTTP code: 429
Response time: 474 ms
Resolved: Packages Repository is back up in 4c48abe.
| gharchive/issue | 2023-05-27T05:18:57 | 2025-04-01T06:46:12.416542 | {
"authors": [
"waybackarchiver"
],
"repo": "wabarc/status",
"url": "https://github.com/wabarc/status/issues/601",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1514855978 | 🛑 Wagestream UK Public Website is down
In 92ef1c5, Wagestream UK Public Website (https://wagestream.com/en/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Wagestream UK Public Website is back up in c5da018.
| gharchive/issue | 2022-12-30T21:34:07 | 2025-04-01T06:46:12.438036 | {
"authors": [
"wagestream-deploy"
],
"repo": "wagestream/statuspage",
"url": "https://github.com/wagestream/statuspage/issues/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
212666458 | Top left masthead text (The Wagtail Bakery) should link to homepage
Rather than to "#".
Was closed with PR #98
| gharchive/issue | 2017-03-08T08:14:22 | 2025-04-01T06:46:12.438859 | {
"authors": [
"heymonkeyriot",
"shacker"
],
"repo": "wagtail/bakerydemo",
"url": "https://github.com/wagtail/bakerydemo/issues/96",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1422325668 | In page anchor links sometimes don't work
Example here:
https://guide.wagtail.org/en-latest/how-to/copying-and-aliasing-pages/#copying-pages
When clicking copying pages, you don't get anchored down the page as the id doesn't match.
The ID on the heading is draftjs_exporterenginesstringelt-object-at-0x7fba786ee680 so it looks like drafttail is doing something weird.
This seems to happen if the heading contains subelements, e.g. if you use bold/italic inside the heading. Fix in #225.
| gharchive/issue | 2022-10-25T11:41:21 | 2025-04-01T06:46:12.440575 | {
"authors": [
"laymonage",
"nicklee"
],
"repo": "wagtail/guide",
"url": "https://github.com/wagtail/guide/issues/201",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1057741616 | Sphinx v4: use js files included with sphinx; fix documentation_options
Sphinx 1.8 moved DOCUMENTATION_OPTIONS from the theme templates into a JavaScript template. Sphinx 4 then started adding the JavaScript template into script_files, causing a 404 when built with Sphinx 4.
Since sphinx_wagtail_theme does not inherit from a built-in sphinx theme, it never received the documentation_options.js file upstream (from the base theme). Therefore that file was 404-ing, and some of the newer DOCUMENTATION_OPTIONS were missing.
This change switches the theme to inherit from Sphinx's basic theme, which means we now get all the glorious JS files directly from sphinx rather than providing our own copies (which were taken from some unknown version of sphinx). This will also prevent the theme from diverging from Sphinx functionality (i.e. search) over time.
See: https://www.sphinx-doc.org/en/master/changes.html?highlight=documentation_options#id35
Alternatively, we could drop support for sphinx older than 4, if this comes across as too hacky.
I wouldn't call it too hacky, but seeing as Sphinx 4 came out six months ago now, I would be fine with dropping support for it in the next release of this theme.
Yea, generally themes will inherit from "basic" so as to automatically get whatever that version of sphinx provides. Since we are not inheriting we will want to keep an eye on the basic theme changes with each major release of sphinx, to remain compatible. That being said, converting the theme to inherit from basic is probably not a chore anyone wants to do unless it becomes necessary.
I also know that Read the Docs tends to encourage people to pin some ancient version of sphinx (1.8.5) because that is their default if you don't specify. You can use newer versions with RTD but have to specify it in requirements.txt. So that is something to consider. Sphinx does have a mechanism for a theme to require a specific version and refuse to build on older versions, if it comes down to it. Which might be the path of least resistance.
Sphinx does have a mechanism for a theme to require a specific version and refuse to build on older versions, if it comes down to it. Which might be the path of least resistance.
This seems reasonable to me, too.
Upon further looking into it... I think a lot of the other javascript files (search, jquery, doctools, etc.) are all "hard-coded" in this theme, from whatever version of sphinx they were originally copied from. I might do a bit more investigation to see if we can truly inherit from the basic theme, as recommended by sphinx, which will make all of these concerns mostly go away.
Ok - just made another push. I removed the js file which were copied from (unknown version) of sphinx. Change the theme config to inherit from the basic theme.
So now, js files (or anything else which is not included in our theme) will automatically come directly from whatever version of sphinx the user has installed. This should greatly reduce any compatibility issues between sphinx versions, and maintenance burden.
I tested this on the included /docs/, and also on my project's docs, each works cleanly with no errors. I would like to test this with something deployed to Read The Docs, and also wagtail's docs, to err on the side of caution.
Also tested with Sphinx version 1 and 4. Works in both.
To help test, I have created a build of a custom project using all my open PRs merged together: https://vince-temp.codered.cloud/
@tbrlpld could I get a quick review/approval to merge this in?
@vsalvino looks good to me, merging in
| gharchive/pull-request | 2021-11-18T19:54:23 | 2025-04-01T06:46:12.452636 | {
"authors": [
"Scotchester",
"lb-",
"vsalvino"
],
"repo": "wagtail/sphinx_wagtail_theme",
"url": "https://github.com/wagtail/sphinx_wagtail_theme/pull/114",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2093336152 | StreamField block usage report
Is your proposal related to a problem?
It would be nice to see a block usage report in the built in reports in Wagtail admin.
Describe the solution you'd like
Reporting of all block types, initially just an inventory.
Ability to filter pages by block usage (or even all models).
Ability to configure (maybe in code) the kinds of blocks that will be listed.
Describe alternatives you've considered
It may make sense for this not to be a core report and is better served by an external package, due to both nuance of requirements and performance concerns.
Additional context
See a historic PR from 2022, a starting point for this reporting approach https://github.com/wagtail/wagtail/pull/9267 (and the trigger for raising this issue).
The PR was just a DRAFT / POC and there has been lots of improvements to how reports are built and even universal listings that may change how this kind of thing would be implemented.
https://github.com/cfpb/wagtail-inventory does already serve mostly this purpose and is well maintained by CFPB, there may not be a need for a core Wagtail solution.
Working on this
There may need to be some consensus on if this is a feature to have in the core code first, please comment if you are willing to work on this and how you would see this useful in your current production projects.
This would require an understanding of how StreamField works across all models, specifically pages.
Some performance optimisations would be required and extensive testing of all block types in unit tests.
View our contributing guidelines, add a comment to the issue once you’re ready to start.
@willbarton - any thoughts or insights for this enhancement request? raised from #9267
Tagging @chosak also.
@jhonatan-lopes I believe you had a version of this around the same time as #10850 – what do you think of your version as a candidate to become a built-in report?
Hey @thibaudcolas, I would be happy to. The only caveat is that the version that I implemented for Moz relies on the PageBlock model from wagtail-inventory.
We need a model that tracks the relationship between pages and blocks and that's exactly what the PageBlock model does in wagtail-inventory. I initially thought that a "Blocks usage report" would live better under that package.
👋 The wagtail-inventory report is implemented as a Wagtail Report in the most recent few versions, and looks like this these days:
@jhonatan-lopes I'd be very curious to see what your report looks like compared to the wagtail-inventory version! Would it be possible for you to share a screenshot?
I agree that this would be a useful addition to Wagtail core. The PageBlock approach from wagtail-inventory maintains a reference index of all (page, block type) pairs which makes it easy to search for pages containing a certain block type -- but this is actually somewhat limited. Ideally users would be able to search for block types and values. And it'd be nice if there were a way to leverage the JSON structure of StreamField data to do JSON-based queries and avoid the need to maintain an index at all (we've played with this a bit here but only in a Postgres-specific way). That said, even the basic block usage report that wagtail-inventory provides has been extremely useful for us over the years.
Hey @chosak, sure I'll put a screenshot here. For a little bit of context, the idea of a new report for block types usage came while working together with the design team on Mozilla Foundation during a CMS clean up. They were making an audit of their block types and wanted to quickly answer the following questions:
What are all block types that we use? Can we get a list of those?
What are the block types that we use the most?
Which block types do we use the least?
Is a block type builtin (from Wagtail core) or was it defined by us? Can we quickly check that?
Can we have a list of pages that use that block type?
The repo was already setup with wagtail-inventory and they had access to the "block inventory" that you've shown. However, this is what it looks like for them (with fake data):
In production, it's even worse because there are more languages, so you end up with 8 duplicates (English, French, German, Dutch, Portuguese,...) before moving on to another page.
From the report builtin to wagtail-inventory, it can be seen that getting answers to the questions above is not trivial and it would require either a dev to compile the data or lots of manual labour from an editor to go through the data that is available now.
For this reason, I've scoped up a new "block types usage" report, focusing on displaying the blocks themselves at the centre and how they are used throughout the pages. This is what it looks like with fake data:
The names of the blocks are simplified for better presentation. You can hover over the block's name and a tooltip will display the block's "full name":
It's also possible to get a list of all page types that use that block:
Finally, there's a flag that indicate whether this is a "Core" (Wagtail builtin) app or a "Custom" one and a counter indicating how many pages are using that block.
Implementation
As I've said, the issue here is that this was built specifically for the Mozilla Foundation, so I was able to leverage the PageBlock model and the Postgres database and assemble a queryset without going through the JSON data.
I agree that it would be awesome to query the page data directly using the JSON but I would have to take a very cautious approach, especially considering:
The different database backends
How to get an unique aggregate for all block types (this is easily done with wagtail-inventory by leveraging PageBlock.objects.all().values("block").annotate(....))
How to keep the queries simple
Limitations
The current Moz implementation is still lacking filters, especially for a locale selection.
Thanks to @chosak for reminding me again to reply to this thread. I have a lot of thoughts, but this is something we've been actively working on at CFPB as part of a larger content audit. That's still very much in progress, and we're like to publish the tools we're working on these as additional Wagtail reports, but they're not really more than internal proofs-of-concept right now.
Specific to these questions:
What are all block types that we use? Can we get a list of those?
What are the block types that we use the most?
Which block types do we use the least?
Is a block type builtin (from Wagtail core) or was it defined by us? Can we quickly check that?
Can we have a list of pages that use that block type?
We've got a work-in-progress QuerySet based on Queryish (AWESOME library for making anything a Wagtail report, FWIW) that produces something we're currently working with in CSV to give an answer to all of these questions.
without going through the JSON data.
Except that we're diving into the JSON to get it. I had experimented with PostgreSQL-specific JSON queries constructing temporary tables with some success to get this info, but found that it's almost as fast for our site (which is not small) to do it in Python. And then it's portable to other databases.
Here's our proof-of-concept for block usage right now:
import csv
from dataclasses import dataclass
from wagtail.blocks import (
BoundBlock,
ListBlock,
StreamBlock,
StreamValue,
StructBlock,
StructValue,
)
from wagtail.blocks.list_block import ListValue
from wagtail.models import get_page_models, Site
from queryish import Queryish
dotted_name = lambda cls: ".".join((cls.__module__, cls.__qualname__))
@dataclass
class AuditedBlock:
page_model: type
field: str
path: str
block: type
pages: list
total_occurrences: int = 0
pages_count: int = 0
pages_live_count: int = 0
pages_in_default_site_count: int = 0
# Traverse a stream field and yield back each available block type
def traverse_streamblock(page_model, block, parent=None):
"""Walk model stream block objects to get initial AuditedBlocks"""
audited_block = AuditedBlock(
page_model=dotted_name(page_model),
field=None,
path=(parent + "." + block.name if parent is not None else block.name),
block=dotted_name(block.__class__),
pages=[],
)
yield audited_block
# If this is a StreamBlock or StructBlock it'll have child blocks
if isinstance(block, (StreamBlock, StructBlock)):
for child_block_field_name in block.child_blocks:
yield from traverse_streamblock(
page_model,
block.child_blocks[child_block_field_name],
parent=audited_block.path,
)
elif isinstance(block, ListBlock):
yield from traverse_streamblock(
page_model,
block.child_block,
parent=audited_block.path,
)
else:
pass
# Traverse a stream field's value and yield back each block type in use
def traverse_streamvalue(value, parent=None):
"""Walk model stream value objects to get AuditedBlocks in-use"""
# This is a block type
if isinstance(value, BoundBlock):
path = (
parent + "." + value.block.name
if parent is not None
else value.block.name
)
yield path
yield from traverse_streamvalue(value.value, parent=path)
# This is a StructValue
elif isinstance(value, StructValue):
for child in value.bound_blocks.values():
yield from traverse_streamvalue(child, parent=parent)
elif isinstance(value, ListValue):
for child in value.bound_blocks:
yield from traverse_streamvalue(child, parent=parent)
# This is a sequence of blocks
elif isinstance(value, StreamValue):
for child in value:
yield from traverse_streamvalue(child, parent=parent)
class BlockUsageQuerySet(Queryish):
"""Return a QuerySet-like object for querying block type usage"""
def get_filtered_page_models(self):
global_page_models = get_page_models()
page_models = [
val
for key, val in self.filters
if key == "page_model" and val in global_page_models
]
return page_models if len(page_models) > 0 else global_page_models
def get_filtered_streamfield_names(self, page_model):
all_streamfields = page_model.get_streamfield_names()
streamfields = [
val
for key, val in self.filters
if key == "field" and val in all_streamfields
]
return streamfields if len(streamfields) > 0 else all_streamfields
def audit_blocks_for_page_model(self, page_model):
# Get the StreamFields on the page model
streamfields = self.get_filtered_streamfield_names(page_model)
# A dictionary to hold counts of each block on a page
page_blocks = {}
# First populate all available blocks
for streamfield_name in streamfields:
streamfield = page_model._meta.get_field(streamfield_name)
page_blocks[streamfield_name] = {}
# Traverse child blocks of this stream field (this avpathoids capturing
# the containing stream block, so we can line-up with stream values
# below)
for child_block_name in streamfield.stream_block.child_blocks:
for audited_block in traverse_streamblock(
page_model,
streamfield.stream_block.child_blocks[child_block_name],
):
audited_block.field = streamfield_name
page_blocks[streamfield_name][
audited_block.path
] = audited_block
# Get the default Wagtail site (this avoids the Trash)
site = Site.objects.get(is_default_site=True)
# Get a queryset for all pages of this type
page_queryset = page_model.objects.exact_type(page_model)
# Loop through the queryset, and traverse each streamfield
for page in page_queryset:
for streamfield_name in streamfields:
streamfield_value = getattr(page, streamfield_name)
streamfield_dict = page_blocks[streamfield_name]
for block_path in traverse_streamvalue(streamfield_value):
# Get the AuditedBlock object for this path
audited_block = streamfield_dict[block_path]
audited_block.total_occurrences += 1
if page not in audited_block.pages:
audited_block.pages.append(page)
audited_block.pages_count += 1
if page.live:
audited_block.pages_live_count += 1
if page.is_descendant_of(site.root_page):
audited_block.pages_in_default_site_count += 1
return page_blocks
def run_query(self):
audited_blocks = []
for page_model in self.get_filtered_page_models():
page_blocks = self.audit_blocks_for_page_model(page_model)
# Flatten the dictionary of dictionaries of blocks to a list and
# extend the audited block list
audited_blocks.extend(
block
for streamfield, blocks in page_blocks.items()
for block in blocks.values()
)
# self.ordering
# Slice based on queryset slicing offset/limit
return audited_blocks[
self.offset : self.offset + self.limit if self.limit else None
]
This then works like a queryset, with filtering and slicing, providing the AuditedBlock dataclass as a result.
from mysite.models import FooPage
queryset = BlockUsageQuerySet().filter(page_model=FooPage, field="content").all()
We're using this to output a simple CSV with a form like:
page_model
field
path
block
occurrences
pages
live
in site
v1.models.learn_page.DocumentDetailPage
sidefoot
related_metadata.content.date
wagtail.blocks.struct_block.StructBlock
1925
1082
1075
1079
v1.models.learn_page.DocumentDetailPage
sidefoot
related_metadata.content.date.heading
wagtail.blocks.field_block.CharBlock
1925
1082
1075
1079
v1.models.learn_page.DocumentDetailPage
sidefoot
related_metadata.content.date.date
wagtail.blocks.field_block.DateBlock
1925
1082
1075
1079
ask_cfpb.models.answer_page.AnswerPage
answer_content
text
wagtail.blocks.struct_block.StructBlock
1821
1715
1418
1710
And so on. Because we're capturing the pages using a block as a list, we can introspect those as well, though right now we're just doing counts.
Emphasizing again that this is very much a proof-of-concept, maybe it's helpful as an alternative that does dive into the JSON?
Right now I don't see this as something that replaces wagtail-inventory either, but complements it, and becomes part of that larger set of audit tools we've been building as we work through what our needs are.
Those other audit tools are around content specifically, like being able to search in blocks, learn their streamfield-migration-compatible paths, etc. I want to do a larger write-up with our content management lead about this whole process of auditing our blocks and content, especially given our Wagtail site has 8 years work of content to wade through now. But we're very far from done.
Wow, thanks so much for sharing that code @willbarton! :tada: :rocket: that is certainly impressive!
It seems like using queryish can indeed be a great alternative. I believe this would allow us to dive into the JSON without resorting to custom/complex SQL, which is what I was kinda hoping to avoid (not necessarily the JSON itself). Bonus points for being portable to other DB backends as well.
For implementing a report like this into Wagtail, I think we would have to divide the approach into two:
First, a piece of work to leverage the queryish package to gather block data in a structure format like you have presented here (pure Wagtail/queryish, no wagtail-inventory of course)
Second, the block types usage report itself -- built on top of this queryish api where we could easily manipulate the data
My only concern would be doing aggregations/annotations on the data. I could be wrong, but I'm not sure that queryish support those yet. In any case, it would be a matter of doing those in pure Python.
@thibaudcolas what do you think? Is this worth exploring as part of Wagtail itself?
FWIW, we've published the block usage and page search audit tools we've been using here: https://github.com/cfpb/wagtail-content-audit
It's still a work-in-progress, and right now the block usage and page search tools are only usable as management commands or directly in code. I still want to wrap them in a report.
Been a while but my take: yes it’s a good fit for Wagtail’s features because StreamField is such a critical part of its approach to content management. From a maintainers’ perspective, I’d worry about whether this can be accomplished within a reasonable code complexity budget, and whether this can be done with compatibility for all databases we support.
| gharchive/issue | 2024-01-22T08:17:15 | 2025-04-01T06:46:12.485576 | {
"authors": [
"Scotchester",
"chosak",
"jhonatan-lopes",
"lb-",
"thibaudcolas",
"willbarton"
],
"repo": "wagtail/wagtail",
"url": "https://github.com/wagtail/wagtail/issues/11506",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
752466205 | Document migrations steps from wagtail-modeltranslation
Is your proposal related to a problem?
Upgrading an existing site from wagtail 2.10 with infoportugal/wagtail-modeltranslation to wagtail 2.11 native i18n without losing existing translations.
Describe the solution you'd like
Documented steps to perform the upgrade and data migration?
Describe alternatives you've considered
(Write your answer here.)
Additional context
(Write your answer here.)
This is a complex process, and probably more for a blog post than documentation.
There shouldn't be anything stopping wagtail-modeltranslation from running 2.11 with Wagtail's native i18n disabled.
@kaedroho Did you ever end up doing that migration? I'm going to have to bring a few sites up to date with considerable multilingual content so hoping others have blazed the trail ahead of me. That said, the apps use django-modeltranslation anyway, so the migration may not necessarily save much overhead in the end.
@kaedroho Can I work with you on this? you mentioned it could be quite complex and I figured I would need help.
Hey, I haven't worked on a wagtail-modeltranslation site before, but here is a guess.
The main difference between wagtail-modeltranslation and Wagtail's built-in i18n is that Wagtail will use separate pages for each translation in separate trees, rather than using additional fields on the same page.
So I think the first step would be to duplicate your pages for each language. Hopefully, if wagtail-modeltranslation follows Django's current language, this might work:
from django.conf import settings
from django.utils import translation
from wagtail.models import Page
from wagtail.locales.models import Locale
default_locale = Locale.get_default()
for language_code, language_name in settings.LANGUAGES:
locale, created = Locale.objects.get_or_create(language_code=language_code)
with translation.override(language_code):
for page in Page.objects.filter(locale=default_locale, depth__gte=2).specific():
page.copy_for_translation(locale)
Hopefully, this leaves you with pages for each locale where the primary fields have been change to their specific locale.
| gharchive/issue | 2020-11-27T20:46:32 | 2025-04-01T06:46:12.492139 | {
"authors": [
"dameeolawuyi",
"denisvlr",
"dwasyl",
"kaedroho"
],
"repo": "wagtail/wagtail",
"url": "https://github.com/wagtail/wagtail/issues/6593",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1344168439 | Make admin page titles more useful for admin users
Is your proposal related to a problem?
Every single Wagtail admin screen puts “Wagtail” at the front (left) of the <title> – which occludes the more useful information of which page you might have open as soon as you have a few tabs.
The example tab I have open is the page editor for a page titled: “Organising A Museum Website”.
The page title is: "Wagtail - Editing News Article Page - Organising A Museum Website".
This is what I see:
If you have a few Wagtail admin tabs open they all say "Wagtail". This necessitates flipping between them (or at least hovering over every tab) to figure out which tab has which page.
Describe the solution you'd like
For an experienced content editor, the favicon uniquely identifies Wagtail tabs, so for most pages, I think omitting "Wagtail" or shifting it to the end of the <title> would allow more salient information to be visible
I think the most significant information presented in the <title> of a page editor tab is the title of the page being edited. The favicon tells you you're in the admin, so "Editing" is redundant.
Simply reordering the elements from most significant to least significant (from the perspective of a content editor) would improve the UI. To persist with my example:
<title>Organising A Museum Website (News Article Page) - Editing - Wagtail</title>
Describe alternatives you've considered
I have not considered any alternatives.
Additional context
I looked for related issues, but couldn't find any.
I think re-ordering the elements is a good idea.
However, you can already override the word 'Wagtail'
https://docs.wagtail.org/en/stable/advanced_topics/customisation/admin_templates.html#branding-title
Labelling accessibility. This would be a win for browser tab usability but also for better compliance.
Links
https://www.w3.org/WAI/WCAG21/Understanding/page-titled.html
https://www.w3.org/WAI/WCAG21/Techniques/general/G88
https://accessibility.iu.edu/creating-content/web-content/titles.html
Approach proposal
https://github.com/wagtail/wagtail/blob/520dfa8005d8b11580dc91c4ff625afddf9c60ad/wagtail/admin/templates/wagtailadmin/skeleton.html#L8
<title>{% block titletag %}{% endblock %} - {% block branding_title %}Wagtail{% endblock %} </title>
Should be an easy enhancement, labelling good first issue also.
I would try this. In the admin screen, I would type the following
def get_admin_display_title(self): #pulled from wagtail sourcecode in home models.py return
I would try this. In the admin screen, I would type the following def get_admin_display_title(self): #pulled from wagtail sourcecode in home models.py return
The return is going to be in another line and indented obv. I think I messed up the code in sending.
Resolved now - this is what an editing page looks like.
If we want to further refine the layout (move the 'page type') out to after the title - best that we cover that in a new issue.
| gharchive/issue | 2022-08-19T09:09:49 | 2025-04-01T06:46:12.501237 | {
"authors": [
"basic5253",
"lb-",
"timjmansfield"
],
"repo": "wagtail/wagtail",
"url": "https://github.com/wagtail/wagtail/issues/9054",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
501616673 | Update setup.py
remove white spaces for clear readability.
Thanks for contributing to Wagtail! 🎉
Before submitting, please review the contributor guidelines http://docs.wagtail.io/en/latest/contributing/index.html and check the following:
Do the tests still pass? (http://docs.wagtail.io/en/latest/contributing/developing.html#testing)
Does the code comply with the style guide? (Run make lint from the Wagtail root)
For Python changes: Have you added tests to cover the new/fixed behaviour?
For front-end changes: Did you test on all of Wagtail’s supported browsers? Please list the exact versions you tested.
For new features: Has the documentation been updated accordingly?
It looks like this user is running for a Hacktoberfest t-shirt by creating a bunch of useless PRs on several repositories on this new GitHub account.
My advice would be to close this PR and add the invalid tag as well. Hacktoberfest processes the invalid tag by invalidating the contribution the user has made and won't count toward Hacktoberfest.
More information about this:
https://hacktoberfest.digitalocean.com/details#spam
The other PR's:
https://github.com/django/django/pull/11864/files
https://github.com/getsentry/sentry/pull/14908/files
https://github.com/wagtail/wagtail/pull/5598/files
https://github.com/fabric/fabric/pull/2020/files
https://github.com/home-assistant/home-assistant/pull/27124/files
| gharchive/pull-request | 2019-10-02T17:10:20 | 2025-04-01T06:46:12.507532 | {
"authors": [
"MrAfrazKhan",
"frenck"
],
"repo": "wagtail/wagtail",
"url": "https://github.com/wagtail/wagtail/pull/5598",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1217805718 | Stream field docs to markdown
Changed all rst documents in docs/reference/streamfield to markdown. This was my first time contributing to the documentation and some of the rst didn't have an exact match to md but did my best to format it similarly (hope the links work as intended!).
To help with issue #8383
[x] Do the tests still pass?[^1] -- I've have had some problems running the codebase through vagrant VM on windows 11 but there were no tests found for the docs directory so presuming my changes will have no affect.
[x] Does the code comply with the style guide?
[x] Run make lint from the Wagtail root.
rest not applicable:
[ ] For Python changes: Have you added tests to cover the new/fixed behaviour?
[ ] For front-end changes: Did you test on all of Wagtail’s supported environments?[^2]
[ ] Please list the exact browser and operating system versions you tested:
[ ] Please list which assistive technologies [^3] you tested:
[ ] For new features: Has the documentation been updated accordingly?
[^1]: Development Testing
[^2]: Browser and device support
[^3]: Accessibility Target
Thanks for taking the time to contribute @benedictFaw - I will aim to get this reviewed soon and give some feedback.
Thanks for taking the time to contribute @benedictFaw - I will aim to get this reviewed soon and give some feedback.
| gharchive/pull-request | 2022-04-27T19:44:10 | 2025-04-01T06:46:12.513189 | {
"authors": [
"benedictFaw",
"lb-"
],
"repo": "wagtail/wagtail",
"url": "https://github.com/wagtail/wagtail/pull/8444",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
678834224 | Add license to docker images and job templates
extend models for docker_image and job_template to include a license field
docker_images.csv files have been extended already (using the name of the license)
methods iterate_docker_images and add_initial_docker_images in _docker.py have been adapted to work with the extended CSV files
Implemented in commit c7effb455d44c1c03754fc1e94e71af51c05971a
The python client still seems to be missing the license parameter for docker images and job templates.
Sorry forgot to push. Done now
| gharchive/issue | 2020-08-14T01:30:49 | 2025-04-01T06:46:12.543428 | {
"authors": [
"csterling",
"fracpete"
],
"repo": "waikato-ufdl/ufdl-backend",
"url": "https://github.com/waikato-ufdl/ufdl-backend/issues/45",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2203800611 | 🛑 cyconet.org is down
In 8101a3f, cyconet.org (https://www.cyconet.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: cyconet.org is back up in f220e2b after 2 hours, 1 minute.
| gharchive/issue | 2024-03-23T09:54:43 | 2025-04-01T06:46:12.554177 | {
"authors": [
"waja"
],
"repo": "waja/cyconet-upptime",
"url": "https://github.com/waja/cyconet-upptime/issues/1501",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2216684033 | 🛑 cyconet.org is down
In 1bd8a38, cyconet.org (https://www.cyconet.org) was down:
HTTP code: 0
Response time: 0 ms
Resolved: cyconet.org is back up in 25ef80b after 55 minutes.
| gharchive/issue | 2024-03-30T22:21:09 | 2025-04-01T06:46:12.557401 | {
"authors": [
"waja"
],
"repo": "waja/cyconet-upptime",
"url": "https://github.com/waja/cyconet-upptime/issues/1542",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1192669768 | feat: provide chunkName to iife
Feature Request
Pass the chunk name into iise in order to further provide additional context that can be used to determine the correct publicUrl.
Describe Preferred Solution
function determinePublicPath(entry, chunk) {
if (chunk !== "main") {
return `/services/${entry}/`;
}
return "/";
}
addWebpackPlugin(new DynamicContainerPathPlugin({
iife: determinePublicPath,
entry: projectName,
})),
Additional Context
I've forked and have a PR that I'd like to submit to resolve this issue. Please see: https://github.com/justinm/dynamic-container-path-webpack-plugin/tree/feature/chunk-based-paths
If the feature request is approved, would you be willing to submit a PR?
(Help can be provided if you need assistance submitting a PR)
[X] Yes
[ ] No
Looks good to me, I'll merge your PR when you're ready.
released via https://www.npmjs.com/package/dynamic-container-path-webpack-plugin/v/1.2.0
| gharchive/issue | 2022-04-05T06:17:11 | 2025-04-01T06:46:12.621421 | {
"authors": [
"justinm",
"waldronmatt"
],
"repo": "waldronmatt/dynamic-container-path-webpack-plugin",
"url": "https://github.com/waldronmatt/dynamic-container-path-webpack-plugin/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
133462944 | Update rules for ESLint 2.0.0
After upgrade to ESLint 2.0.0, there were these failures:
1:1 error Rule 'no-empty-label' was removed and replaced by: no-labels no-empty-label
1:1 error Rule 'space-after-keywords' was removed and replaced by: keyword-spacing space-after-keywords
1:1 error Rule 'space-before-keywords' was removed and replaced by: keyword-spacing space-before-keywords
1:1 error Rule 'space-return-throw-case' was removed and replaced by: keyword-spacing space-return-throw-case
These rules were removed.
This pull request aims to migrate configuration for removed rules.
Merge please.
Hi @lkrnac! Thanks so much for the PR - I dono what's going on but I didn't get an email for this. Sorry for the delay. I have a few comments though before I can merge it:
The package.json only gets bumped when a release is cut, not on every PR so that part will have to go.
Please rebase your commits if you could - I like to try to keep history as clean as I can.
Unfortunately, this PR is not sufficient to upgrade to ESLint 2.0 by itself but it's a great step and is a huge help to me so thanks a bunch for your work. I'll create a new issue to track the rest of it. Once you've made those two changes I'll merge this.
Thanks again.
Here's the comment that outlines the remaining work to get on ESLint 2.0. PRs are of course welcome but I'll be getting to it as I have time
https://github.com/walmartlabs/eslint-config-defaults/issues/69#issuecomment-189036324
@bear, OK, np. I closed this one and opened separate PR (https://github.com/walmartlabs/eslint-config-defaults/pull/71), without problematic commit.
Thanks
| gharchive/pull-request | 2016-02-13T19:23:47 | 2025-04-01T06:46:12.654833 | {
"authors": [
"Panda-ref",
"baer",
"lkrnac"
],
"repo": "walmartlabs/eslint-config-defaults",
"url": "https://github.com/walmartlabs/eslint-config-defaults/pull/68",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
69618490 | Auto-refresh: make it possible to add watched .js files
ScriptCraft could watch and auto-refresh .js files that a user is actively working with.
The user registers what files/folders to watch.
This is already available. See classroom.js and utils.js
| gharchive/issue | 2015-04-20T15:52:14 | 2025-04-01T06:46:12.660513 | {
"authors": [
"carlrobert",
"walterhiggins"
],
"repo": "walterhiggins/ScriptCraft",
"url": "https://github.com/walterhiggins/ScriptCraft/issues/237",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
474036739 | Suppress info messages
Weights and Biases version: 0.8.5
Python version: 3.7.3
Operating System: Linux
Description
WANDB prints a lot infos to the jupyter cell like:
INFO - wandb.run_manager - file/dir modified: /home/ubuntu/Projects/XX/XX/wandb/run-20190729_122935-r8x77nbl/wandb-summary.json
INFO - wandb.run_manager - file/dir modified: /home/ubuntu/Projects/XX/XX/wandb/run-20190729_122935-r8x77nbl/wandb-history.jsonl
INFO - wandb.run_manager - file/dir modified: /home/ubuntu/Projects/XX/XX/wandb/run-20190729_122935-r8x77nbl/model-best.h5
Is there a way to suppress these messages?
What I Did
Read the entire doc, but did not found something related
This shouldn't be happening in the most recent release. Is there anything special about your jupyter environment or instructions on how I could recreate it?
@vanpelt I am using W&B in my own pytorch pipeline and I am very particular with the output on my stdout. I really need to suppress all terminal outputs from W&B. Is there a kwarg or a global settings for the functions to suppress this?? Really need it.
Hey @AakashMallik are you running in Windows or Unix? Also are you always running python from the command line or within a jupyter kernel? If it's unix and the command line, you can disable stdout mirroing by adding this before your call to wandb.init:
from wandb import io_wrap
def redirect_noop(self):
pass
io_wrap.FileRedirector.redirect = redirect_noop
If it's one of the other scenarios, let me know and I'll share some code.
Hey @AakashMallik are you running in Windows or Unix? Also are you always running python from the command line or within a jupyter kernel? If it's unix and the command line, you can disable stdout mirroing by adding this before your call to wandb.init:
from wandb import io_wrap
def redirect_noop(self):
pass
io_wrap.FileRedirector.redirect = redirect_noop
If it's one of the other scenarios, let me know and I'll share some code.
Silly me. Just found this thread https://github.com/wandb/client/issues/337
Now I am setting the env variable from the pipeline itself. All good now. Thanks for your reply.
Hi, I have the same problem, and I'm running on ipython. I suppose it's the same as jupyter. I tried setting os.environ['WANDB_SILENT'] = 'true', but that did not work either. @vanpelt, you have some code to share that may solve the problem?
Actually, I found that the info can be suppressed following the instructions here: https://docs.wandb.com/library/integrations/jupyter#disable-info-messages-from-wandb-in-jupyter
We're working on a big new release that will address this issue for good. It's due out at the end of the month!
Hey folks,
This issue has been resolved in the present wandb package.
Closing this thread.
Please feel free to comment on this thread if the issue persists any longer. :smile:
| gharchive/issue | 2019-07-29T12:57:32 | 2025-04-01T06:46:12.671701 | {
"authors": [
"AakashMallik",
"ariG23498",
"gismo07",
"tshmak",
"vanpelt"
],
"repo": "wandb/client",
"url": "https://github.com/wandb/client/issues/457",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2299379083 | Using openui with AzureOpenai
I have Azureopenai Credentials and i was wondering if the openui can be used with Azureopenai
I'm using the native OpenAI client which I imagine supports it. You can just change the code here.
i tried changing the code as you mentioned ,but still facing a lot of errors
Thanks @vanpelt works perfectly for me with gpt-4o.
@NikitaRawat2003 the fastest way to try it, is to create on your azure instance a deployment that as the same name as the model name in openai ("gpt-3.5-turbo", "gpt-4o", "gpt-4-turbo") and then import from AsyncAzureOpenAI
and replace:
# openai = AsyncOpenAI(
# base_url=config.OPENAI_BASE_URL,
# api_key=config.OPENAI_API_KEY)
openai = AsyncAzureOpenAI(
api_key=config.OPENAI_API_KEY,
api_version="2024-02-01",
azure_endpoint=config.OPENAI_BASE_URL)
Thanks @vanpelt works perfectly for me with gpt-4o. @NikitaRawat2003 the fastest way to try it, is to create on your azure instance a deployment that as the same name as the model name in openai ("gpt-3.5-turbo", "gpt-4o", "gpt-4-turbo") and then import AsyncAzureOpenAI from openai and replace:
# openai = AsyncOpenAI(
# base_url=config.OPENAI_BASE_URL,
# api_key=config.OPENAI_API_KEY)
openai = AsyncAzureOpenAI(
api_key=config.OPENAI_API_KEY,
api_version="2024-02-01",
azure_endpoint=config.OPENAI_BASE_URL)
did you add your endpoint and key in the config.py file? can you show me how you did it?
I created a pull request for use Azure OpenAI.
https://github.com/wandb/openui/pull/216
| gharchive/issue | 2024-05-16T05:21:52 | 2025-04-01T06:46:12.676409 | {
"authors": [
"NikitaRawat2003",
"WangYiDFO",
"bxav",
"gyxter",
"vanpelt"
],
"repo": "wandb/openui",
"url": "https://github.com/wandb/openui/issues/128",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
930428463 | [CLI-975] make distinguishing between uniform and int_uniform distributions safe
When a user passes integermin/max values for a sweep parameter explicitly declared with distribution: "uniform" the backend throws warnings like:
wandb: WARNING To avoid this, please fix the sweep config schema violations below:
wandb: WARNING Violation 1. {'distribution': 'uniform', 'max': -18, 'min': -20} is not valid under any of the given schemas
wandb: WARNING Violation 2. {'distribution': 'uniform', 'max': 0.01, 'min': 0} is not valid under any of the given schemas
wandb: WARNING Violation 3. {'distribution': 'uniform', 'max': 0.05, 'min': 0} is not valid under any of the given schemas
wandb: WARNING Violation 4. {'distribution': 'uniform', 'max': -18, 'min': -20} is not valid under any of the given schemas
wandb: WARNING Violation 5. {'max': -18, 'min': -20, 'distribution': 'uniform'} is not valid under any of the given schemas
wandb: WARNING Violation 6. {'distribution': 'uniform', 'max': 0.01, 'min': 0} is not valid under any of the given schemas
wandb: WARNING Violation 7. {'distribution': 'uniform', 'max': -18, 'min': -20} is not valid under any of the given schemas
wandb: WARNING Violation 8. {'distribution': 'uniform', 'max': -18, 'min': -20} is not valid under any of the given schemas
wandb: WARNING Violation 9. {'distribution': 'uniform', 'max': 0.02, 'min': 0} is not valid under any of the given schemas
wandb: WARNING Violation 10. {'max': -6, 'min': -8, 'distribution': 'uniform'} is not valid under any of the given schemas
wandb: WARNING Violation 11. {'distribution': 'uniform', 'max': -18, 'min': -20} is not valid under any of the given schemas
This PR updates the schema to allow integer min/max when the distribution: uniform is explicitly specified
Fixes https://github.com/wandb/client/issues/2317.
Moving back to draft and will re-open for review once I have made the changes discussed in slack thread
| gharchive/pull-request | 2021-06-25T19:00:40 | 2025-04-01T06:46:12.679092 | {
"authors": [
"dannygoldstein"
],
"repo": "wandb/sweeps",
"url": "https://github.com/wandb/sweeps/pull/16",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1424868946 | [Q] question about matplotlib plot figure logging
Hello, I'm trying to log scatter figure for multiple speaker embeddings distribution like below.
But the list of label displayed like picture below. and marker displays trace_[NUM].
I want to change it to display marker's label when I touch it by mouse. Is there any way to change like I want?
def wandb_log_scatter(data, x_label="x1", y_label="x2"):
"""
data: Dict[[str, List[List[float]]]]
"""
colors = cm.rainbow(np.linspace(0, 1, len(data)))
fig = plt.figure(figsize=(16, 12))
pls = []
labels = []
for c, (label, values) in zip(colors, data.items()):
x = [v[0] for v in values]
y = [v[1] for v in values]
labels.append(label)
pls.append(plt.scatter(x, y, color=c))
plt.legend(pls, labels)
return wandb.log({"Scatter Output": fig})
Thank you.
Hi @LEECHOONGHO,
The best way to get an interactive chart of this form would be to use a Custom Chart. This uses Vega, a visualization grammar built to generate interactive charts.
Hi @LEECHOONGHO, We wanted to follow up with you regarding your support request as we have not heard back from you. Please let us know if we can be of further assistance or if your issue has been resolved.
Hi @LEECHOONGHO, since we have not heard back from you we are going to close this request. If you would like to re-open the conversation, please let us know!
| gharchive/issue | 2022-10-27T00:51:59 | 2025-04-01T06:46:12.683097 | {
"authors": [
"LEECHOONGHO",
"ramit-wandb"
],
"repo": "wandb/wandb",
"url": "https://github.com/wandb/wandb/issues/4415",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1956084011 | [App]: "No selected runs have saved code." when attempting to visualize code
Current Behavior
When I add a "Code" panel to my workspace, I see the following:
As you can see at this link, there is in fact code saved for this run.
As you can see here, I have enabled code saving for my profile:
Expected Behavior
I should be able to visualize some code with this panel.
Steps To Reproduce
Click "add panel"
Choose "Code"
Screenshots
No response
Environment
OS: macos
Browsers: Chrome
Version: 118.0.5993.96
Additional Context
No response
Hi @ethanabrooks ,
Thank you for reaching out for support. I'll be glad to assist you with this. we'll investigate and get back to with updates.
Regards,
Carlo Argel
Hi @ethanabrooks ,
Upon checking, the "Enable code saving by default" is disabled/untoggled for your teams. You will need to toggle it for you to able to save for your teams.
regards,
Carlo Argel
Hi @Carlo-Argel, I've been having the same issue. I have code artifacts saved, and code saving is enabled in my personal account, but the code panel displays "no selected runs have saved code". I don't have any teams, and I'm not part of any teams in wandb.
I switched on "Enable code saving by default" in my teams settings, but I still get the same issue.
Hi @ethanabrooks , assisting my colleague Carlo further on this. Are you still having issues with saving your code? I checked the wandb teams you are a member of, https://wandb.ai/l2m/settings, https://wandb.ai/gpt-rl/settings and neither of these teams has code saving enabled.
@poltimmer, I checked one of your projects and it appears code saving is now working. If you still have issue do let us know.
Hi Mohammad. I am running all of my current projects in https://wandb.ai/rldl where I have code saving enabled. Those other two teams are not in use at the moment. I am able to view my main.py file for this run:
But not for this run:
Also, I was hoping to actually view my config file (which I am logging using run.loc_code, not my main.py file, but perhaps this functionality has not been implemented.
I confirm. Code is presented in artifacts, but dedicated panel in workspace shows an error "No selected runs have saved code.".
@ethanabrooks , to compare run configs between runs, utilize a run comparer panel. Additionally, I was able to see what you do now with the code saving. Do you have the debug.log and debug-internal.log located in your working directory under the wandb folder available to review? If yes please send them to support@wandb.com and include in subject "Attention Mohammad: Wandb github issue #6485". I suspect the run did not successfully upload the code files prior to crashing. These files will help us confirm.
@MBakirWB unfortunately I deleted my wandb/ directory since launching that run. Since then, my runs have been displaying my main.py file so perhaps the issue is resolved. Thanks for suggesting the "run comparer" -- that is a tool that I will use in the future. That said, my goal in displaying the config file is not to compare configs between runs (the reasons are particular to my setup, so I won't detail them here). If there is any way to display this code in the workspace, let me know. Otherwise, it looks like I can access it by clicking on artifacts.
Hi @ethanabrooks , thank you for confirming that you run files are not appearing. If this issue does happen again please let me know and do send over those debug files for review. Additionally, you can utilize a weave panel to display the run configurations of all/specific runs. See example video below. Considering this resolved but do let me know if you have any other questions.
https://github.com/wandb/wandb/assets/104453360/adbd1625-b021-48de-b76d-f49f72b9550d
@MBakirWB thank you for the suggestion. The goal of displaying the config file is actually not to see what parameters I used but to be able to easily copy/paste the logic used to produce the config. This may lay outside the scope of this issue though.
Hi @ethanabrooks , thanks for the additional clarification. This will have to be a feature request. I've recorded one in our system and will keep you updated once there has been movement on it. Do write back in please if there is anything else we could assist you with.
Perfect! Thank you.
WandB Internal User commented:
ethanabrooks commented:
Hi Mohammad. I am running all of my current projects in https://wandb.ai/rldl where I have code saving enabled. Those other two teams are not in use at the moment. I am able to view my main.py file for this run:
But not for this run:
Also, I was hoping to actually view my config file (which I am logging using run.loc_code, not my main.py file, but perhaps this functionality has not been implemented.
WandB Internal User commented:
ethanabrooks commented:
@MBakirWB unfortunately I deleted my wandb/ directory since launching that run. Since then, my runs have been displaying my main.py file so perhaps the issue is resolved. Thanks for suggesting the "run comparer" -- that is a tool that I will use in the future. That said, my goal in displaying the config file is not to compare configs between runs (the reasons are particular to my setup, so I won't detail them here). If there is any way to display this code in the workspace, let me know. Otherwise, it looks like I can access it by clicking on artifacts.
| gharchive/issue | 2023-10-22T22:41:55 | 2025-04-01T06:46:12.701395 | {
"authors": [
"Carlo-Argel",
"MBakirWB",
"PgLoLo",
"ethanabrooks",
"poltimmer",
"sydholl"
],
"repo": "wandb/wandb",
"url": "https://github.com/wandb/wandb/issues/6485",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2723714006 | Add frontend for search
Description
Fixes WB-NNNNN
Fixes #NNNN
What does the PR do? Include a concise description of the PR contents.
Testing
How was this PR tested?
[!WARNING]
This pull request is not mergeable via GitHub because a downstack PR is open. Once all requirements are satisfied, merge this PR as a stack on Graphite.
Learn more
#3165 👈 (View in Graphite)
#3164
master
This stack of pull requests is managed by Graphite. Learn more about stacking.
| gharchive/pull-request | 2024-12-06T18:50:55 | 2025-04-01T06:46:12.707013 | {
"authors": [
"gtarpenning"
],
"repo": "wandb/weave",
"url": "https://github.com/wandb/weave/pull/3165",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
55127825 | stop codis-proxy and immediately start codis-proxy and set proxy online
the codis-proxy process is dead.
one error in log:
2015/01/22 16:24:47 main.go:106: [info] running on 0.0.0.0:19000
2015/01/22 16:24:47 router.go:567: [info] &{proxyId:proxy_1 productName:test zkAddr:localhost:2181 f:}
2015/01/22 16:24:47 router.go:589: [info] proxy_info:{Id:proxy_1 Addr:node3:19000 LastEvent: LastEventTs:0 State:offline Description: DebugVarAddr:node3:11000}
2015/01/22 16:24:47 router.go:560: [fatal] zk: node already exists
wait long time start codis-proxy and set proxy online is ok.
Default expire time is 30 seconds on zookeeper.
have some method to set session invaild?
I write a LSB bash script to start codis-proxy
nop, you could remove it manually by using zookeeper client
ok,thank you.
| gharchive/issue | 2015-01-22T08:44:26 | 2025-04-01T06:46:12.714868 | {
"authors": [
"ngaut",
"wu0407"
],
"repo": "wandoulabs/codis",
"url": "https://github.com/wandoulabs/codis/issues/81",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2230327059 | 我想在工具栏中加入历史记录浏览功能 但是显示历史记录脑图后 主编辑脑图就不能使用了 我应该怎么修改
https://github.com/wanglin2/mind-map/assets/70259086/d06da165-d358-4d45-9e01-65adc73caf01
代码截一下,看视频我也不知道为啥
是个bug,下个版本会修复
已修复,v0.9.11+生效,也可以拉取feature分支获取最新的代码
| gharchive/issue | 2024-04-08T05:59:14 | 2025-04-01T06:46:12.725812 | {
"authors": [
"wanglin2",
"yangda666"
],
"repo": "wanglin2/mind-map",
"url": "https://github.com/wanglin2/mind-map/issues/604",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1706567546 | [FEATURE] add/remove/list actor claims with wash
Note that this should probably not start until #538 has been decided upon and implemented.
It would be nice to be able to add claims/capabilities to an actor's wasmcloud.toml file without having to edit it directly. Less chance for squishy human errors.
wash claims add <name>
wash claims list
wash claims remove <name>
As a stretch goal, autocomplete for the argument would be nice. Could suggest official wasmcloud contracts for add and list existing claims for remove
Just tagging #538 so we don't forget, I think I proposed removing claims but it would remove some functionality, so I'd be fine to add that back in.
Could possibly go under the config section? Or if that is specifically for wash config, then maybe a new section specific to actors/contracts/interfaces?
Off-the-cuff: wash mod? (Short for "modify"; almost all the top-level commands are verbs now.) Or wash edit?
Some potential examples:
wash mod path/to/wasmcloud.toml (open the file in the editor)
wash mod add wasmcloud:httpserver (add to wasmcloud.toml in cwd)
wash mod add --bin path/to/provider/binary --archive path/to/pargz (add to par.gz)
As an example, kustomize uses edit:
Edits a kustomization file
Usage:
kustomize edit [command]
Examples:
# Adds a configmap to the kustomization file
kustomize edit add configmap NAME --from-literal=k=v
# Sets the nameprefix field
kustomize edit set nameprefix <prefix-value>
# Sets the namesuffix field
kustomize edit set namesuffix <suffix-value>
Available Commands:
add Adds an item to the kustomization file
alpha-list-builtin-plugin [Alpha] List the builtin plugins
fix Fix the missing fields in kustomization file
remove Removes items from the kustomization file
set Sets the value of different fields in kustomization file
Flags:
-h, --help help for edit
Global Flags:
--stack-trace print a stack-trace on error
Use "kustomize edit [command] --help" for more information about a command.
Also I think this is one of those areas where having verbs go first is a weakness -- It's a bit unclear what you're editing, where something like wash cfg set ... is a little clearer.
That said, wash add claims could work, as well as wash mod .../wash edit ...
Popping in here, I was developing today and had an awesome time with cargo add wasmcloud-interface-httpclient just to find that my next invocation was blocked because I didn't have the capability. Would be cool to include that capability if we're already intentionally adding the crate!
Some thoughts from the other issue (#856) related to claim mgmt, which was trying to enable users to:
List existing claims on a built & signed wasm module
Add a claim to an existing built & signed module
Remove a claim from a built & signed module
As for the additional context (for anyone who might start this effort):
The code for wash inspect is a great place to reference
Consider using an example like echo-wasi-preview2 for testing the build process. Note that it has one claim which is for the httpserver contract
| gharchive/issue | 2023-05-11T21:15:04 | 2025-04-01T06:46:12.823749 | {
"authors": [
"brooksmtownsend",
"connorsmith256",
"lachieh",
"vados-cosmonic"
],
"repo": "wasmCloud/wash",
"url": "https://github.com/wasmCloud/wash/issues/553",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1837288797 | Feat/update actor
Feature or Problem
This PR adds the ability to update/replace running actors with an atomic command
Related Issues
Resolves https://github.com/wasmCloud/wasmCloud/issues/374
Release Information
Next
Consumer Impact
Users can now use wash ctl update actor with the Rust host
Testing
Built on platform(s)
[ ] x86_64-linux
[ ] aarch64-linux
[ ] x86_64-darwin
[ ] aarch64-darwin
[ ] x86_64-windows
Tested on platform(s)
[ ] x86_64-linux
[ ] aarch64-linux
[ ] x86_64-darwin
[ ] aarch64-darwin
[ ] x86_64-windows
Unit Test(s)
Acceptance or Integration
I didn't write a new integration test. In order to test this manually (see below), I relied on having two actors built with the same public key but different implementation details. So in order to include an automated test of this fashion, I would need to add two test fixture actors and reproduce the manual test below. I think we'll want that eventually, but for now I'm okay leaving it out.
Manual Verification
wash ctl start actor -x default --host-id NDYEPXYQTKD3GAFCCCU2HKCC3FEXW23WH4LPX2LSOSHDNS4REUXDJYAI oci://localhost:5000/v2/hello1:0.1.0
>>> ⢀⠀ Starting actor oci://localhost:5000/v2/hello1:0.1.0 ...
Actor [MBLRZ3JJ7GVEYHZ6T5VLHZG4WJAEO26YOQBBKG6HEPHNUBZ2P7MAKYKR] (ref: [oci://localhost:5000/v2/hello1:0.1.0]) started on host [NDYEPXYQTKD3GAFCCCU2HKCC3FEXW23WH4LPX2LSOSHDNS4REUXDJYAI]
curl localhost:8080/echo
Hello, World 1!%
wash ctl update actor NDYEPXYQTKD3GAFCCCU2HKCC3FEXW23WH4LPX2LSOSHDNS4REUXDJYAI MBLRZ3JJ7GVEYHZ6T5VLHZG4WJAEO26YOQBBKG6HEPHNUBZ2P7MAKYKR oci://localhost:5000/v2/hello2:0.1.0
Actor MBLRZ3JJ7GVEYHZ6T5VLHZG4WJAEO26YOQBBKG6HEPHNUBZ2P7MAKYKR updated to oci://localhost:5000/v2/hello2:0.1.0
curl localhost:8080/echo
Hello, World 2!%
@rvolosatovs next time can you let me decide when to merge my PR? It's technically part of our PR lifecycle rules
@rvolosatovs next time can you let me decide when to merge my PR? It's technically part of our PR lifecycle rules
Sure - I did not expect an approved "ready-to-review" PR not to be ready for merge
| gharchive/pull-request | 2023-08-04T20:17:44 | 2025-04-01T06:46:12.831255 | {
"authors": [
"connorsmith256",
"rvolosatovs"
],
"repo": "wasmCloud/wasmCloud",
"url": "https://github.com/wasmCloud/wasmCloud/pull/439",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
506093497 | wasm-terminal: Android keyboard blur, jumps terminal history to top
On blur, we should call this.xterm.scrollToBottom(). 😄
I don't believe this is relevant now that we've rewritten the package as @wasmer/sdk.
| gharchive/issue | 2019-10-11T23:17:47 | 2025-04-01T06:46:12.832553 | {
"authors": [
"Michael-F-Bryan",
"torch2424"
],
"repo": "wasmerio/wasmer-js",
"url": "https://github.com/wasmerio/wasmer-js/issues/118",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
806790963 | [Decompiler] expression sequences in function arguments are ugly
Imagine you write code like this:
(func (begin <expr1> <expr2>))
or anything where a function argument's value involves multiple expressions that can't be combined, but aren't grouped in a single parent value form. This isn't very common, and I've only seen it where there was likely a macro in a function argument with a let in that macro.
The decompiler won't attempt to group it up like this. Originally it did, but it was a huge mess and turned things into one giant terrifying expression that was hard to read. I still think is the right choice by default and works in 99% of cases, but you have some weird things like:
(set! t9-3 format)
(set! a0-4 (quote #t))
(set! a1-3 "~Tparent: ~A~%")
(set! v1-0 (-> arg0 parent))
(t9-3 a0-4 a1-3 (if v1-0 (-> v1-0 0 self)))
which is actually
(format #t "blah" (let ((v1-0 (-> arg0 parent))) (if v1-0 (-> v1-0 0 self))))
There's a few approaches we could take:
Ignore it for now. The code works and this happens rarely so far.
Have a mode "try harder than usual to make this into a single expression" and use it for function calls.
Revisit once the decompiler inserts lets and see if we can be smarter about this.
Right now I'm leaning toward "ignore", unless we run into a bunch of these.
Seems very rare and would require a huge amount of effort to correctly detect and fix.
| gharchive/issue | 2021-02-11T22:24:14 | 2025-04-01T06:46:12.847784 | {
"authors": [
"water111"
],
"repo": "water111/jak-project",
"url": "https://github.com/water111/jak-project/issues/254",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
} |
1798088180 | [Planting Flow - Design] Upload Photo on Add Info
will need to add a photo upload to the planting feature where the information input is. will reuse the current upload photos design plus a carousal or a list of photos uploaded.
Will need to consider what is the primary image and if we want to give it a description and title per image or just upload an image and worry about the details later.
will work on figma design
Add upload component to the information portion of the planting feature. not sure where exactly the right spot for the upload is. is it better to start with it or after inputting the species name and other stats about the tree?
Here are the two examples.
@zoobot which of the two do you prefer or do you have a different solution?
view of uploaded photos with the ability to upload more and delete photos.
example of how the ui would look for the upload image.
mobile preview
| gharchive/issue | 2023-07-11T04:48:49 | 2025-04-01T06:46:12.864094 | {
"authors": [
"ri0nardo"
],
"repo": "waterthetrees/wtt_front",
"url": "https://github.com/waterthetrees/wtt_front/issues/729",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
208330260 | With Cisco Spark
Hi, I'm Kotaro. I wanna use this plugin with Cisco Spark but It seems "unsupported" yet. If possible, please give me a solution. I wanna know how I can use. Which part should I change??
I've done. Never mind.
| gharchive/issue | 2017-02-17T04:10:41 | 2025-04-01T06:46:12.866145 | {
"authors": [
"arootkm"
],
"repo": "watson-developer-cloud/botkit-middleware",
"url": "https://github.com/watson-developer-cloud/botkit-middleware/issues/35",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
299862040 | Revamp the UI of the top bar
I have revamped the UI of the top bar to make it more visually appealing as can be seen in the screenshots below.
I made the following changes:
Put the logo into centre and removed "Wave in a box" text.
Put "Online" and "Saved" status indicators underneath the address of the user.
Put the Language option and the Sign-in button into a jquery UI Settings dialog.
Removed the border on the top bar.
Changed the background colour of the main app.
I like the goal of improving the top bar, but I feel like the gear icon should open a menu in the corner, just below the icon. It would better visually connect the dialog to the action the user just took, and it would be more consistent with well-known apps/sites like Google, Microsoft, Facebook, etc.
Also, does this version still allow for a server title?
Great changes in general, however:
We need to try and maximize the vertical screen space to allow more space for the UI.
I would like to retain the red/green color on the status, as the user should be aware when he is working offline. All offline changes will be lost if the wave is closed before saved.
@ZMYaro I have updated the Settings dialog so it appears underneath the settings cog but without the title bar (See picture below). What do you mean by a server title?
@vega113 I changed it so that the colour for the 'Online' indicator is grey as this is the normal status but when it is Offline it still appears Red and Bold (See picture below)
I have noticed an issue when you first press the cog. The issue is that the dialog displays over the cog, stopping the ability to close it (See screenshot). I will try and solve the issue.
Now I think that it's an issue with $( "#dialog *" ).css("display","unset"); and that calling it before the creation of the dialog will fix it.
Doesn't seem to work with IE11.
Problem with Firefox. Overflow bars shown and Language box too wide.
Do you think it is ready to be merged?
I just realized this PR adds jQuery, which was not present before. Can you not use the dialog framework already present in WIAB?
@ZMYaro Sorry for the delay, I aggree and I will now try to remove JQuery and JQueryUI.
@ZMYaro I have now made the changes.
Do you think it is ready to be merged?
Sorry for the delay, I will check the changes and merge.
LGTM
Thanks
| gharchive/pull-request | 2018-02-23T21:52:37 | 2025-04-01T06:46:12.875575 | {
"authors": [
"PeterWarrington",
"ZMYaro",
"vega113"
],
"repo": "wave-protocol/wave",
"url": "https://github.com/wave-protocol/wave/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
131407267 | Frame Rate Issue
If the frame rate is lowered below 24, a Broken pipe error happens.
Hmm, I can't test right now but for when I get a second, what framerate are you trying? (and I assume you're modifying the framerate by tweaking the FRAMERATE= line at the top of server.py rather than anywhere else?)
Yes. I set it to several, but specifically I'm trying somewhere in the 12-15 range. I am using the broadcasting thread to save the data to a file as well and to keep the file size down I want to reduce the frame rate. Bit rate is more important than frame rate in my case. I did look in the jsmpg.js file and noticed the PICTURE_RATE array does not go below 23.976, so I assume that is the limiting factor.
Ah, found it - avconv bombs out with the following error (which you don't see because its output is suppressed):
[mpeg1video @ 0x10e7c20] MPEG1/2 does not support 15/1 fps
The above was when I attempted 15fps but you get much the same for anything below 24fps. I don't know the MPEG1/2 standards well enough to know what frame rates are supported, but I'd hazard a guess that they bottom out at 24fps from what you've described.
I've found a usable way around it (at least for my needs). I put an if statement around the broadcast output write to basically do it every 2nd time (increment an integer or alternate a boolean) and also cut the buffer read size in half in the broadcast thread. This effectively records and displays at 12 fps. The video playback speed is 2x though since it still plays at 24 fps. So 1 minute of recording makes a 30 second video. For my purposes this is fine since the video size is reduced by half and it makes it easier to sift through hours of footage as it runs faster.
Heh - well, it's a cunning method! Glad you found something that'll work for you.
| gharchive/issue | 2016-02-04T16:49:27 | 2025-04-01T06:46:12.888435 | {
"authors": [
"doogerk",
"waveform80"
],
"repo": "waveform80/pistreaming",
"url": "https://github.com/waveform80/pistreaming/issues/5",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
603086928 | "Cannot use import statement outside a module" when using require
After installing local-echo as a node module using npm install, I tried to import local-echo using require within an electron app but encountered this error:
var LocalEchoController = require('local-echo')
Exception thrown in developer console:
Uncaught D:\electron-app\node_modules\local-echo\index.js:1
(function (exports, require, module, __filename, __dirname, process, global, Buffer) { return function (exports, require, module, __filename, __dirname) { import LocalEchoController from "./lib/LocalEchoController";
^^^^^^
SyntaxError: Cannot use import statement outside a module
at new Script (vm.js:84:7)
at createScript (vm.js:258:10)
at Object.runInThisContext (vm.js:306:10)
at Module._compile (internal/modules/cjs/loader.js:884:26)
at Object.Module._extensions..js (internal/modules/cjs/loader.js:986:10)
at Module.load (internal/modules/cjs/loader.js:816:32)
at Module._load (internal/modules/cjs/loader.js:728:14)
at Module._load (electron/js2c/asar.js:717:26)
at Function.Module._load (electron/js2c/asar.js:717:26)
at Module.require (internal/modules/cjs/loader.js:853:19)
Can you fix this error? Or at least provide a solution on how to fix this error? Thanks
Hello, I also encountered a similar problem. Import cannot be used in Es5, and the exception similar to that reported by you cannot be used with require
| gharchive/issue | 2020-04-20T09:29:52 | 2025-04-01T06:46:12.892166 | {
"authors": [
"fisyher",
"zhubinsheng"
],
"repo": "wavesoft/local-echo",
"url": "https://github.com/wavesoft/local-echo/issues/20",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
61265417 | Smarty does a poor job at German apostrophes
Configuring it with the correct values for 'left-single-quote' and 'right-single-quote', it gives wrong apostrophes, for example:
"O'Reilly" → "O‘Reilly" (correct is "O’Reilly")
"Alice'" → "Alice‘" (correct is "Alice’")
{
'markdown.extensions.smarty': {
'substitutions': {
'left-single-quote': '‚',
'right-single-quote': '‘',
'left-double-quote': '„',
'right-double-quote': '“'
}
}
}
So you should be able to tell smarty: I want apostrophes instead of right-single-quotes.
Thank you for your report.
Unfortunately, I am unsure of the exact problem you are reporting. Is this connected to #397 or is that separate? It occurs to me that if your were passing in the configs incorrectly due to the documentation error, that that was causing the wrong type of quote to be used. If so, then this may not actually be an issue and can be closed.
If, on the other hand, this is an actual separate issue, all I can say is that I can't possibly factor in every edge case for every language out there. It is not clear to me if you need to set a different substitution in the configs or if the logic which selects the type of quote needs to be improved for your specific needs. If it is the later, my inclination is to suggest that the extension should be forked and maintained as a language specific fork that meets all of the nuances of that specific language.
In fact, the existing support for substitutions was provided by a third party contributor who also provided the (apparently broken) documentation. That person informed me that the example in the docs worked for German. Personally, I have no knowledge of German or what requirements that language has. If the issue is that we need better docs, then I am happy to accept patches (although I don't what to document German (or any language's) syntax rules). However, as stated above, anything except very minor code changes (which benefit users of all languages) are best addressed with a fork maintained by someone who understands the specific needs of that fork.
Taking another look at your report, it would appear that you require apostrophes and right single quotes to be two separate and different characters. In English, those two uses generally call for the same character. As there is no difference in practice, the code does not differentiate between them. Presumably, a whole additional feature would need to be added to the extension to support apostrophes. I suppose the English default could be for "apostrophes" and "right-single-quotes" to have the same character assigned to them and offer the option to provide independent substitutions for each for other languages. That would certainly be an addition that could potentially support multiple languages.
That said, I have neither the time, the need, nor the interest in working on this myself. Of course, patches are welcome.
I think we just need to make closingSingleQuotesRegex2 produce an apostrophe, not a single quote.
However this way not all language problems will be solved, i.e. that regex contains a case for English-specific 's ending.
After looking at it again, it is not that simple — that regex matches quotes on word endings and the cases like Paul's.
I know a bit of German, but I am not an expert — are all use cases for apostrophe limited by 's, or are there other ones? If there are no other cases, then I will just split that regex into two regexes and make one return an apostrophe instead of closing-single-quote.
I think that is a reasonable approach to take if we are going to change anything. I would suggest that the "apostrophe" regex cover the uses discussed by Wikipedia. Don't forget about non-English Proper Names, which are buried deep in that page. I would think that would mostly be pretty straightforward (eg: "\w'\w"). However, a couple issues may arise with plural years (eg: the 1990's or the '90's) and nouns ending in an "s" (or "s" or "z" sound) where a second "s" is not added after the the apostrophe (whether it is plural or not).
The non-plural ending-in-s is a rare case. But even the common plural case can be tricky. Consider these examples:
"He stated that they were his brother's tools." (the tools belong to one brother)
"He stated that they were his 'brother's' tools." (same as 1 except that 'brother' is in quotes because it is a title -- perhaps not his 'real brother')
"He stated that they were his brothers' tools." (the tools belong to multiple brothers -- this is an apostrophe not a right single quote)
"He stated that they were his 'brothers'' tools." (same as 2 with multiple 'brothers' (as a quoted title) - this is an apostrophe followed by a right single quote)
It is not clear to me how to consistently determine which is an apostrophe and which is a right quote in every one of those cases with a regex. If it is not reasonably possible, then I say we should not bother making any change.
I know a bit of German, but I am not an expert — are all use cases for apostrophe limited by 's, or are there other ones?
German use cases:
Alice’ Abenteuer im Wunderland
Max’ und Moritz’ Scherze
Johannes’ Aufgabe
Dern’sches Gelände
Andrea’s Blabla (actually a “greengrocer’s apostrophe”)
Wie geht’s?
Ku’damm
’ne Kippe
Kommen S’ nur herein!
mit’m Fahrrad fahr’n
Might not cover all of them.
@gandaro Thanks! In half of your examples the apostrophe is at the beginning (or end) of the word, so it will be impossible to distinguish it from opening/closing quote...
That's what I expected. Therefore I am closing this as wontfix.
| gharchive/issue | 2015-03-13T21:57:28 | 2025-04-01T06:46:12.914531 | {
"authors": [
"gandaro",
"mitya57",
"waylan"
],
"repo": "waylan/Python-Markdown",
"url": "https://github.com/waylan/Python-Markdown/issues/396",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
240419795 | Weird treatment of trailing HTML
The following input
<!DOCTYPE html>
<html><body>
some text here
</body></html>
when run through the CLI creates the output
<!DOCTYPE html>
<html><body>
<p>some text here</p>
<p></body></html></p>
Obviously, the <p> tags around the trailing HTML code are wrong. Weirdly, if the leading and trailing HTML code are a simple <div> tag, these superfluous <p> tags are not produced.
For reference, the Python CommonMark implementation does the right thing here.
The problem you are seeing is specifically because Python Markdown doesn't handle body or html. It's supposed to not convert content between block tags, but it doesn't account for body or html.
Usually people are running content through and inserting it in HTML (or basically wrapping body and html around it after conversion), not usually running a fully formed HTML document through the parser. You'll notice Python Markdown never wraps content in body and html when it outputs content. It really is a content parser...
With that said, maybe it should at least account for html and body so it doesn't surprise people. If @waylan agrees, I wouldn't mind issuing a pull request, but if not, my best advice is don't feed your content through the parser wrapped in html and body.
But why does it work for the opening html and body tags, but not the closing ones?
Also I thought that Markdown is intended to leave any inline HTML alone, which is why I thought the above behavior is a bug.
I haven't looked at the code for specifics, but content between blocks should not get parsed. In your example, the content and the closing tags are getting parsed. This is telling me the html blocks html and body aren't getting handled like block tags. Most likely they are getting handled like inline tags, but since you have empty new lines after the last opening tag and before the first closing tag, the opening and closing tags are in different paragraph blocks breaking the inline tag treatment.
Basically, html and body are probably being treated as inline tags, but are spanning across multiple paragraphs which breaks parsing.
@facelessuser is correct. The html and body tags are seen as inline tags. The reason the opening tags are not wrapped in a p element is because of the doctype. If, instead, you do this:
<!DOCTYPE html>
<html><body>
some text here
</body></html>
Then you get the following output:
<!DOCTYPE html>
<p><html><body></p>
<p>some text here</p>
<p></body></html></p>
And if you provide this input:
<!DOCTYPE html>
<html><body>
some text here
</body></html>
You get this output:
<!DOCTYPE html>
<p><html><body>
some text here
</body></html></p>
Notice that with no blank lines, the parser sees the entire block as a single paragraph. This is the confirmation that html and body tags are not considered block tags. And this is correct behavior as Markdown should only ever be content between body tags. To include a body (or its html parent) would generate invalid HTML. This is the Garbage-in, garbage-out principle. You can't expect valid output if you give it invalid input.
Note that the reference implementation (markdown.pl) follows the same behavior with one exception. It does not consider the doctype to be a block level element and also wraps it in a p element. If there is any bug here, it is that we recognize the doctype as a block level element. But personally, I'm fine leaving that as-is.
Note that this is a Markdown implementation which follows these rules. It is NOT a CommonMark implementation, nor will it ever be (CommonMark is not Markdown).
The document you cited states this:
Note that Markdown formatting syntax is not processed within block-level HTML tags. E.g., you can’t use Markdown-style emphasis inside an HTML block.
So the observed behavior is still a bug because the inner text within the body tag is treated as Markdown, which it shouldn't be according to the spec you cited.
Another, more relevant example: the input
<div>
some **text** here
</div>
turns into
<div>
some **text** here
</div>
Clearly, div is a block tag, so the inner text should not be parsed as Markdown.
That depends on what is considered a "block-level HTML tag". As the reference implementation does not consider body or html to be block level, we don't either. And we will not change until the reference implementation does and/or the rules are updated.
| gharchive/issue | 2017-07-04T13:23:16 | 2025-04-01T06:46:12.927634 | {
"authors": [
"c-f-h",
"facelessuser",
"waylan"
],
"repo": "waylan/Python-Markdown",
"url": "https://github.com/waylan/Python-Markdown/issues/572",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2236010818 | systemctl start wazuh-indexer not working
I ve installed Wazuh on my virtuell Debian environment but when i give the last command systemctl start wazuh-indexer, i take the error in picture which i added.
Does anyone has any idea?
thanks
Have you checked the logs as suggested?
Yes but was not able to identify the error
[
]
@AlexRuiz7 Can you help
Check the service logs with on journald
journalctl --no-pager --since yesterday -xeu wazuh-indexer
Also, check the logs at /var/log/wazuh-indexer/wazuh-cluster.log
grep -iE "error" /var/log/wazuh-indexer/wazuh-cluster.log
Closed due to inactivity.
| gharchive/issue | 2023-07-11T11:45:15 | 2025-04-01T06:46:12.957229 | {
"authors": [
"AlexRuiz7",
"amarkhade96",
"orhancevik64"
],
"repo": "wazuh/wazuh-indexer",
"url": "https://github.com/wazuh/wazuh-indexer/issues/203",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
239493739 | New features to next version
Some features planned next:
Returning length of steps on back and next callback;
Support to custom navigator, instead use our own;
Adding a progress bar;
Removing dependency of use a <form/>;
when do you want to release a new update?
Hi @bypit , what kind of feature do you need?
| gharchive/issue | 2017-06-29T14:14:54 | 2025-04-01T06:46:13.062600 | {
"authors": [
"bypit",
"eduardojmatos",
"wbotelhos"
],
"repo": "wbotelhos/stepy",
"url": "https://github.com/wbotelhos/stepy/issues/50",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
246228866 | how to save the chart in local?
how to save the chart in local?
it's simple:
// your graph
file, _ := os.Create("demo.png")
_ = graph.Render(chat.PNG, file)
Render's second parameter is an io.Writer
| gharchive/issue | 2017-07-28T03:35:03 | 2025-04-01T06:46:13.071870 | {
"authors": [
"136025747",
"qgymje"
],
"repo": "wcharczuk/go-chart",
"url": "https://github.com/wcharczuk/go-chart/issues/49",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
259032858 | SVG: title is incorrectly positioned/rotated when using YAxis
When I added a YAxis Name to my chart it caused the title to be rotated and positioned weirdly. This only happens while using the charts.SVG renderer. Switching to charts.PNG fixes the issue.
Example file: https://fn.lc/s/0.svg
NameStyle: chart.StyleShow()
No NameStyle
Relevant Code
graph := chart.Chart{
Title: fmt.Sprintf("%s vs. %s", gp.Name(dim), gp.OutputName()),
TitleStyle: chart.StyleShow(),
XAxis: chart.XAxis{
Name: gp.Name(dim),
NameStyle: chart.StyleShow(),
Style: chart.StyleShow(),
},
YAxis: chart.YAxis{
Name: gp.OutputName(),
NameStyle: chart.StyleShow(), // Commenting this line fixes the issue.
Style: chart.StyleShow(),
},
Background: chart.Style{
Padding: chart.Box{
Top: 20,
Left: 20,
Bottom: 20,
Right: 20,
},
},
}
graph.Elements = []chart.Renderable{
chart.Legend(&graph),
}
graph.Series = append(
graph.Series,
chart.ContinuousSeries{
Name: "Mean",
XValues: x,
YValues: means,
},
chart.ContinuousSeries{
Name: "+1σ",
XValues: x,
YValues: uppers,
},
chart.ContinuousSeries{
Name: "-1σ",
XValues: x,
YValues: lowers,
},
)
graph.Series = append(
graph.Series,
chart.ContinuousSeries{
Name: "Known",
XValues: knownX,
YValues: knownY,
Style: chart.Style{
Show: true,
StrokeWidth: chart.Disabled,
DotWidth: 5,
},
},
)
if err := graph.Render(chart.SVG, w); err != nil {
return err
}
This also appears to be the same issue at https://github.com/wcharczuk/go-chart/issues/44, except #44 doesn't mention this is incorrect behavior.
Any work on this (seeing as it has been 3 years)? Having the same issue, would love to have this fixed.
| gharchive/issue | 2017-09-20T03:59:31 | 2025-04-01T06:46:13.075870 | {
"authors": [
"d4l3k",
"tiJy"
],
"repo": "wcharczuk/go-chart",
"url": "https://github.com/wcharczuk/go-chart/issues/55",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1361093624 | Unexpected end of JSON input 這問題是?
指令如下,也在 cookie 用 touch 建立 coin 檔案 。
docker run -it -v ~/cookie/coin:/cookie hyperbola/shopee-coins-bot:1 -u '' -p '' -c /cookie
會出現 Unexpected end of JSON input
Start shopee coins bot v1.1.0.
Start to load cookies.
Failed to load cookies: Unexpected end of JSON input
Start to login shopee.
Failed to login. Missing username or password.
用的是甲骨文的 ARM。
指令如下,也在 cookie 用 touch 建立 coin 檔案 。
docker run -it -v ~/cookie/coin:/cookie hyperbola/shopee-coins-bot:1 -u '' -p '' -c /cookie
會出現 Unexpected end of JSON input
Start shopee coins bot v1.1.0.
Start to load cookies.
Failed to load cookies: Unexpected end of JSON input
Start to login shopee.
Failed to login. Missing username or password.
用的是甲骨文的 ARM。
抱歉,我覺得有些文不對題。標題寫出現了 Unexpected end of JSON input 的錯誤訊息,內文給的錯誤訊息卻是 Failed to login. Missing username or password.。
是否可以再麻煩你釐清,感謝!
好,我看懂了。
你的 cookie 格式是爛掉了,因為那是一個空檔案。然而,你給的帳號密碼都是空字串,這當然登不進去。
已成功用上 感謝
在ubuntu上 沒用-u的話 cookie寫不進去
| gharchive/issue | 2022-09-04T09:27:33 | 2025-04-01T06:46:13.090501 | {
"authors": [
"kai456789",
"wdzeng"
],
"repo": "wdzeng/shopee-coins-bot",
"url": "https://github.com/wdzeng/shopee-coins-bot/issues/38",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
358001166 | Excluding Goland idea based IDE meta files
Incomplete .gitignore.
Please may you accept the additional ignore line.
This should be in your ~/.gitignore
| gharchive/pull-request | 2018-09-07T09:58:30 | 2025-04-01T06:46:13.124599 | {
"authors": [
"casibbald",
"stefanprodan"
],
"repo": "weaveworks/flux",
"url": "https://github.com/weaveworks/flux/pull/1346",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
190020259 | Spawning terminals doesn't work most of the time
After Spawning the Socks Shop demo on ECS through Cloud Formation everything works in Scope except for terminals. Spawning them only works half of the time.
When spawning fails, the terminal is created in the UI but remains blank. The UI successfully creates the websocket to the app and the Chrome dev-tools reveal how characters are sent correctly but nothing comes back.
The problem appears when spawning both container terminals and host terminals.
Here's a probe excerpt when spawning a host terminal:
<probe> INFO: 2016/11/17 11:15:51.119072 Pipe pipe-767150125140527997 connection to scope.weave.local. starting
<probe> ERRO: 2016/11/17 11:16:12.845758 Error processing event for container ff2e0079907c724de574df942abedd2a7a6fc1c2ab3d820206ffe2e85cf84f46: API error (500): devmapper: Unknown device 207fc457a149fc81248abeb0b344435c763ca47a8aecd0ab8add4c258eee55cb
<probe> ERRO: 2016/11/17 11:16:51.121778 Error doing pipe-767150125140527997 for scope.weave.local., backing off 1s: read tcp 172.31.1.110:48912->172.31.0.202:4040: i/o timeout
<app> INFO: 2016/11/17 11:17:06.571640 Timing out pipe pipe-767150125140527997
<probe> INFO: 2016/11/17 11:17:17.438070 Host shell closed.
<probe> INFO: 2016/11/17 11:17:17.438096 Pipe pipe-767150125140527997 connection to scope.weave.local. exiting
<probe> ERRO: 2016/11/17 11:17:17.438340 Error waiting on host shell: signal: killed
And here's the corresponding App excerpt:
<app> INFO: 2016/11/17 11:17:17.435722 Timing out pipe pipe-767150125140527997
A similar output happens when a container terminal fails:
<probe> INFO: 2016/11/17 11:25:03.470493 Pipe pipe-1115508876840658752 connection to scope.weave.local. starting
<probe> ERRO: 2016/11/17 11:26:03.473427 Error doing pipe-1115508876840658752 for scope.weave.local., backing off 1s: read tcp 10.32.0.1:41246->10.46.0.0:4040: i/o timeout
<probe> INFO: 2016/11/17 11:26:06.576067 Pipe pipe-1115508876840658752 connection to scope.weave.local. exiting
Uhm, I noticed that all the Apps (ECS deploys both an app and probe on each host) show the same error, although I am only connecting from to one of the apps.
<app> INFO: 2016/11/17 11:26:06.571626 Timing out pipe pipe-1115508876840658752
<app> INFO: 2016/11/17 11:27:17.435735 Timing out pipe pipe-1115508876840658752
<app> INFO: 2016/11/17 11:26:14.119499 Timing out pipe pipe-1115508876840658752
I would expect the probe to only connect to the app initiating the control request. Could this be part of the problem?
The problem (at least partly) is that the ECS Cloud Formation Template uses an ELB infront of the standalone app which is completely wrong since it leads to a split-brain.
Yep, as soon as I start using the IP of one of the instances instead of the ELB endpoint everything works.
| gharchive/issue | 2016-11-17T11:29:08 | 2025-04-01T06:46:13.131041 | {
"authors": [
"2opremio"
],
"repo": "weaveworks/scope",
"url": "https://github.com/weaveworks/scope/issues/2016",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1832814284 | chore: fix example code
Summary
Currently, react-with-less and react-with-sass examples cannot be started normally.
Modify their rspack.config.js so that they can be started easily.
Test Plan
Goto these example dirs and execute pnpm dev. Then check that the view has rendered normally.
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
thanks
| gharchive/pull-request | 2023-08-02T09:35:38 | 2025-04-01T06:46:13.138024 | {
"authors": [
"CLAassistant",
"daydayhappychao",
"hardfist"
],
"repo": "web-infra-dev/rspack",
"url": "https://github.com/web-infra-dev/rspack/pull/3887",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2730189719 | refactor: simplify file system code
Summary
cherry pick refactor changes in #8643
refactor: Unify the return values of ThreadsafeNodeFS methods into promises to simplify the code
refactor: change fn <'a>() -> BoxFuture<'a, Result<()>>; to async fn() -> Result<()>
refactor: remove unnecessary async blocks
refactor(rspack_fs_node): create map_error_to_fs_error util to replace duplicate closures
refactor(rspack_fs_node): create new_fs_error to simplify code
fix: NodeFileStats lacks the is_symlink field
Checklist
[ ] Tests updated (or not required).
[ ] Documentation updated (or not required).
!bench
thanks again for your great work!
| gharchive/pull-request | 2024-12-10T13:53:28 | 2025-04-01T06:46:13.142123 | {
"authors": [
"hardfist",
"nilptr"
],
"repo": "web-infra-dev/rspack",
"url": "https://github.com/web-infra-dev/rspack/pull/8654",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2754479409 | fix: handle glob patterns ending with /** in CopyRspackPlugin
Summary
fix #8801
A glob pattern ending with /** should match all files within a directory, not just the directory itself.
Since the standard glob only matches directories (refer to https://github.com/rust-lang/glob/issues/129), we append /* to align with webpack's behavior.
Checklist
[x] Tests updated (or not required).
[ ] Documentation updated (or not required).
if the glob library behaves different from webpack's glob library, maybe we should consider align glob crate behavior to globby, I'm not sure the fast-glob behaves the same as globby and maybe we use fast-glob in copyRspackPlugin.
BTW this is actually a breaking and should be emphasized in notice
I'm not sure the fast-glob behaves the same as globby
Yes, it aligns with the glob pattern used by webpack and includes corresponding test cases.
BTW this is actually a breaking and should be emphasized in notice
Okay, sometimes it's hard to distinguish between fixing and breaking. I'll revise the corresponding description later to emphasize that.
@shulaoda fast-glob's API looks very concise. Can you help me confirm if it supports the globOptions configuration for use in the CopyRspackPlugin?
Okay, sometimes it's hard to distinguish between fixing and breaking. I'll revise the corresponding description later to emphasize that.
yeah this is tricky maybe we should have some label like “alignment breaking” which mean the breaking introduced for alignment which is allowed in minor release
@shulaoda fast-glob's API looks very concise. Can you help me confirm if it supports the globOptions configuration for use in the CopyRspackPlugin?
Yes, fast-glob only supports simple matching and doesn't support additional extensions, as these could lead to performance degradation. For more details, you can refer to #7556 related attempts made by @symind.
| gharchive/pull-request | 2024-12-22T04:16:43 | 2025-04-01T06:46:13.149364 | {
"authors": [
"hardfist",
"inottn",
"shulaoda"
],
"repo": "web-infra-dev/rspack",
"url": "https://github.com/web-infra-dev/rspack/pull/8803",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2536802776 | Prepare v1.4.0 release
Diff for changes in the release (as of right now): https://gist.github.com/ddbeck/57cdad030951e1114debaa458cc19ca4
Please review and (if applicable) approve this PR, but don't merge; I'll merge it myself when I'm ready to do the rest of the release checklist.
I ended up not getting to this on Friday, so I merged the latest BCD. Diff for this release: https://gist.github.com/ddbeck/f6c4a07ae0f896ac3dc6c45c6b8f6a30
| gharchive/pull-request | 2024-09-19T16:16:20 | 2025-04-01T06:46:13.151357 | {
"authors": [
"ddbeck"
],
"repo": "web-platform-dx/web-features",
"url": "https://github.com/web-platform-dx/web-features/pull/1814",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2649790669 | Session not created
what does mean of this error?
Try restarting your container a few times (worked for me)
10|gradient-9 | at parseHttpResponse (/app/node_modules/selenium-webdriver/lib/http.js:514:13)
10|gradient-9 | at Executor.execute (/app/node_modules/selenium-webdriver/lib/http.js:446:28)
10|gradient-9 | at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
10|gradient-9 | at async Driver.execute (/app/node_modules/selenium-webdriver/lib/webdriver.js:744:17)
10|gradient-9 | at async getProxyIpInfo (/app/app.js:177:5)
10|gradient-9 | at async /app/app.js:215:9 {
10|gradient-9 | remoteStacktrace: '#0 0x5557ff0216da \n' +
10|gradient-9 | '#1 0x5557feb35f80 \n' +
10|gradient-9 | '#2 0x5557feb2e181 \n' +
10|gradient-9 | '#3 0x5557feb1e8b9 \n' +
10|gradient-9 | '#4 0x5557feb205ea \n' +
10|gradient-9 | '#6 0x5557feb1e3df \n' +
10|gradient-9 | '#7 0x5557feb1e2d2 \n' +
10|gradient-9 | '#9 0x5557feb1c99a \n' +
10|gradient-9 | '#10 0x5557feb38879 \n' +
10|gradient-9 | '#12 0x5557feba6bb2 \n' +
10|gradient-9 | '#13 0x5557febc569d \n' +
10|gradient-9 | '#14 0x5557feba6953 \n' +
10|gradient-9 | '#15 0x5557feb7572e \n' +
10|gradient-9 | '#16 0x5557feb7679e \n' +
10|gradient-9 | '#17 0x5557fefecf1b \n' +
10|gradient-9 | '#18 0x5557feff0eb8 \n' +
10|gradient-9 | '#19 0x5557fefda42c \n' +
10|gradient-9 | '#20 0x5557feff1a37 \n' +
10|gradient-9 | '#21 0x5557fefbefef \n' +
10|gradient-9 | '#22 0x5557ff00fad8 \n' +
10|gradient-9 | '#23 0x5557ff00fca0 \n' +
10|gradient-9 | '#24 0x5557ff020556 \n' +
10|gradient-9 | '#25 0x7f3d10f9b144 \n'
still error
already restart the docker multiple times
| gharchive/issue | 2024-11-11T16:21:09 | 2025-04-01T06:46:13.211497 | {
"authors": [
"Lukey372",
"MHendriF",
"iamspa"
],
"repo": "web3bothub/gradient-bot",
"url": "https://github.com/web3bothub/gradient-bot/issues/32",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1034265245 | App: SFX pack
We should get an SFX pack and create a sound system to trigger them. Sound is easy but pretty important for the final product.
This sfx pack for anime games should be interesting for Webaverse:
https://assetstore.unity.com/packages/audio/sound-fx/anime-epic-combat-sound-effect-pack-206061
Maybe this is also interesting
https://assetstore.unity.com/packages/audio/sound-fx/vintage-anime-sfx-158863
Here we have some basic sfx sounds
https://assetstore.unity.com/packages/audio/sound-fx/eastern-sfx-74043
All of the assets can be seen here:
https://assetstore.unity.com/top-assets/top-paid?category=audio&q=anime&orderBy=1&page=1
| gharchive/issue | 2021-10-23T21:24:43 | 2025-04-01T06:46:13.220434 | {
"authors": [
"avaer",
"skaljac"
],
"repo": "webaverse/app",
"url": "https://github.com/webaverse/app/issues/1633",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
787711219 | FE: add NFT collaborators
The same thing as https://github.com/webaverse/app/issues/590, except for hash-based metadata collaborators on ASSET NFTs on the frontend.
https://github.com/webaverse/contracts/blob/99b56bde896d98aaa8d63e077678dc486133a857/ethereum/WebaverseERC721.sol#L210
added https://github.com/webaverse/webaverse.com/compare/7aed11b89f44...be266fbf1696
needs https://github.com/webaverse/app/issues/598 to be functional as expected
added https://github.com/webaverse/webaverse.com/compare/7aed11b89f44...be266fbf1696
needs https://github.com/webaverse/app/issues/598 to be functional as expected
| gharchive/issue | 2021-01-17T13:56:04 | 2025-04-01T06:46:13.223507 | {
"authors": [
"avaer",
"chrislatorres"
],
"repo": "webaverse/app",
"url": "https://github.com/webaverse/app/issues/592",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1890034486 | [BUG] Relayer Fails to Reset When Users Switch to a Different Bridge
Currently, in the withdrawal/transfer process, the Hubble bridge automatically selects a relayer based on the initial state. However, when users switch to a different bridge (e.g., opt for a different chain or pool), the relayer neither resets nor updates. This results in an issue when users select a relayer that is not supported by the chosen bridge.
Expected Behavior
Whenever users switch to a different bridge, the relayer should reset, or alternatively, select a relayer that is supported by the new bridge.
Current Behavior
The relayer neither resets nor reselects when a different bridge is chosen; it merely remains in the state of the previously selected bridge.
Possible Solution
Implement a watch function to monitor changes in the chain or pool, and accordingly, reset or reselect the supported relayer.
Closed in https://github.com/webb-tools/webb-dapp/pull/1610.
| gharchive/issue | 2023-09-11T09:08:45 | 2025-04-01T06:46:13.226704 | {
"authors": [
"AtelyPham"
],
"repo": "webb-tools/webb-dapp",
"url": "https://github.com/webb-tools/webb-dapp/issues/1603",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
622095689 | '/usr/bin/git' failed with exit code 128
We're seeing this error consistently. We have successfully setup the unity-test-runner, but are struggling with the builder. This message below is the only thing we see in the logs, the unity-builder step takes 1s to complete, so I would assume there are no Unity logs.
Note that our project is made of 2 sibling repos: my-project-unity and my-project-server, both are folders in the root. We want to build my-project-unity, which inside contains the folders Assets, ProjectSettings and Packages.
Any idea what's going on? Thanks!
Run webbertakken/unity-builder@v0.15
with:
projectPath: my-project-unity
unityVersion: 2019.2.11f1
targetPlatform: iOS
buildMethod: MyNamespace.MyStaticMethodToBuild
env:
UNITY_EMAIL: ***
UNITY_PASSWORD: ***
UNITY_SERIAL: ***
##[warning]Error: The process '/usr/bin/git' failed with exit code 128
##[error]The process '/usr/bin/git' failed with exit code 128
@ivan-omnidrone Thanks a lot for reporting your issue with unity builder!
Technically it would be helpful if it were to give a little bit more output, so that is something to be improved.
To me it looks like the commands (that you find in versioning.js of this repo) can not find a git repository in the root of your project, as you have just checked out 2 repos in these folders.
I believe this problem can be solved by running all the git commands in the folder that's specified using projectPath.
@ivan-omnidrone Could you try again with unity-builder@3c040dd?
Yes, it gave me this:
Run webbertakken/unity-builder@3c040dd
with:
projectPath: my-project-unity
unityVersion: 2019.2.11f1
targetPlatform: iOS
buildMethod: MyNamespace.MyStaticMethodToBuild
env:
UNITY_EMAIL: ***
UNITY_PASSWORD: ***
UNITY_SERIAL: ***
##[warning]Error: The process '/usr/bin/git' failed with exit code 1
M .gitattributes
M .github/PULL_REQUEST_TEMPLATE.md
M .github/workflows/build.yml
M .gitignore
M .gitmodules
M .vscode/settings.json
M Assets/.editorconfig
M Assets/.gitignore
... [Bunch of other files from our repo]
##[error]Branch is dirty. Refusing to base semantic version on uncommitted changes
Our current versioning strategy doesn't seem to follow any of the available ATM, but that's another topic. I'm setting it to None for now and it's building.
with:
versioning: None
A related issue might be that documentation states that Strategies only work when no custom buildMethod is specified but that doesn't seems to be true anymore -didn't really followed the code to see intention-.
Awesome work, BTW, really excited and looking forward into how this action evolves :)
@ivan-omnidrone Thank you for testing it, and thanks for your compliments.
The error that you are now getting is as designed; Branch is dirty means that you checked out your repository but then changed files after the checkout. It fails for that because it's considered bad practice to let the pipeline change any files (unless if there's a valid reason to do it).
Versioning: None will work, but actually I would recommend sticking with the default of Semantic unless you already have worked out versioning for your project by yourself.
I'll explain why:
Incorporating your own versioning scheme can be hard and tedious and takes time away from working on the project itself.
The option Semantic will work out of the box; if you don't do anything, and you have also never set a tag before, the version will be 0.0.257 where 257 is the amount of commits in your repository at the time of your build.
You can update major and minor versions as simple as creating a tag on your repository. (for example v1.0 or v0.1, then your next build will use version 1.0.0 or 0.1.0 respectively, and increment the last digit for the amount of commits since that tag.
This will help you easily identify mismatches between server and client (in case they're in the same repository)
Lastly, if you use a bug reporting system, the version will act as an identifier, that will be unique for every build (in your case different for your server as for the client, as they are 2 repositories).
The only thing you would have to do is make sure that your repository is not dirty (or we could add a flag that allows for dirty builds).
Don't get me wrong; of course you're free to use the actions as you please. I'm just curious about the use case and whether or not knowing the above you would indeed still choose for None. Perhaps we can improve the actions for everyone's sake.
To come back on your comment about strategies;
A related issue might be that documentation states that Strategies only work when no custom buildMethod is specified but that doesn't seems to be true anymore -didn't really followed the code to see intention-.
It's still true that they only work with the default buildMethod, or in case you implement a function that applies the version to the build in your custom build method, like illustrated below;
However there is a pre-build step that determines the version for the build. This pre-build step happens outside of the buildMethod and even before the unity container is started. This is why it failed on this pre-build step for you and not in the build itself. If you have a custom buildMethod, you'll also have to pass the version argument into your build, as can be seen here.
I think you're absolutely right about versioning, and of course we'll need something. Currently we have a build version number in Jenkins and major and minor in Unity. While attempting to get the builds done with Github actions, I was just delaying this problem for a moment, but sounds like the Semantic default will work for us.
About the repo not being clean, our case is that we're cloning the repo and then deleting StreamingAssets folder for Android or iOS depending on which platform are we building. That will show the repo as dirty, but it might be a legit case.
@ivan-omnidrone I would call that a legit use case indeed. I added the flag for this.
could you test your build once more using unity-builder@a808c06 with:
allowDirtyBuild: true
And versioning set back to the default Semantic?
Enabling both things worked correctly!
There's an unrelated issue that I'll need to debug in our end.
Thank you!
| gharchive/issue | 2020-05-20T21:28:28 | 2025-04-01T06:46:13.240296 | {
"authors": [
"ivan-omnidrone",
"webbertakken"
],
"repo": "webbertakken/unity-builder",
"url": "https://github.com/webbertakken/unity-builder/issues/89",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1120476167 | Wrap xhr::abort and add tests for angular
Hi there!
Following my investigations on the issue https://github.com/webdriverio-community/wdio-intercept-service/issues/204, I added a simple test to ensure everything goes well when using angular HTTP client.
Hey, so I tried to follow your comment here and I updated the code to wrap xhr::abort.
If we accept this test, how easily can maintainers ensure we are testing against "latest stable" Angular, and/or test against "supported major versions"? Ideally one would be able to expand from this test and provide assurances that this library can function alongside a "reasonable" range of Angular versions.
I think it's already a good start to have a test that checks that the lib is doing fine when working with these two angular versions. I can improve the tests I added a bit but my goal is not to add a complete angular testing solution.
@chmanie @christian-bromann can we get the checks run on this PR?
Hey @tehhowch, do you have any more review comments? I would love to move forward on this if possible.
Nothing I can do; @chmanie and @christian-bromann are the maintainers. When either of them gets time, they will hopefully approve checks and then merge / release.
@chmanie @christian-bromann mind merging and releasing?
@lildesert please update the changelog
@tehhowch How can I do that? Just update my PR description? Sorry, I'm not used to the process ^^
@lildesert In another commit in your PR, please amend this file under the v.next section with the feature you're adding (and add a thanks with your username 😀).
@christian-bromann @chmanie bump for checks & merge 🙂
@tehhowch sorry for taking so long. If it was for me, I'd add you as a maintainer so that you can approve & merge yourself. But only @christian-bromann has the power to make that happen :)
| gharchive/pull-request | 2022-02-01T10:18:56 | 2025-04-01T06:46:13.367151 | {
"authors": [
"chmanie",
"lildesert",
"tehhowch"
],
"repo": "webdriverio-community/wdio-intercept-service",
"url": "https://github.com/webdriverio-community/wdio-intercept-service/pull/225",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1980589987 | Flipped toybox normal map along x and y-axis.
See issue #317
It looks like after you flipped the height map, you then regenerated the normal-map, instead of just flipping that. For some reason, the normals for this texture then point the wrong way. I don't know why this texture appears wrong, when the others look okay.
I wanted to confirm that this was an issue with the texture and not a bug in the code, so I spent some time going through the sample's code. I think the code was sound, but I cleaned up the code as I went along. PR here: #320
Pull request closed due to issues being more comprehensively addressed by #320.
| gharchive/pull-request | 2023-11-07T05:43:33 | 2025-04-01T06:46:13.426065 | {
"authors": [
"ben-clayton",
"cmhhelgeson"
],
"repo": "webgpu/webgpu-samples",
"url": "https://github.com/webgpu/webgpu-samples/pull/319",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
515386199 | Update Clear-Site-Data value as quoted string
The Clear-Site-Data HTTP header's values "MUST comply with the quoted-string grammar", as defined in https://tools.ietf.org/html/rfc7230#section-3.2.6.
Fix #855.
Thanks @Malvoz !
@Malvoz is there a way we can talk to you privately about an idea we've had?
Maybe on gitter or anywhere else? My DMs are open on Twitter and my email is anton at gmail dot com.
Thanks!
@molant I've sent you a message on Gitter. ^^
| gharchive/pull-request | 2019-10-31T12:18:44 | 2025-04-01T06:46:13.546659 | {
"authors": [
"Malvoz",
"molant"
],
"repo": "webhintio/webhint.io",
"url": "https://github.com/webhintio/webhint.io/pull/856",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1386298732 | Compressed or Uncompressed size, or offset exceeds the maximum value
When using the Compress-Citem without UseShell on files that are larger than 2GB, the following error is thrown:
Compressed or Uncompressed size, or offset exceeds the maximum value. Consider setting the UseZip64WhenSaving property on the ZipFile instance.
Is it possible to implement something like:
else
{
$zipFile = New-Object 'Ionic.Zip.ZipFile'
if ($Use64Bit)
{
$zipFile.UseZip64WhenSaving = [Ionic.Zip.Zip64Option]::Always
}
}
With an additional parameter added
[Switch]
Uses the 64 bit option with 'DotNetZip' library.
$Use64Bit
| gharchive/issue | 2022-09-26T15:36:35 | 2025-04-01T06:46:13.551440 | {
"authors": [
"Gijsreyn"
],
"repo": "webmd-health-services/Carbon",
"url": "https://github.com/webmd-health-services/Carbon/issues/127",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
295564671 | [0.12][WIP] Small upstream updates
Each commit mirrors one of the smaller commits in graphql-js. Mostly test additions or remove comments etc.
Coverage remained the same at 92.84% when pulling 621eccf54bd9080dac787103ecd0ec0e93b12ffa on danez:012-small-stuff into 94525c00252696b47673db1ca56fe0436c5949be on webonyx:master.
| gharchive/pull-request | 2018-02-08T15:54:41 | 2025-04-01T06:46:13.604668 | {
"authors": [
"coveralls",
"danez"
],
"repo": "webonyx/graphql-php",
"url": "https://github.com/webonyx/graphql-php/pull/241",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
276916255 | Different behavior between server-side and client-side?
Hey, guys
I use the latest version of file-loader (1.1.5) with my spa and encounter this issue. I require a png image in my component for my og:image tag, but it acts differently between server-side and client-side. In server-side, it seems like that it requires the image in base64 format. But in client-side, it requires the image in relative path which is desired for me because of og:image tag does not support base64 format.
here is the website
https://www.cakeresume.com/search?locale=en
You can see the different format in og:image tag if you disable javascript in the browser console.
my webpack config:
module: {
loaders: [
{
test: /\.js$/,
exclude: /node_modules/,
loader: 'babel-loader',
},
{
test: /\.(woff(2)?|ttf|eot|svg)(\?v=[0-9]\.[0-9]\.[0-9])?$/,
loader: 'url-loader'
},
{
test: /\.(png|jpg)(\?v=[0-9]\.[0-9]\.[0-9])?$/,
loader: 'file-loader',
options: {
name: '[name]-[hash].[ext]',
},
},
],
},
and the code I require image:
const defaultOgImage = require(`../../../images/TalentSearch/default_og_image.${locale}@2x.png`);
Here I use require instead of import because I want to load different image by different locale which is dynamic. Perhaps this is where this issue comes.
thanks.
p.s. I manage meta tags with react-helmet
Did you find solution in the meantime ? This definitely needs more info about your current setup and reproducible example at best. Upfront this is likely a issue in somewhere in you setup...
| gharchive/issue | 2017-11-27T06:46:28 | 2025-04-01T06:46:13.619845 | {
"authors": [
"frozenfung",
"michael-ciniawsky"
],
"repo": "webpack-contrib/file-loader",
"url": "https://github.com/webpack-contrib/file-loader/issues/227",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
371968269 | ERROR in ./assets/js/commons.js from UglifyJs Unexpected token: keyword (const)
I use uglifyjs-webpack-plugin.
I get this error :"ERROR in ./assets/js/commons.js from UglifyJs
Unexpected token: keyword (const) [./assets/js/commons.js:13526,0]
"
what's the solution to this ?
version info :
nodejs: v8.12.0
npm 6.4.1,
webpack: "^4.21.0",
"uglifyjs-webpack-plugin": "^2.0.1",
"babel-loader": "^8.0.4",
babel : 7
+1
same problem
You code contains ES6, uglifyjs-webpack-plugin use uglify-js to compress code and uglify-js doesn't support ES6, use https://github.com/webpack-contrib/terser-webpack-plugin
Execution of "yarn add uglifyjs-webpack-plugin" can be temporarily resolved. The specific reasons need to be further searched.
| gharchive/issue | 2018-10-19T13:55:01 | 2025-04-01T06:46:13.627614 | {
"authors": [
"evilebottnawi",
"kk-fan",
"m-korkut",
"mhavci"
],
"repo": "webpack-contrib/uglifyjs-webpack-plugin",
"url": "https://github.com/webpack-contrib/uglifyjs-webpack-plugin/issues/367",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
88252249 | Please add a link for "Don't forget to polyfill require"
This statement is absolutely not what someone who's looking for information on why loading json isn't working hopes to find. It just makes them more angry because no one's explaining how the hell to use this loader =)
I agree. Related issue: https://github.com/webpack/webpack/issues/140
+1. I have no idea how to use var pkg = require('../package.json'); in a Node project. I set target: 'node' and externals: [ /\.json$/ ],, but it doesn't work. I don't know if json-loader helps me here?
This doesn't seem to work when I use require(path.join(process.cwd(), 'package.json'));. I created a separate issue for that: https://github.com/webpack/webpack/issues/2413
@donaldpipowitch requires with magic names will always be a problem because the bundler has no idea what you're asking for (it would have to actually run the code to figure out which thing you're actually requiring) so you're going to have to use relative, explicit strings like require('./package.json').
I don't want to bundle my JSON file in that case. I want a vanilla Node require. Webpack should do nothing. (I have a hack in webpack/webpack#2413 which works.)
if it's code that will end up in your bundle, there is no vanilla node, because the bundle will run in the browser. If it's code that never ends up bundled in, then you don't need a webpack loader, node already knows how to require .json files
As mentioned above it is target: 'node' and will not be executed in the browser. The JSON should not be bundled, but required at runtime.
node already knows how to require .json files
That is exactly what I meant with Nodes "vanilla require". I use webpack just for other things then JSON.
in that case I don't quite understand why you're commenting on a thread that is specifically about webpack's json-loader, which is necessary only when you need to bundle. If your require is in code that doesn't need to be bundled, then this library is not relevant to your needs, but more importantly even if it is, this particular issue --about a README improvement-- has nothing to do with the question you really want to ask. Better to open a new issue, possibly not even on this repo.
Sorry, I don't know if you read my comments. In my first commend I said "I don't know if json-loader helps me here?". So yes, maybe this was the wrong issue in the first place but nobody corrected me, so I thought I was correct. That was nearly a month ago. Now I know json-loader was the wrong tool for my problem. I DID created a new issue in a different repository. I just commented here for people who look at this loader for the same problem in the first place - just like me. I don't even use json-loader. But if you ask someone how to load JSON in a webpack based Node project, everyone seems to redirect you to this repository as the only valid solution...
You even said "this statement is absolutely not what someone who's looking for information on why loading json isn't working hopes to find." This is what drove me here, too. Loading JSON in a Node project wasn't working for me. However this README doesn't tell what this loader really does (or not does which was more important for me as it turns out). It just says "don't forget the require polyfill". Browsers don't have a require which could be polyfilled. Only Node has require. So yeah, that's why I thought this missing polyfill was the reason why I couldn't load JSON in Node when using webpack, because I have no clue if webpack alters Nodes require. If not, why wouldn't it work (especially with target: 'node'), right? Sorry for that! But seriously... I still don't know what this require polyfill even does? Why do I need a polyfill for bundling (as in inlining) JSON?
but... but this is the webpack/json-loader repository. It's for webpack, that part should be relatively obvious?
As an alternative to using a require polyfill for runtime loading of json resources that you aren't using json-loader to inline with webpack for whatever reason (they're too big and you don't want to slow down app loading, they change at runtime, etc.), you might be interested in the solution I posted in response to the issue @donaldpipowitch referenced: https://github.com/webpack/webpack/issues/2413#issuecomment-216381287
If you're using webpack to bundle your code and you need to bundle in a json resource, then you want this loader. If you're not forming a webpacked bundle, then you this library is not what you were looking for.
I do use webpack to bundle code, but this doesn't mean that I want to bundle my JSON, because it isn't static. I just wanted clarify that. It is not either bundling and using webpack or not bundling and don't use webpack. You can use webpack for bundling most things (like your code), but not everything (like some JSON). And it is quite hard to figure out how to opt-out from bundling in some cases.
@rob3c Thanks, I'll look into that.
I would like to vote to remove the statement entirely, instead of adding a link, because this sentence makes it seem that if I do want to bundle the JSON with my generated library, then I need to do some additional work in order to make it function in Node environments—which does not seem to be the case.
In webpack/webpack#140 @sokra goes so far as to state "The sentence is pretty old, I should remove it from all the loaders.", indicating this is a legacy of an earlier time.
@kadamwhite thanks for your insights. I've removed the statement.
| gharchive/issue | 2015-06-14T21:35:23 | 2025-04-01T06:46:13.640153 | {
"authors": [
"Pomax",
"SpaceK33z",
"dmitry",
"donaldpipowitch",
"kadamwhite",
"rob3c"
],
"repo": "webpack/json-loader",
"url": "https://github.com/webpack/json-loader/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
362341573 | Update filelist.md plugin example in write-a-plugin.md
In the filelist.md the callback parameter should be compilation, not compiler.
Duplicate of https://github.com/webpack/webpack.js.org/pull/2531, please check PRs before submitting yours.
| gharchive/pull-request | 2018-09-20T19:55:02 | 2025-04-01T06:46:13.791183 | {
"authors": [
"danielna",
"montogeek"
],
"repo": "webpack/webpack.js.org",
"url": "https://github.com/webpack/webpack.js.org/pull/2535",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
702863744 | corejs error after update to beta 30
Bug report
What is the current behavior?
Our build worked with beta 29 and previous, but is failing with beta 30 with the following error:
[
{
"moduleIdentifier": "/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/extends.js",
"moduleName": "../node_modules/@babel/runtime-corejs2/helpers/esm/extends.js",
"loc": "1:0-57",
"message": "Module not found: Error: Can't resolve '../../core-js/object/assign' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'",
"moduleId": "../node_modules/@babel/runtime-corejs2/helpers/esm/extends.js",
"moduleTrace": [
{
"originIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/utils/bootstrapUtils.js",
"originName": "../node_modules/react-bootstrap/es/utils/bootstrapUtils.js",
"moduleIdentifier": "/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/extends.js",
"moduleName": "../node_modules/@babel/runtime-corejs2/helpers/esm/extends.js",
"dependencies": [
{
"loc": "2:0-66"
},
{
"loc": "55:24-32"
},
{
"loc": "93:24-32"
}
],
"originId": "../node_modules/react-bootstrap/es/utils/bootstrapUtils.js",
"moduleId": "../node_modules/@babel/runtime-corejs2/helpers/esm/extends.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/Tooltip.js",
"originName": "../node_modules/react-bootstrap/es/Tooltip.js",
"moduleIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/utils/bootstrapUtils.js",
"moduleName": "../node_modules/react-bootstrap/es/utils/bootstrapUtils.js",
"dependencies": [
{
"loc": "8:0-84"
},
{
"loc": "71:24-36"
},
{
"loc": "75:31-42"
},
{
"loc": "91:17-23"
},
{
"loc": "94:17-23"
},
{
"loc": "103:15-22"
}
],
"originId": "../node_modules/react-bootstrap/es/Tooltip.js",
"moduleId": "../node_modules/react-bootstrap/es/utils/bootstrapUtils.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/index.js",
"originName": "../node_modules/react-bootstrap/es/index.js",
"moduleIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/Tooltip.js",
"moduleName": "../node_modules/react-bootstrap/es/Tooltip.js",
"dependencies": [
{
"loc": "137:0-33"
},
{
"loc": "138:0-31"
}
],
"originId": "../node_modules/react-bootstrap/es/index.js",
"moduleId": "../node_modules/react-bootstrap/es/Tooltip.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/CreateUserModal.jsx",
"originName": "./src/apps/admin/components/CreateUserModal.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/index.js",
"moduleName": "../node_modules/react-bootstrap/es/index.js",
"dependencies": [
{
"loc": "10:22-48"
}
],
"originId": "./src/apps/admin/components/CreateUserModal.jsx",
"moduleId": "../node_modules/react-bootstrap/es/index.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/UsersPage.jsx",
"originName": "./src/apps/admin/components/UsersPage.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/CreateUserModal.jsx",
"moduleName": "./src/apps/admin/components/CreateUserModal.jsx",
"dependencies": [
{
"loc": "14:46-74"
}
],
"originId": "./src/apps/admin/components/UsersPage.jsx",
"moduleId": "./src/apps/admin/components/CreateUserModal.jsx"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/app.jsx",
"originName": "./src/apps/admin/app.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/UsersPage.jsx",
"moduleName": "./src/apps/admin/components/UsersPage.jsx",
"dependencies": [
{
"loc": "27:40-73"
}
],
"originId": "./src/apps/admin/app.jsx",
"moduleId": "./src/apps/admin/components/UsersPage.jsx"
}
],
"details": "resolve '../../core-js/object/assign' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'\n using description file: /usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/package.json (relative path: .)\n Field 'browser' doesn't contain a valid alias configuration\n using description file: /usr/src/app/node_modules/@babel/runtime-corejs2/package.json (relative path: ./core-js/object/assign)\n Field 'browser' doesn't contain a valid alias configuration\n /usr/src/app/node_modules/@babel/runtime-corejs2/core-js/object/assign doesn't exist",
"stack": "ModuleNotFoundError: Module not found: Error: Can't resolve '../../core-js/object/assign' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'\n at /usr/src/app/node_modules/webpack/lib/Compilation.js:1470:28\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:647:13\n at eval (eval at create (/usr/src/app/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:8:1)\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:233:22\n at eval (eval at create (/usr/src/app/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:7:1)\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:357:22\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:116:11\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:576:24\n at finishWithoutResolve (/usr/src/app/node_modules/enhanced-resolve/lib/Resolver.js:284:11)\n at /usr/src/app/node_modules/enhanced-resolve/lib/Resolver.js:350:15"
},
{
"moduleIdentifier": "/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/inheritsLoose.js",
"moduleName": "../node_modules/@babel/runtime-corejs2/helpers/esm/inheritsLoose.js",
"loc": "1:0-57",
"message": "Module not found: Error: Can't resolve '../../core-js/object/create' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'",
"moduleId": "../node_modules/@babel/runtime-corejs2/helpers/esm/inheritsLoose.js",
"moduleTrace": [
{
"originIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/Tooltip.js",
"originName": "../node_modules/react-bootstrap/es/Tooltip.js",
"moduleIdentifier": "/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/inheritsLoose.js",
"moduleName": "../node_modules/@babel/runtime-corejs2/helpers/esm/inheritsLoose.js",
"dependencies": [
{
"loc": "3:0-78"
},
{
"loc": "49:2-16"
}
],
"originId": "../node_modules/react-bootstrap/es/Tooltip.js",
"moduleId": "../node_modules/@babel/runtime-corejs2/helpers/esm/inheritsLoose.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/index.js",
"originName": "../node_modules/react-bootstrap/es/index.js",
"moduleIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/Tooltip.js",
"moduleName": "../node_modules/react-bootstrap/es/Tooltip.js",
"dependencies": [
{
"loc": "137:0-33"
},
{
"loc": "138:0-31"
}
],
"originId": "../node_modules/react-bootstrap/es/index.js",
"moduleId": "../node_modules/react-bootstrap/es/Tooltip.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/CreateUserModal.jsx",
"originName": "./src/apps/admin/components/CreateUserModal.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/index.js",
"moduleName": "../node_modules/react-bootstrap/es/index.js",
"dependencies": [
{
"loc": "10:22-48"
}
],
"originId": "./src/apps/admin/components/CreateUserModal.jsx",
"moduleId": "../node_modules/react-bootstrap/es/index.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/UsersPage.jsx",
"originName": "./src/apps/admin/components/UsersPage.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/CreateUserModal.jsx",
"moduleName": "./src/apps/admin/components/CreateUserModal.jsx",
"dependencies": [
{
"loc": "14:46-74"
}
],
"originId": "./src/apps/admin/components/UsersPage.jsx",
"moduleId": "./src/apps/admin/components/CreateUserModal.jsx"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/app.jsx",
"originName": "./src/apps/admin/app.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/UsersPage.jsx",
"moduleName": "./src/apps/admin/components/UsersPage.jsx",
"dependencies": [
{
"loc": "27:40-73"
}
],
"originId": "./src/apps/admin/app.jsx",
"moduleId": "./src/apps/admin/components/UsersPage.jsx"
}
],
"details": "resolve '../../core-js/object/create' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'\n using description file: /usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/package.json (relative path: .)\n Field 'browser' doesn't contain a valid alias configuration\n using description file: /usr/src/app/node_modules/@babel/runtime-corejs2/package.json (relative path: ./core-js/object/create)\n Field 'browser' doesn't contain a valid alias configuration\n /usr/src/app/node_modules/@babel/runtime-corejs2/core-js/object/create doesn't exist",
"stack": "ModuleNotFoundError: Module not found: Error: Can't resolve '../../core-js/object/create' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'\n at /usr/src/app/node_modules/webpack/lib/Compilation.js:1470:28\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:647:13\n at eval (eval at create (/usr/src/app/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:8:1)\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:233:22\n at eval (eval at create (/usr/src/app/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:7:1)\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:357:22\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:116:11\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:576:24\n at finishWithoutResolve (/usr/src/app/node_modules/enhanced-resolve/lib/Resolver.js:284:11)\n at /usr/src/app/node_modules/enhanced-resolve/lib/Resolver.js:350:15"
},
{
"moduleIdentifier": "/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/objectWithoutPropertiesLoose.js",
"moduleName": "../node_modules/@babel/runtime-corejs2/helpers/esm/objectWithoutPropertiesLoose.js",
"loc": "1:0-53",
"message": "Module not found: Error: Can't resolve '../../core-js/object/keys' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'",
"moduleId": "../node_modules/@babel/runtime-corejs2/helpers/esm/objectWithoutPropertiesLoose.js",
"moduleTrace": [
{
"originIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/Tooltip.js",
"originName": "../node_modules/react-bootstrap/es/Tooltip.js",
"moduleIdentifier": "/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/objectWithoutPropertiesLoose.js",
"moduleName": "../node_modules/@babel/runtime-corejs2/helpers/esm/objectWithoutPropertiesLoose.js",
"dependencies": [
{
"loc": "2:0-108"
},
{
"loc": "69:16-45"
}
],
"originId": "../node_modules/react-bootstrap/es/Tooltip.js",
"moduleId": "../node_modules/@babel/runtime-corejs2/helpers/esm/objectWithoutPropertiesLoose.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/index.js",
"originName": "../node_modules/react-bootstrap/es/index.js",
"moduleIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/Tooltip.js",
"moduleName": "../node_modules/react-bootstrap/es/Tooltip.js",
"dependencies": [
{
"loc": "137:0-33"
},
{
"loc": "138:0-31"
}
],
"originId": "../node_modules/react-bootstrap/es/index.js",
"moduleId": "../node_modules/react-bootstrap/es/Tooltip.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/CreateUserModal.jsx",
"originName": "./src/apps/admin/components/CreateUserModal.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/react-bootstrap/es/index.js",
"moduleName": "../node_modules/react-bootstrap/es/index.js",
"dependencies": [
{
"loc": "10:22-48"
}
],
"originId": "./src/apps/admin/components/CreateUserModal.jsx",
"moduleId": "../node_modules/react-bootstrap/es/index.js"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/UsersPage.jsx",
"originName": "./src/apps/admin/components/UsersPage.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/CreateUserModal.jsx",
"moduleName": "./src/apps/admin/components/CreateUserModal.jsx",
"dependencies": [
{
"loc": "14:46-74"
}
],
"originId": "./src/apps/admin/components/UsersPage.jsx",
"moduleId": "./src/apps/admin/components/CreateUserModal.jsx"
},
{
"originIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/app.jsx",
"originName": "./src/apps/admin/app.jsx",
"moduleIdentifier": "/usr/src/app/node_modules/babel-loader/lib/index.js??ruleSet[1].rules[0].use[0]!/usr/src/app/admin/src/apps/admin/components/UsersPage.jsx",
"moduleName": "./src/apps/admin/components/UsersPage.jsx",
"dependencies": [
{
"loc": "27:40-73"
}
],
"originId": "./src/apps/admin/app.jsx",
"moduleId": "./src/apps/admin/components/UsersPage.jsx"
}
],
"details": "resolve '../../core-js/object/keys' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'\n using description file: /usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm/package.json (relative path: .)\n Field 'browser' doesn't contain a valid alias configuration\n using description file: /usr/src/app/node_modules/@babel/runtime-corejs2/package.json (relative path: ./core-js/object/keys)\n Field 'browser' doesn't contain a valid alias configuration\n /usr/src/app/node_modules/@babel/runtime-corejs2/core-js/object/keys doesn't exist",
"stack": "ModuleNotFoundError: Module not found: Error: Can't resolve '../../core-js/object/keys' in '/usr/src/app/node_modules/@babel/runtime-corejs2/helpers/esm'\n at /usr/src/app/node_modules/webpack/lib/Compilation.js:1470:28\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:647:13\n at eval (eval at create (/usr/src/app/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:8:1)\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:233:22\n at eval (eval at create (/usr/src/app/node_modules/tapable/lib/HookCodeFactory.js:33:10), <anonymous>:7:1)\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:357:22\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:116:11\n at /usr/src/app/node_modules/webpack/lib/NormalModuleFactory.js:576:24\n at finishWithoutResolve (/usr/src/app/node_modules/enhanced-resolve/lib/Resolver.js:284:11)\n at /usr/src/app/node_modules/enhanced-resolve/lib/Resolver.js:350:15"
}
]
What is the expected behavior?
Build should continue to work as it did before
Other relevant information:
webpack version: 5.0.0-beta.30
Node.js version: 12.18.3
Operating System: debian 10.5
Additional tools: babel-loader 8.1.0
Please use search https://github.com/webpack/webpack/issues/11467
Thanks for the quick response! I added the recommended rule and it's now working:
https://github.com/webpack/webpack/issues/11467#issuecomment-691982478
It is bug from babel
| gharchive/issue | 2020-09-16T15:29:25 | 2025-04-01T06:46:13.804002 | {
"authors": [
"daveisfera",
"evilebottnawi"
],
"repo": "webpack/webpack",
"url": "https://github.com/webpack/webpack/issues/11484",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
35019387 | webpack-dev-server stuck in invalid state
This is unpredictable, but there are times when webpack-dev-server becomes stuck in a compilation loop. The dev server invalidates the build which causes a re-build, then it invalidates it again, and this loop continues indefinitely. I don't have enough telemetry to know what is going on yet, but I wanted to document this issue as we have seen it multiple times now.
propably fixed
| gharchive/issue | 2014-06-05T01:01:12 | 2025-04-01T06:46:13.805800 | {
"authors": [
"maspwr",
"sokra"
],
"repo": "webpack/webpack",
"url": "https://github.com/webpack/webpack/issues/297",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
208924220 | SyntaxError: Unexpected character '`' [0.0972f70db899961b2a15.hot-update.js:16,9]
Do you want to request a feature or report a bug?
Bug
What is the current behavior?
When I try to pack using --optimize-minimize it fails if I have strings written with grave accent character (template string)
If the current behavior is a bug, please provide the steps to reproduce.
Try to print a string using template characters instead of regular string characters
alert('regular string, works')
alert(`template string containing ${variables} wont work`)
What is the expected behavior?
It should pack normally
If this is a feature request, what is motivation or use case for changing the behavior?
Please mention other relevant information such as the browser version, Node.js version, webpack version and Operating System.
NodeJS v6.9.5
Windows 10.0.14393
Webpack 2.2.1
UglifyJs2 doesn't support ES5 yet.
As workaround run npm install https://github.com/mishoo/UglifyJS2#harmony to get the harmony branch. Or transpile with babel.
| gharchive/issue | 2017-02-20T16:16:40 | 2025-04-01T06:46:13.810341 | {
"authors": [
"sebfindling",
"sokra"
],
"repo": "webpack/webpack",
"url": "https://github.com/webpack/webpack/issues/4318",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1047781880 | fix regexp in DefaultStatsPrinterPlugin.js
What kind of change does this PR introduce?
bugfix
fixes #14678
Did you add tests for your changes?
no
Does this PR introduce a breaking change?
nothing
What needs to be documented once your changes are merged?
nothing
Thanks
| gharchive/pull-request | 2021-11-08T18:32:56 | 2025-04-01T06:46:13.813174 | {
"authors": [
"sokra",
"vankop"
],
"repo": "webpack/webpack",
"url": "https://github.com/webpack/webpack/pull/14679",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
417022546 | Adding support for output.libraryTarget "system"
Resolves #8833. cc @sokra @guybedford
What kind of change does this PR introduce?
A new feature
Did you add tests for your changes?
Yes
Does this PR introduce a breaking change?
No -- completely new feature
What needs to be documented once your changes are merged?
We'll need to update the output.libraryTarget docs to indicate that "system" is now a valid output option.
@sokra any feedback on this pull request? Are the tests sufficient? Any things I should change in SystemMainTemplatePlugin?
I'm attempting to fix the codecov integration failure, but unfortunately am having troubles clicking on the Details link -- codecov's website is giving a network protocol error and I think is timing out or something.
CI failure seems to be due to flaky unrelated issues as far as I can tell.
Just got the tests to pass -- the azure ones are flaky but i finally got a build where everything succeeded! This is ready for review from a webpack maintainer.
Just updated this with some feedback from sokra. Now using Template.asString and Template.indent. Also, changed spaces -> tabs in my test files.
Any update on this? @sokra
Thanks
Amazing thank you so much @joeldenning and @sokra for the work on this one.
Oo thanks! I’m excited to use this.
| gharchive/pull-request | 2019-03-04T22:29:52 | 2025-04-01T06:46:13.818275 | {
"authors": [
"guybedford",
"joeldenning",
"sokra"
],
"repo": "webpack/webpack",
"url": "https://github.com/webpack/webpack/pull/8864",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2193276432 | Allow using inline code comments to override config
It would be nice if we could use code comments to tell knip to ignore a file, export, variable, etc, similar to how eslint works
Knip tries not to introduce new/proprietary syntax: https://knip.dev/explanations/why-use-knip#unobtrusive
You still have some options, though:
https://knip.dev/reference/cli#--experimental-tags
https://knip.dev/reference/jsdoc-tsdoc-tags
Knip tries not to introduce new/proprietary syntax: https://knip.dev/explanations/why-use-knip#unobtrusive
You still have some options, though:
https://knip.dev/reference/cli#--experimental-tags
https://knip.dev/reference/jsdoc-tsdoc-tags
Missed that part in the docs, thanks for the quick reply!
| gharchive/issue | 2024-03-18T20:45:52 | 2025-04-01T06:46:13.822740 | {
"authors": [
"sbalay",
"webpro"
],
"repo": "webpro/knip",
"url": "https://github.com/webpro/knip/issues/557",
"license": "ISC",
"license_type": "permissive",
"license_source": "github-api"
} |
138120450 | run collider with tls failed
Hello,
I'm building a signal service with collider, when i run
collider_root/bin/collidermain -port=8089 -tls=true,
it says:
Run: http2: TLSConfig.CipherSuites index 4 contains an HTTP/2-approved cipher suite (0xc030), but it comes after unapproved cipher suites. With this configuration, clients that don't support previous, approved cipher suites may be given an unapproved one and reject the connection.
then it stops. plz help.
Have you tried updating go?
Hi,
I comes with the same error, have you resolved it?
What version of Go are you using?
What happens if you remove https://github.com/webrtc/apprtc/blob/master/src/collider/collider/collider.go#L58 and then run it again?
@zhangyashen @TaoistKing I've got the same issue when I was using Go 1.6. Switch to 1.5.3 and problem solved.
I see, most likely means they are deprecating less secure cipher suits. These were added for iOS i believe, @jiayliu @tkchin do you know if we can remove the cipher suites Go 1.6 complains about?
We can remove unapproved cipher suite as long as all clients support remaining suites. On iOS:
https://developer.apple.com/library/prerelease/ios/documentation/General/Reference/InfoPlistKeyReference/Articles/CocoaKeys.html
We care about this line:
The negotiated TLS connection cipher suite must support forward secrecy (FS) and be one of the following:
TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384
TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256
TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384
TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA
TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256
TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA
TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384
TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256
TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384
TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256
TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA
@TaoistKing , can you help me install on my machine the collider or websocket?
@michaelluk @KaptenJansson I've got the same issue for http2 running on ubuntu server, please help.
It works on any go version but it require one cipher which is "TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256"
And there's a list of bad ciphers that we will avoid https://github.com/bradfitz/http2/blob/b6255645465a25b25f804acb9b3a54009e80c2a4/server.go#L228-L302
@TaoistKing does it work for you using the cipher above?
| gharchive/issue | 2016-03-03T09:11:37 | 2025-04-01T06:46:13.839974 | {
"authors": [
"KaptenJansson",
"TaoistKing",
"mechazod",
"michaelluk",
"tkchin",
"zhangyashen"
],
"repo": "webrtc/apprtc",
"url": "https://github.com/webrtc/apprtc/issues/272",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.