id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
649846258 | unable to use on tar.xz
I am getting checksum and other errors when trying to read a tar.xz.
Is it not supported ?
microtar only supports tar archives -- tar.xz, tar.gz and friends are compressed tar archives which would have to be decompressed first, the resultant decompressed data should work with microtar.
Thanks for the reply ! can you suggest what might be the best route to do that ? Should I use some other library os System command ?
basically tar collects all the files and creates a single archive, that archive is then compressed 7z format, thus .tar.7z is produced, or .tar.gz in gzip, or .tar.xz when using xz file format...
libarchive is the best library to work with all sorts of archives...
| gharchive/issue | 2020-07-02T11:11:18 | 2025-04-01T06:45:42.334022 | {
"authors": [
"mkandulavm",
"pegvin",
"rxi"
],
"repo": "rxi/microtar",
"url": "https://github.com/rxi/microtar/issues/20",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1421465579 | Not consistent behaviour with column mutator and accessor
Found this possible issue while trying to use together Spatie's Laravel Translatable and Orbit but I think it's extensive to any stored column with mutator/accessor. (maybe related to this comment? https://github.com/ryangjchandler/orbit/pull/56#issuecomment-819048606)
I'm not sure if the behaviour for mutator/accessor is the expected for simple cases, but it doesn't work with Laravel Translatable.
Steps to reproduce
To keep this example simple, I'll use a simple attribute. Create a Test model with this schema and attribute:
class Test extends Model
{
use Orbital;
protected $fillable = [
'title',
'amount',
];
public static function schema(Blueprint $table)
{
$table->id();
$table->integer('amount');
}
protected function amount(): Attribute
{
return Attribute::make(
get: fn($value) => $value / 100,
set: fn($value) => $value * 100,
);
}
}
When you create a new model like this:
$model = Test::create([
'amount' => 12.34,
]);
If you retrieve the model later, it'll work and pass the tests, because it's stored correctly on the SQLite as 1234 but not in the Markdown, stored as amount: 12.34.
Even if you remove the cache after creating and before retrieving it still works because the code uses the setter to create the SQLite.
If this is the expected behaviour, it won't work with packages like Laravel Translatable because they're managing JSON columns with multiple languages but assigning a default locale on the getter, so you lose other translations on every save. I can expand this issue with a replicable example for Laravel Translate if the previous behaviour for attributes is the expected one.
Proposed solution
Don't use the getter during the file store and the setter during the get from the file to create the SQLite in order to store the same original content on SQLite and files.
Don't use the Model getter to store the file. That would keep the 1234 value on the file and also the JSON string for translations. That happens on this line of the FileDriver: https://github.com/ryangjchandler/orbit/blob/2a812a57ce295283e3fe7840bf51e7c08990bc11/src/Drivers/FileDriver.php#L88
Don't use the setter to retrieve values from files to create the SQLite: https://github.com/ryangjchandler/orbit/blob/2a812a57ce295283e3fe7840bf51e7c08990bc11/src/Concerns/Orbital.php#L185
With these changes, SQLite and files should store the data the same way. I've found some collateral benefits of this, like datetimes that would be stored in the same format as the database, but that would be a breaking change for existing installations.
I've tried to find a non-breaking solution to make it compatible with Spatie's Laravel Translatable but didn't find it.
I can work on a PR if you confirm which should be the expected behaviour for accessors and mutators on Orbit. Thanks!
Thanks for looking into this @dtorras.
It definitely looks like a problem with how data is transformed during serialisation.
Regarding the breaking changes, I think this could fit into a new major version which has been worked on sparsely.
This has been fixed in 2.x.
| gharchive/issue | 2022-10-24T21:20:06 | 2025-04-01T06:45:42.371590 | {
"authors": [
"dtorras",
"ryangjchandler"
],
"repo": "ryangjchandler/orbit",
"url": "https://github.com/ryangjchandler/orbit/issues/148",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
3425279 | PEP8 Cleanup, More Verbosness
Rid of a lot of libs not being used
Changing Exceptions to prefix with "Twython", just safer in case
other apps have "AuthError", etc. for some reason.
I'll merge this, as I think the Exception name changes are good, but the README (at the very least) should be updated to note that this changed. I'll probably do a larger release with these changes so it's evident that... well, something public-facing changed.
Just wanted to let you know that the Exceptions names WERE changed without a version bump or README update. :P
| gharchive/issue | 2012-02-28T20:17:09 | 2025-04-01T06:45:42.393999 | {
"authors": [
"michaelhelmick",
"ryanmcgrath"
],
"repo": "ryanmcgrath/twython",
"url": "https://github.com/ryanmcgrath/twython/issues/62",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
602418732 | Graph cli
fixed error from build 213.1
change option to argument
| gharchive/pull-request | 2020-04-18T09:27:28 | 2025-04-01T06:45:42.407738 | {
"authors": [
"viktordremio"
],
"repo": "rymurr/dremio_client",
"url": "https://github.com/rymurr/dremio_client/pull/52",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1320793548 | Feature/update deps resolves #7
Updates package-lock.json to version 2 (npm 8/node 16)
Updates dependencies (npm outdated and npm audit)
Thanks. I squashed and merged your changes - https://github.com/rzaba0/soldat-launcher/commit/71ee08b2e1f2bdeb64b2228e64d5b7a160a5a645
| gharchive/pull-request | 2022-07-28T11:31:33 | 2025-04-01T06:45:42.438699 | {
"authors": [
"Shoozza",
"rzaba0"
],
"repo": "rzaba0/soldat-launcher",
"url": "https://github.com/rzaba0/soldat-launcher/pull/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
580084307 | Doesn't work well with Spring Tool Suite, it slows that tool down
Refer : https://github.com/spring-projects/sts4/issues/417
In all fairness, we don't know for sure whether Lombok is the cause of the problem as @bismoy2013 has not tried whether adding/removing lombok makes a difference.
We also have not yet recieved information such as JVM stacktraces that could help us trouble-shoot and determine why the UI thread is being blocked and pinpoint a potential cause.
The problem may be caused by Lombok or it may not. We really don't know either way.
Even if Lombok is the problem, we probably need a project to reproduce, because as it seems to affect only a minority of users, it is likely that this is caused by a single Lombok annotation or a special way how one annotation is used.
@janrieke I have closed the issue because I don't think I would be able to devote much time tracing logs and stuff , I can clearly understand if this issue remains unresolved because in the end it's free , be it's lombok or Spring Tool Suite . I have downloaded another version of STS provided by Martin and it was running fine, as I added lombok , it started behaving bit weirdly , just basic "method implementation" navigation by doing (Ctrl+Click on method name) , STS immediately freezes . This is an open source community , I do not have the right to judge anyone as I am not paying , but I just thought if intellij works fine , STS also should have worked out well. I am providing a link where people did discuss about lombok and STS issue in the past - https://github.com/rzwitserloot/lombok/issues/1717
Your not supposed to go through logs etc. However, it would be nice if you could give at least information that could help us reproduce the issue:
Which version of Lombok do you have installed (you'll find it in the eclipse about dialogue)?
Which version(s) of STS are affected?
Which Java version do you use?
Lombok-Version: 1.18.12
STS Version : 4
Java version : OpenJDK 1.8.0_242 , did not try with others .
Hope this helps.
STS Version : 4
That is a bit vague. You can get a more precise version number by opening the 'About' dialog from the 'Help' menu. If you are not sure what to look for... click on the 'Installation Details' button in the bottom right of the dialog. Select the 'Configuration' tab and just copy-paste all the text that appears. It contains every possible information about the Eclipse installation, from the version of OS, JVM you are running to exact version of every single individual Eclipse plugin you have installed and more.
Version: 4.5.1.RELEASE
Build Id: 202001211302
A very crucial bit of information still missing is which version of Eclipse the STS version is based on. In fact there are multiple builds on different versions of Eclipse for every STS release. I am guessing the version/build of Eclipse @bismoy2013 is using was based on Eclipse 4.14. I think Martin also provided him a link with a STS 4.5.1.RELEASE based on Eclipse 4.13 (the reasoning being that if Lombok was the cause it might work better on an older version of Eclipse).
@kdvolder yes this is the same STS that I downloaded from the link provided by Martin . Hope this clarifies .
I beleave the one Martin provided was based on Eclipse 4.13 whereas the 'default' distrubution downloaded from official download page would be based on Eclipse 4.14. The base Eclipse version is also displayed in the 'about' dialog. A lot of confusing could/can be avoided if you just followed my advice on copying all the info from the 'About' dialogs 'Installation Details', 'Configuration' tab. All that information and more is included there.
*** System properties:
applicationXMI=org.eclipse.ui.workbench/LegacyIDE.e4xmi
awt.toolkit=sun.lwawt.macosx.LWCToolkit
com.genuitec.eclipse.theming.ui.isDark=true
com.genuitec.eclipse.theming.ui.isDevStyle=true
com.genuitec.eclipse.theming.ui.isMojaveDark=true
ds.delayed.keepInstances=true
ds.delayed.keepInstances.default=true
eclipse.application=org.eclipse.ui.ide.workbench
eclipse.buildId=4.5.1.202001211302-RELEASE
eclipse.commands=-os
macosx
-ws
cocoa
-arch
x86_64
-showsplash
-launcher
/Applications/SpringToolSuite4.app/Contents/MacOS/SpringToolSuite4
-name
SpringToolSuite4
--launcher.library
/Applications/SpringToolSuite4.app/Contents/MacOS//../Eclipse/plugins/org.eclipse.equinox.launcher.cocoa.macosx.x86_64_1.1.1100.v20190907-0426/eclipse_1902.so
-startup
/Applications/SpringToolSuite4.app/Contents/MacOS//../Eclipse/plugins/org.eclipse.equinox.launcher_1.5.500.v20190715-1310.jar
--launcher.overrideVmargs
-product
org.springframework.boot.ide.branding.sts4
-keyring
/Users/bismoymurasing/.eclipse_keyring
-vm
/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/bin/../jre/lib/server/libjvm.dylib
eclipse.home.location=file:/Applications/SpringToolSuite4.app/Contents/Eclipse/
eclipse.launcher=/Applications/SpringToolSuite4.app/Contents/MacOS/SpringToolSuite4
eclipse.launcher.name=SpringToolSuite4
eclipse.p2.data.area=@config.dir/../p2/
eclipse.p2.profile=DefaultProfile
eclipse.product=org.springframework.boot.ide.branding.sts4
eclipse.startTime=1584079431922
eclipse.stateSaveDelayInterval=30000
eclipse.vm=/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/bin/../jre/lib/server/libjvm.dylib
eclipse.vmargs=-Dosgi.requiredJavaVersion=1.8
-Xms256m
-Xmx1024m
-XX:+UseG1GC
-XX:+UseStringDeduplication
-XstartOnFirstThread
-Dorg.eclipse.swt.internal.carbon.smallFonts
-Xdock:icon=../Resources/sts4.icns
-javaagent:/Applications/SpringToolSuite4.app/Contents/Eclipse/lombok.jar
-Djava.class.path=/Applications/SpringToolSuite4.app/Contents/MacOS//../Eclipse/plugins/org.eclipse.equinox.launcher_1.5.500.v20190715-1310.jar
equinox.init.uuid=true
file.encoding=UTF-8
file.encoding.pkg=sun.io
file.separator=/
ftp.nonProxyHosts=local|.local|169.254/16|.169.254/16
gopherProxySet=false
gosh.args=--nointeractive
guice.disable.misplaced.annotation.check=true
java.awt.graphicsenv=sun.awt.CGraphicsEnvironment
java.awt.printerjob=sun.lwawt.macosx.CPrinterJob
java.class.path=/Applications/SpringToolSuite4.app/Contents/MacOS//../Eclipse/plugins/org.eclipse.equinox.launcher_1.5.500.v20190715-1310.jar:/Applications/SpringToolSuite4.app/Contents/Eclipse/lombok.jar
java.class.version=52.0
java.endorsed.dirs=/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/endorsed
java.ext.dirs=/Users/bismoymurasing/Library/Java/Extensions:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/ext:/Library/Java/Extensions:/Network/Library/Java/Extensions:/System/Library/Java/Extensions:/usr/lib/java
java.home=/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre
java.io.tmpdir=/var/folders/ty/t18v_frd6jg82r1r4zqprk5w0000gn/T/
java.library.path=/Users/bismoymurasing/Library/Java/Extensions:/Library/Java/Extensions:/Network/Library/Java/Extensions:/System/Library/Java/Extensions:/usr/lib/java:.
java.runtime.name=OpenJDK Runtime Environment
java.runtime.version=1.8.0_242-b08
java.specification.name=Java Platform API Specification
java.specification.vendor=Oracle Corporation
java.specification.version=1.8
java.vendor=AdoptOpenJDK
java.vendor.url=http://java.oracle.com/
java.vendor.url.bug=http://bugreport.sun.com/bugreport/
java.version=1.8.0_242
java.vm.info=mixed mode
java.vm.name=OpenJDK 64-Bit Server VM
java.vm.specification.name=Java Virtual Machine Specification
java.vm.specification.vendor=Oracle Corporation
java.vm.specification.version=1.8
java.vm.vendor=AdoptOpenJDK
java.vm.version=25.242-b08
line.separator=
maven.defaultProjectBuilder.disableGlobalModelCache=true
org.eclipse.debug.ui.breakpoints.toggleFactoriesUsed=true
org.eclipse.equinox.simpleconfigurator.configUrl=file:org.eclipse.equinox.simpleconfigurator/bundles.info
org.eclipse.m2e.log.dir=/Users/bismoymurasing/Documents/workspace-spring-tool-suite-4-4.5.1.RELEASE/.metadata/.plugins/org.eclipse.m2e.logback.configuration
org.eclipse.swt.internal.carbon.smallFonts=
org.eclipse.swt.internal.deviceZoom=200
org.eclipse.update.reconcile=false
org.osgi.framework.executionenvironment=OSGi/Minimum-1.0,OSGi/Minimum-1.1,OSGi/Minimum-1.2,JavaSE/compact1-1.8,JavaSE/compact2-1.8,JavaSE/compact3-1.8,JRE-1.1,J2SE-1.2,J2SE-1.3,J2SE-1.4,J2SE-1.5,JavaSE-1.6,JavaSE-1.7,JavaSE-1.8
org.osgi.framework.language=en
org.osgi.framework.os.name=MacOSX
org.osgi.framework.os.version=10.15.3
org.osgi.framework.processor=x86-64
org.osgi.framework.storage=/Applications/SpringToolSuite4.app/Contents/Eclipse/configuration
org.osgi.framework.system.capabilities=osgi.ee; osgi.ee="OSGi/Minimum"; version:List="1.0, 1.1, 1.2",osgi.ee; osgi.ee="JRE"; version:List="1.0, 1.1",osgi.ee; osgi.ee="JavaSE"; version:List="1.0, 1.1, 1.2, 1.3, 1.4, 1.5, 1.6, 1.7, 1.8",osgi.ee; osgi.ee="JavaSE/compact1"; version:List="1.8",osgi.ee; osgi.ee="JavaSE/compact2"; version:List="1.8",osgi.ee; osgi.ee="JavaSE/compact3"; version:List="1.8"
org.osgi.framework.system.packages=java.applet,java.awt,java.awt.color,java.awt.datatransfer,java.awt.dnd,java.awt.event,java.awt.font,java.awt.geom,java.awt.im,java.awt.im.spi,java.awt.image,java.awt.image.renderable,java.awt.print,java.beans,java.beans.beancontext,java.io,java.lang,java.lang.annotation,java.lang.instrument,java.lang.invoke,java.lang.management,java.lang.ref,java.lang.reflect,java.math,java.net,java.nio,java.nio.channels,java.nio.channels.spi,java.nio.charset,java.nio.charset.spi,java.nio.file,java.nio.file.attribute,java.nio.file.spi,java.rmi,java.rmi.activation,java.rmi.dgc,java.rmi.registry,java.rmi.server,java.security,java.security.acl,java.security.cert,java.security.interfaces,java.security.spec,java.sql,java.text,java.text.spi,java.time,java.time.chrono,java.time.format,java.time.temporal,java.time.zone,java.util,java.util.concurrent,java.util.concurrent.atomic,java.util.concurrent.locks,java.util.function,java.util.jar,java.util.logging,java.util.prefs,java.util.regex,java.util.spi,java.util.stream,java.util.zip,javax.accessibility,javax.activation,javax.activity,javax.annotation,javax.annotation.processing,javax.crypto,javax.crypto.interfaces,javax.crypto.spec,javax.imageio,javax.imageio.event,javax.imageio.metadata,javax.imageio.plugins.bmp,javax.imageio.plugins.jpeg,javax.imageio.spi,javax.imageio.stream,javax.jws,javax.jws.soap,javax.lang.model,javax.lang.model.element,javax.lang.model.type,javax.lang.model.util,javax.management,javax.management.loading,javax.management.modelmbean,javax.management.monitor,javax.management.openmbean,javax.management.relation,javax.management.remote,javax.management.remote.rmi,javax.management.timer,javax.naming,javax.naming.directory,javax.naming.event,javax.naming.ldap,javax.naming.spi,javax.net,javax.net.ssl,javax.print,javax.print.attribute,javax.print.attribute.standard,javax.print.event,javax.rmi,javax.rmi.CORBA,javax.rmi.ssl,javax.script,javax.security.auth,javax.security.auth.callback,javax.security.auth.kerberos,javax.security.auth.login,javax.security.auth.spi,javax.security.auth.x500,javax.security.cert,javax.security.sasl,javax.sound.midi,javax.sound.midi.spi,javax.sound.sampled,javax.sound.sampled.spi,javax.sql,javax.sql.rowset,javax.sql.rowset.serial,javax.sql.rowset.spi,javax.swing,javax.swing.border,javax.swing.colorchooser,javax.swing.event,javax.swing.filechooser,javax.swing.plaf,javax.swing.plaf.basic,javax.swing.plaf.metal,javax.swing.plaf.multi,javax.swing.plaf.nimbus,javax.swing.plaf.synth,javax.swing.table,javax.swing.text,javax.swing.text.html,javax.swing.text.html.parser,javax.swing.text.rtf,javax.swing.tree,javax.swing.undo,javax.tools,javax.transaction,javax.transaction.xa,javax.xml,javax.xml.bind,javax.xml.bind.annotation,javax.xml.bind.annotation.adapters,javax.xml.bind.attachment,javax.xml.bind.helpers,javax.xml.bind.util,javax.xml.crypto,javax.xml.crypto.dom,javax.xml.crypto.dsig,javax.xml.crypto.dsig.dom,javax.xml.crypto.dsig.keyinfo,javax.xml.crypto.dsig.spec,javax.xml.datatype,javax.xml.namespace,javax.xml.parsers,javax.xml.soap,javax.xml.stream,javax.xml.stream.events,javax.xml.stream.util,javax.xml.transform,javax.xml.transform.dom,javax.xml.transform.sax,javax.xml.transform.stax,javax.xml.transform.stream,javax.xml.validation,javax.xml.ws,javax.xml.ws.handler,javax.xml.ws.handler.soap,javax.xml.ws.http,javax.xml.ws.soap,javax.xml.ws.spi,javax.xml.ws.spi.http,javax.xml.ws.wsaddressing,javax.xml.xpath,org.ietf.jgss,org.omg.CORBA,org.omg.CORBA_2_3,org.omg.CORBA_2_3.portable,org.omg.CORBA.DynAnyPackage,org.omg.CORBA.ORBPackage,org.omg.CORBA.portable,org.omg.CORBA.TypeCodePackage,org.omg.CosNaming,org.omg.CosNaming.NamingContextExtPackage,org.omg.CosNaming.NamingContextPackage,org.omg.Dynamic,org.omg.DynamicAny,org.omg.DynamicAny.DynAnyFactoryPackage,org.omg.DynamicAny.DynAnyPackage,org.omg.IOP,org.omg.IOP.CodecFactoryPackage,org.omg.IOP.CodecPackage,org.omg.Messaging,org.omg.PortableInterceptor,org.omg.PortableInterceptor.ORBInitInfoPackage,org.omg.PortableServer,org.omg.PortableServer.CurrentPackage,org.omg.PortableServer.POAManagerPackage,org.omg.PortableServer.POAPackage,org.omg.PortableServer.portable,org.omg.PortableServer.ServantLocatorPackage,org.omg.SendingContext,org.omg.stub.java.rmi,org.w3c.dom,org.w3c.dom.bootstrap,org.w3c.dom.css,org.w3c.dom.events,org.w3c.dom.html,org.w3c.dom.ls,org.w3c.dom.ranges,org.w3c.dom.stylesheets,org.w3c.dom.traversal,org.w3c.dom.views,org.w3c.dom.xpath,org.xml.sax,org.xml.sax.ext,org.xml.sax.helpers
org.osgi.framework.uuid=8b776263-78e7-4a0f-95d3-93376d327b09
org.osgi.framework.vendor=Eclipse
org.osgi.framework.version=1.9.0
org.osgi.supports.framework.extension=true
org.osgi.supports.framework.fragment=true
org.osgi.supports.framework.requirebundle=true
os.arch=x86_64
os.name=Mac OS X
os.version=10.15.3
osgi.arch=x86_64
osgi.bundles=reference:file:org.eclipse.osgi.compatibility.state_1.1.600.v20190814-1451.jar,reference:file:org.eclipse.equinox.simpleconfigurator_1.3.300.v20190716-0825.jar@1:start
osgi.bundles.defaultStartLevel=4
osgi.compatibility.bootdelegation=true
osgi.compatibility.bootdelegation.default=true
osgi.configuration.area=file:/Applications/SpringToolSuite4.app/Contents/Eclipse/configuration/
osgi.framework=file:/Applications/SpringToolSuite4.app/Contents/Eclipse/plugins/org.eclipse.osgi_3.15.0.v20190830-1434.jar
osgi.framework.extensions=reference:file:org.eclipse.osgi.compatibility.state_1.1.600.v20190814-1451.jar
osgi.framework.shape=jar
osgi.framework.useSystemProperties=true
osgi.frameworkClassPath=., file:/Applications/SpringToolSuite4.app/Contents/Eclipse/plugins/org.eclipse.osgi.compatibility.state_1.1.600.v20190814-1451.jar
osgi.install.area=file:/Applications/SpringToolSuite4.app/Contents/Eclipse/
osgi.instance.area=file:/Users/bismoymurasing/Documents/workspace-spring-tool-suite-4-4.5.1.RELEASE/
osgi.instance.area.default=file:/Users/bismoymurasing/Documents/workspace-spring-tool-suite-4-4.5.1.RELEASE/
osgi.logfile=/Users/bismoymurasing/Documents/workspace-spring-tool-suite-4-4.5.1.RELEASE/.metadata/.log
osgi.nl=en_IN
osgi.os=macosx
osgi.requiredJavaVersion=1.8
osgi.splashLocation=/Applications/SpringToolSuite4.app/Contents/Eclipse/plugins/org.springframework.boot.ide.branding_4.5.1.202001211302-RELEASE/splash.bmp
osgi.splashPath=platform:/base/plugins/org.springframework.boot.ide.branding
osgi.syspath=/Applications/SpringToolSuite4.app/Contents/Eclipse/plugins
osgi.tracefile=/Users/bismoymurasing/Documents/workspace-spring-tool-suite-4-4.5.1.RELEASE/.metadata/trace.log
osgi.ws=cocoa
path.separator=:
socksNonProxyHosts=local|.local|169.254/16|.169.254/16
sun.arch.data.model=64
sun.awt.enableExtraMouseButtons=true
sun.boot.class.path=/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/resources.jar:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/rt.jar:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/sunrsasign.jar:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/jsse.jar:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/jce.jar:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/charsets.jar:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib/jfr.jar:/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/classes
sun.boot.library.path=/Library/Java/JavaVirtualMachines/adoptopenjdk-8.jdk/Contents/Home/jre/lib
sun.cpu.endian=little
sun.cpu.isalist=
sun.font.fontmanager=sun.font.CFontManager
sun.io.unicode.encoding=UnicodeBig
sun.jnu.encoding=UTF-8
sun.management.compiler=HotSpot 64-Bit Tiered Compilers
sun.os.patch.level=unknown
swt.enable.themedScrollBar=false
swt.enable.wasThemedScrollBar=true
user.country=IN
user.dir=/Applications/SpringToolSuite4.app/Contents/MacOS
user.home=/Users/bismoymurasing
user.language=en
user.name=bismoymurasing
user.timezone=Asia/Kolkata
*** System environment variables:
APP_ICON_2788=../Resources/sts4.icns
HOME=/Users/bismoymurasing
JAVA_STARTED_ON_FIRST_THREAD_2788=1
LOGNAME=bismoymurasing
PATH=/usr/bin:/bin:/usr/sbin:/sbin
SHELL=/bin/zsh
SSH_AUTH_SOCK=/private/tmp/com.apple.launchd.B0dJ8BnEts/Listeners
TMPDIR=/var/folders/ty/t18v_frd6jg82r1r4zqprk5w0000gn/T/
USER=bismoymurasing
XPC_FLAGS=0x0
XPC_SERVICE_NAME=org.springframework.boot.ide.branding.sts4.4896
__CF_USER_TEXT_ENCODING=0x1F5:0x0:0x0
I did not paste the whole thing , just the part that I believe is useful . The whole details would be too big to paste here
For eclipse it is possible to 'run an eclipse inside eclipse', in debug mode, which makes it a lot easier to profile.
However, it's not easy to do profile runs, and STS is both for the community as a while not a big prority (not nearly as commonly used as eclipse itself), nor for @rspilker or I, as we don't use it.
If someone else wants to try to find out, let me know, maybe we can help you a little bit, but most of the heavy lifting would have to be done by you (basically, we can help you get a lombokized eclipse running in a PDE eclipse, and that's as far as we can take you) – otherwise it might be a good long while before we get around this :(
Leaving it open as it is still a problem.
@rzwitserloot I understand you don't feel compelled to invest time into STS specific issue.
However, the question is whether this problem is 'STS specific' though. The truth is that I don't know. But if I were to take a wild guess, I would assume that its probably not. STS is just Eclipse with some additional plugins. So the chances are that if its a problem for STS it is a problem for any Eclipse installation with the same version of JDT.
The way to verify this and know for sure would be to try whether similar problem arises when using a plain Eclipse with Lombok. I'm not sure I can try it (well I can, but TBH I don't know if I know enough to actally be able to reproduce the problem at all).
I.m.o @bismoy2013 who is the one actually experiencing the problem would be the best placed to try whether the problem is STS specific or generally also a problem with a vanilla Eclipse + Lombok setup. I do sense his patience is wearing very thin.
Anyhow, if we can get some clear evidence that the problem is indeed STS specific, it would go a long way in convincingin me or someone else on STS team to invest the time to try and debug it.
As it is though we do not feel like it is our responsibility to debug Lombok, a tool that we do not officially support nor recommend, and do not include with our STS distribution builds. Users can of course install it on their own initiative. and fr be it from us to stop them. However we do not feel responsible for problems it may cause.
@kdvolder you'll be happy to know that millions of eclipses with lombok are out there, every day. Either they're all just accepting the slowness (an extraordinary claim), or nobody is running into this (which makes it either STS specific, or it's not STS specific but nevertheless extremely exotic). It also strongly suggests just faffing about, clicking random buttons, hoping the slowdown occurs is not going to be a fruitful bug hunting exercise either. And that's the only real plan I have here (though, given that it is highly likely STS related, as so far nobody's reported eclipse slowdowns), I'd do that with an STS install and not an eclipse install.
We now have a single plausible report that lombok + some specific STS version has slowdowns, but no real formula to reproduce it. If neither team STS nor lombok currently wants to spend resources on this one, the next step is to wait for more reports, and more detailed reports.
If I or anybody else created the impression that you have a responsibility to reproduce and fix this, I apologize – you certainly do not :) open source licenses are rather explicit in disavowing merchantability and responsibility for pretty much everything.
However, users gonna user, and I'm pretty sure that the % of all lombok users that uses STS is much, much lower than the % of all STS users that uses lombok. Thus: You do not owe an explanation to lombok's maintainers. But you might owe one to STS's users.
Lombok's core maintainers are 2 people who get paid enough to work at minimum wage for a few hours a week max on lombok. I'm pretty sure team STS has more resources than that, so from a perspective of 'who has the most resources to throw at this problem', I don't feel obligated. We must triage our time to work on what's most important to the community and this just isn't making the cut; not without more details on how to reproduce this.
@rzwitserloot I can strongly confirm this is not a lombok issue. It's an internal STS issue , to confirm that I have written an email with log traces to Martin from STS team . The issue was showing up even without lombok addition.
The thing is, even if I want to help further more , I can only help till some extent , sharing personal projects for debugging purpose to external teams (like both of you) would have been too much . I hope @kdvolder agrees, after we are all bounded by any org's rules and guidelines !
Also , there could be many who faces this issue that we all are unaware of, but they wouldn't spend time reporting bugs here, they would easily move to an IDE that works just fine, as simple as that. Many people don't even know the right medium for addressing IDE related bugs .
| gharchive/issue | 2020-03-12T17:08:03 | 2025-04-01T06:45:42.503074 | {
"authors": [
"bismoy2013",
"janrieke",
"kdvolder",
"rzwitserloot"
],
"repo": "rzwitserloot/lombok",
"url": "https://github.com/rzwitserloot/lombok/issues/2395",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
142469260 | pod installation
im having problems installing with pod. it installs and adds your files to my project, but then it doesn't recognize sv() and layout() inside my classes. could you help? thanks
Hey @rostik
Can you confirm your Podfile.lock has - SteviaLayout (2.1.0), which the latest version.
sv() and layout() were introduced recently and their previous counterparts are sv([]) and layout([])
I know it sounds stupid but did you add import SteviaLayout at the top of the file?
adding 2.1.0 to pod file, fixed it, thank you!
On Mon, Mar 21, 2016 at 5:05 PM, S4cha notifications@github.com wrote:
Hey @rostik https://github.com/rostik
Can you confirm your Podfile.lock has - SteviaLayout (2.1.0), which the
latest version.
sv() and layout() were introduced recently and their previous
counterparts are sv([]) and layout([])
I know it sounds stupid but did you add import SteviaLayout at the top of
the file?
—
You are receiving this because you were mentioned.
Reply to this email directly or view it on GitHub
https://github.com/s4cha/Stevia/issues/25#issuecomment-199485604
--
Thanks,
Rostik
| gharchive/issue | 2016-03-21T20:57:15 | 2025-04-01T06:45:42.564833 | {
"authors": [
"rostik",
"s4cha"
],
"repo": "s4cha/Stevia",
"url": "https://github.com/s4cha/Stevia/issues/25",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
502714448 | Data Population in SPS
I am using the annotation approach for writing data to a datablock. When I tried the library I found out that my hardware device, which reads from that datablock, throws an error because a integer variable of that datablock is negative. Indeed, I do not set a negative value to that variable. Also in the TIA portal (monitor mode) the value is positive. For it seems like this variable is negative only for a very short time.
So I am wondering how the data is written to the datablock in the S7. Might it be possible that the data is written "byte-by-byte" so that the variable takes on negative values during the serialize process?
It is possible that the command is not completely "atomic" if the data you are writing is larger than the block size (256 bytes i think...)
How much data are you writing and how does it look like?
I didn't know that atomic values are supported
So it seems that the error is in my algorithm - I will notify you as soon as I have new information
I think I have the problem. It is in my code. It seems like the commands com.github.s7connector.api.S7Serializer.dispense and com.github.s7connector.api.S7Serializer.store are not multi-threadable. So running one thread which executes store and a second one which executes dispense seems not lead to a better performance. This seems to be my issue. Do you have a recommendation how to improve the performance? (one read/write of a datablock takes about 50ms)
It seems like the commands com.github.s7connector.api.S7Serializer.dispense and com.github.s7connector.api.S7Serializer.store are not multi-threadable
I have an issue open about the documentation-rewrite, i'll add a note about that too...
Do you have a recommendation how to improve the performance?
Not really, no. But you could write simultaneously to different data-blocks, but i think the bottle-neck is the S7-CPU though... :/
| gharchive/issue | 2019-10-04T15:52:13 | 2025-04-01T06:45:42.570519 | {
"authors": [
"a13u",
"thomasrudin"
],
"repo": "s7connector/s7connector",
"url": "https://github.com/s7connector/s7connector/issues/72",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1121284456 | Toggle/checkbox don't change state in winicss
Hi.
For some reason it's impossible to enable checkbox/toggle when I use windicss 😞
Here's demo: https://stackblitz.com/edit/vite-tcwtbf?file=index.html
Windi is generating invalid style.
Expected CSS is:
.toggle:checked,.toggle[checked=true]{
--chkbg:hsl(var(--bc));
--tw-bg-opacity:1;
--tw-border-opacity:1;
box-shadow:var(--handleoffset) 0 0 2px hsl(var(--b1)) inset, 0 0 0 2px hsl(var(--b1)) inset, var(--focus-shadow);
}
What Windi generates:
toggle, .toggle[checked=true] {
--chkbg: hsl(var(--bc));
--tw-bg-opacity: 1;
--tw-border-opacity: 1;
box-shadow: var(--handleoffset) 0 0 2px hsl(var(--b1)) inset, 0 0 0 2px hsl(var(--b1)) inset, var(--focus-shadow);
}
I suggest using Tailwind CSS instead of Windi
| gharchive/issue | 2022-02-01T22:56:36 | 2025-04-01T06:45:42.577709 | {
"authors": [
"saadeghi",
"xorik"
],
"repo": "saadeghi/daisyui",
"url": "https://github.com/saadeghi/daisyui/issues/443",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
719856289 | Add config file, themes, colors and multiple columns layout close #2
- Enhance log message of server listening
- Add config.js file
- Remove unnecessary console logs
- Change id of "grad1" to "backgroundCard"
- Change id of "grad2" to "backgroundImage"
- Add "theme" parameter
- Enhance README.md file
Add a screenshot of working model
can i use html instead of svg
of course... Did you catch the points.
yes but i think it will be easier if i do that with regular html and css
and the height should be change with the card's limit
For Now, I will accept the #8 ... Pull request person needs to participate in Hactoberfest. However, Can we make it more beautiful???
I didn't expect this for hactoberfest. We need to reach it more people like GitHub Readme stats
Could you please check the checkbox that you finish
How could we embed this in the Readme.md file???
Yes sure... But It is not compatible with Markdown Main Aim of this project is to embed the cards in Github Readme Profile cards.
However, you gave a nice try. I will mark this PR as 'hacktoberfest-accepted'.. We will close it next mont.. We achieved the goal now. Good luck with your hacktoberfest
I thank you and appreciate that. I will not delete the request because I hope that we will work with each other to make something better
I thank you and appreciate that. I will not delete the request because I hope that we will work with each other to make something better
Don't forget to start this repo... I already implement like this in React. Check it here. For this project please watch the readme of this project.
We almost there
| gharchive/pull-request | 2020-10-13T04:41:30 | 2025-04-01T06:45:42.619147 | {
"authors": [
"AbrahemAlhofe",
"sabesansathananthan"
],
"repo": "sabesansathananthan/github-readme-medium-card",
"url": "https://github.com/sabesansathananthan/github-readme-medium-card/pull/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1437369379 | Transition to new feature viewer
The nextProt feature viewer seems to be barely maintained and not ideal for our purpose.
The typescript version looks much more promising.
However, our dependency combination seems to be incompatible as they require a very old version of d3.
If we put in the work to fix it, their library would be a better choice.
The currently up-to-date sib-nextProt component is also unusable, as it includes unwanted UI elements.
| gharchive/issue | 2022-11-06T11:15:27 | 2025-04-01T06:45:42.645594 | {
"authors": [
"t03i"
],
"repo": "sacdallago/embed.predictprotein.org",
"url": "https://github.com/sacdallago/embed.predictprotein.org/issues/61",
"license": "AFL-3.0",
"license_type": "permissive",
"license_source": "github-api"
} |
119827379 | org.jinstagram.auth.oauth.InstagramService getAccessToken missing proxy
There doesn't seem to be any way to pass a proxy in to a OAuthService.
I don't know where the best place to put it is, either in the org.jinstagram.auth.InstagramAuthService builder or in the org.jinstagram.auth.oauth.InstagramService, adding a proxy() method, or expose the request somehow.
Perhaps the org.jinstagram.auth.InstagramApi should be the carrier from the InstagramAuthService to the InstagramService
| gharchive/issue | 2015-12-01T22:54:50 | 2025-04-01T06:45:42.659594 | {
"authors": [
"dariusj18"
],
"repo": "sachin-handiekar/jInstagram",
"url": "https://github.com/sachin-handiekar/jInstagram/issues/133",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1063270379 | 水平スケールでのディスクのクローン
現在の水平スケールの実装であるServerGroupではスケールアウト時にテンプレートを元にサーバ+ディスクを毎回作成している。
これに対し、元になる1台をクローンしていく方式もサポートしたい。
実現方法:
既存のServerGroupでtemplateの代わりに元になるサーバの情報を指定可能にする?
新たにServerGroupの派生を作る?
v0.2.1の時点ではtemplate.disks.source_diskが指定可能。(ID or Name)
これを使えば特に改修せずとも対応できそう。
検証 & 必要に応じてドキュメント化する。
https://github.com/sacloud/autoscaler/issues/242#issuecomment-988496081 の方法で対応可能/検証済み。
https://github.com/sacloud/docs.usacloud.jp/issues/101 でドキュメント追加対応する。
| gharchive/issue | 2021-11-25T08:07:38 | 2025-04-01T06:45:42.662646 | {
"authors": [
"yamamoto-febc"
],
"repo": "sacloud/autoscaler",
"url": "https://github.com/sacloud/autoscaler/issues/242",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2263212473 | Add settlementContract to Order
Summary
Changes
Closing this one since it's part of an undocumented API from CoW which might be removed (so we can't rely on this field).
| gharchive/pull-request | 2024-04-25T10:15:30 | 2025-04-01T06:45:42.689939 | {
"authors": [
"fmrsabino"
],
"repo": "safe-global/safe-client-gateway",
"url": "https://github.com/safe-global/safe-client-gateway/pull/1457",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2715330849 | [New chain]: Swell Network Mainnet canonical 1.3.0 contracts
Summary
https://github.com/safe-global/safe-deployments/pull/894
Chain ID
1923
Chain detail URL
https://chainlist.org/chain/1923
RPC URL
https://swell-mainnet.alt.technology
Blockscout Client URL
https://explorer.swellnetwork.io/api/v1/graphql
Etherscan Client URL
No response
Etherscan Client API URL
No response
Version
1.3.0
Address (Master copy)
0xd9Db270c1B5E3Bd161E8c8503c55cEABeE709552
Deployment Tx hash (Master copy)
0xdea548203b663a410f073ea4636437c0b09f2cc1369b54a9675b9beb79b42bd5
Block explorer URL (Master copy)
No response
Address (Master copy L2)
0x3E5c63644E683549055b9Be8653de26E0B4CD36E
Deployment Tx hash (Master copy L2)
0x9699b29f29f63f9c018ca11ea3b2607488dcfdb7b785d207577a8e86e6918b45
Block explorer URL (Master copy L2)
No response
Address (Proxy factory)
0xa6B71E26C5e0845f74c812102Ca7114b6a896AB2
Deployment Tx hash (Proxy factory)
0x4f61ad72f7744d79b6498470d4373978016e36a1c51486677c42e8f882acb099
Block explorer URL (Proxy factory)
No response
The Blockscout API is actually hosted on a different domain:
/execute
| gharchive/issue | 2024-12-03T15:25:03 | 2025-04-01T06:45:42.696909 | {
"authors": [
"ElvisKrop",
"falvaradorodriguez"
],
"repo": "safe-global/safe-eth-py",
"url": "https://github.com/safe-global/safe-eth-py/issues/1460",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1266040490 | fix: hover color of TxShareButton
What it solves
The hover color of the share TX Button was not matching the figma one.
Resolves https://github.com/safe-global/safe-react/pull/3943#issuecomment-1150910974
How this PR fixes it
Changes the hover color
How to test it
hover the share tx button
Analytics changes
Screenshots
Pull Request Test Coverage Report for Build 2468247993
0 of 0 changed or added relevant lines in 0 files are covered.
No unchanged relevant lines lost coverage.
Overall coverage remained the same at 36.613%
Totals
Change from base Build 2467061384:
0.0%
Covered Lines:
3845
Relevant Lines:
9542
💛 - Coveralls
Tested, looks good ✅ | gharchive/pull-request | 2022-06-09T12:13:58 | 2025-04-01T06:45:42.703282 | {
"authors": [
"coveralls",
"katspaugh",
"schmanu"
],
"repo": "safe-global/safe-react",
"url": "https://github.com/safe-global/safe-react/pull/3947",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2122950566 | 4337 - Refactor indexers into own package
Currently indexers are on the history package for the transaction service. Analize if they should belong to their own app inside the Django project and move refactor if it's posible
It seems like it will be more difficult than expected at first, as dependency between indexers and history models is quite tight, and we also cannot move the models to other app without breaking the database schema
| gharchive/issue | 2024-02-07T12:42:00 | 2025-04-01T06:45:42.704556 | {
"authors": [
"Uxio0"
],
"repo": "safe-global/safe-transaction-service",
"url": "https://github.com/safe-global/safe-transaction-service/issues/1856",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1366636123 | Airdrop farmer report
Related Safe Addresses
0xfed79ad04bc452420e6df9f40625cec8e8a3373d
0xf03a139c3d0d26a07e2939204ae7f3ea1dbd9919
0xeae5e0d63285f24f3a089fdb6a0082ff6801eba1
0xde7a69f62a0746cf93bf97bca16038d05fb700e5
0xBFA3e74eCAC5b2Ecf20B2BaFcdc02Cd659435a45
0xadca2254b9b40a48ef44140482e912a03cbc7f5d
0xab61f7cb67dbce3612f3eefd617b2df2a27d5215
0x91155176f8911c6a34dbca8537a59c2c4816d2d3
0x8cb1f9ad02e57eca18065d83efdfb6ff0c57392c
0x7f630885074b976570152a712589e0ceca9086d3
0x7ec0740988c8d01ef4acfb8ac280dbc7b1d2462b
0x7c3d8bc495b7d2dcc995a3a0dc6444480483988c
0x70c9fd1651ca952314965764012ab63e60fd4eb9
0x692a7d771c69bf1664690b5a072e04f14254aece
0x676758b7c610acec42d47e414d5335cf6fa11899
0x6673fa2045ae0324c68c0cc6f66e7d91c446d728
0x61f744a03b89f7267d26839d4913693c58eadbde
0x5deaba8f3d9ac196df20987987a3b15eac629927
0x52b14b11ef687dbad8e422d655a45ea551948760
0x521fe1bd7dec39362809cac1a237629d458e459b
0x4153f0145d48bbd3eb2438dbe1d4535734ec78e3
0x3204596e8cd53a791767340c93fb5ee2f11b5840
0x30cd6df997d3c153624fcc560823a32e8886d69a
0x1deafddfc62c7d6113204adff6fda8a2d2e55a69
0x1d02f0a433a999a777d1b11311c4bc9fac33af65
0x199aa8a209884ffb1e0849eaeb072bce573bf871
Reasoning
Created from same address 0x20c05CF93223E39d75B6266b4044E2e5C588d089
Very few transactions only 1 or 2.
Transfer tiny size ETH like 0.001
No erc20 token transfer.
Methodology
with target as (
select *
from url('https://raw.githubusercontent.com/safe-global/safe-user-allocation-reports/main/safe_user_allocations_reworked.csv', CSVWithNames)
), transactions as (
select t1.*
from ethereum.transactions t1
join target t2
on t1.to_address = t2.safe_address
)
select distinct to_address
from transactions
where from_address = '0x20c05CF93223E39d75B6266b4044E2e5C588d089'
Safe Address
0xfADD9521EbA86Df0D89B92C5D105f445E3CA9479
Created from same address 0x20c05CF93223E39d75B6266b4044E2e5C588d089
Very few transactions only 1 or 2.
Transfer tiny size ETH like 0.001
The following 3 addresses do more than
https://gnosis-safe.io/app/eth:0x70c9fd1651ca952314965764012ab63e60fd4eb9/transactions/history
https://gnosis-safe.io/app/eth:0x692a7d771c69bf1664690b5a072e04f14254aece/transactions/history
https://gnosis-safe.io/app/eth:0x4153f0145d48bbd3eb2438dbe1d4535734ec78e3/transactions/history
Nevertheless, thanks for the report @0x9J , we've verified that the following addresses seems to be airdrop farming.
0xfed79ad04bc452420e6df9f40625cec8e8a3373d
0xf03a139c3d0d26a07e2939204ae7f3ea1dbd9919
0xeae5e0d63285f24f3a089fdb6a0082ff6801eba1
0xde7a69f62a0746cf93bf97bca16038d05fb700e5
0xBFA3e74eCAC5b2Ecf20B2BaFcdc02Cd659435a45
0xadca2254b9b40a48ef44140482e912a03cbc7f5d
0xab61f7cb67dbce3612f3eefd617b2df2a27d5215
0x91155176f8911c6a34dbca8537a59c2c4816d2d3
0x8cb1f9ad02e57eca18065d83efdfb6ff0c57392c
0x7f630885074b976570152a712589e0ceca9086d3
0x7ec0740988c8d01ef4acfb8ac280dbc7b1d2462b
0x7c3d8bc495b7d2dcc995a3a0dc6444480483988c
0x676758b7c610acec42d47e414d5335cf6fa11899
0x6673fa2045ae0324c68c0cc6f66e7d91c446d728
0x61f744a03b89f7267d26839d4913693c58eadbde
0x5deaba8f3d9ac196df20987987a3b15eac629927
0x52b14b11ef687dbad8e422d655a45ea551948760
0x521fe1bd7dec39362809cac1a237629d458e459b
0x3204596e8cd53a791767340c93fb5ee2f11b5840
0x30cd6df997d3c153624fcc560823a32e8886d69a
0x1deafddfc62c7d6113204adff6fda8a2d2e55a69
0x1d02f0a433a999a777d1b11311c4bc9fac33af65
0x199aa8a209884ffb1e0849eaeb072bce573bf871
@tschubotz hi ser, it seems i found this airdrop farmer first, plz check!
https://github.com/safe-global/safe-user-allocation-reports/issues/56
@tschubotz I am the first to summit and have more information about this report.
https://github.com/safe-global/safe-user-allocation-reports/issues/56
@tschubotz I am the first to summit and have more information about this report. https://github.com/safe-global/safe-user-allocation-reports/issues/56
Hi @farmerxx, your edit that added the addresses was done on Sep 08, 17:38 UTC, while @0x9J did an edit adding the last addresses on Sep 08, 17:36 UTC. 2 minutes before you, arguably a very unfortunate situation but "first come, first serve".
@tschubotz I am the first to summit and have more information about this report. https://github.com/safe-global/safe-user-allocation-reports/issues/56
Hi @farmerxx, your edit that added the addresses was done on Sep 08, 17:38 UTC, while @0x9J did an edit adding the last addresses on Sep 08, 17:36 UTC. 2 minutes before you, arguably a very unfortunate situation but "first come, first serve".
I changed my title on Sep 08, 16:18 UTC, which you can see above.
Because it is my first time to submit issue in github and i put all my information in the notes so nobody can see that, but the time i change my title don't lie, it shows the time i submitted this address.
@tschubotz I am the first to summit and have more information about this report. https://github.com/safe-global/safe-user-allocation-reports/issues/56
Hi @farmerxx, your edit that added the addresses was done on Sep 08, 17:38 UTC, while @0x9J did an edit adding the last addresses on Sep 08, 17:36 UTC. 2 minutes before you, arguably a very unfortunate situation but "first come, first serve".
I changed my title on Sep 08, 16:18 UTC, which you can see above. Because it is my first time to submit issue in github and i put all my information in the notes so nobody can see that, but the time i change my title don't lie, it shows the time i submitted this address.
Hi @farmerxx
Sorry. Rule is rule. Try use Github more often next time.
The rule is i submitted this address in titile first.
@tschubotz I am the first to summit and have more information about this report. https://github.com/safe-global/safe-user-allocation-reports/issues/56
Hi @farmerxx, your edit that added the addresses was done on Sep 08, 17:38 UTC, while @0x9J did an edit adding the last addresses on Sep 08, 17:36 UTC. 2 minutes before you, arguably a very unfortunate situation but "first come, first serve".
I changed my title on Sep 08, 16:18 UTC, which you can see above. Because it is my first time to submit issue in github and i put all my information in the notes so nobody can see that, but the time i change my title don't lie, it shows the time i submitted this address.
Hi @farmerxx Sorry. Rule is rule. Try use Github more often next time.
The rule is i submitted this address in titile first ser, the time in title don't lie.
@tschubotz I am the first to summit and have more information about this report. https://github.com/safe-global/safe-user-allocation-reports/issues/56
Hi @farmerxx, your edit that added the addresses was done on Sep 08, 17:38 UTC, while @0x9J did an edit adding the last addresses on Sep 08, 17:36 UTC. 2 minutes before you, arguably a very unfortunate situation but "first come, first serve".
I changed my title on Sep 08, 16:18 UTC, which you can see above. Because it is my first time to submit issue in github and i put all my information in the notes so nobody can see that, but the time i change my title don't lie, it shows the time i submitted this address.
Hi @farmerxx Sorry. Rule is rule. Try use Github more often next time.
The rule is i submitted this address in titile first ser, the time in title don't lie.
Hi.
I believe rule of FCFS is the time of report submitted not only the tittle submitted.
@Yarosn Agree. Besides this, i think FCFS is the time you submitted a vaild report which contains 100% sybil attackers, not a report which contains lots of real users, Hop Identify Airdrop Farmers challenge taught me this rule :)
| gharchive/issue | 2022-09-08T16:18:32 | 2025-04-01T06:45:42.725682 | {
"authors": [
"0x9J",
"Parkcora",
"Yarosn",
"d3centra",
"farmerxx",
"tschubotz"
],
"repo": "safe-global/safe-user-allocation-reports",
"url": "https://github.com/safe-global/safe-user-allocation-reports/issues/58",
"license": "CC0-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1519122572 | Implement zip and unzip commands
This PR implements zip/unzip commands as per #223. The only exception is that the maps (--path-map) for zip are not implemented (yet).
wow, thanks a lot. i wanted this one for a long time but was too lazy to implement :)
I'll review it and provide you feedback
wow, thanks a lot. i wanted this one for a long time but was too lazy to implement :)
I'll review it and provide you feedback
Yeah, sorry for such a huge delay
@Red-Teapot loos great. thanks a lot. merging
| gharchive/pull-request | 2023-01-04T14:58:51 | 2025-04-01T06:45:42.743488 | {
"authors": [
"Red-Teapot",
"sagiegurari"
],
"repo": "sagiegurari/duckscript",
"url": "https://github.com/sagiegurari/duckscript/pull/294",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
608625466 | Needed the current hash
Hi I wanted the last commit sha and shortend commit sha.
thanks for the work btw!
cool. will check it out.
done and released in version 0.1.2
thanks a lot for the feedback.
println!(
"Last Commit Hash: {}",
info.head.last_commit_hash.unwrap_or("Unknown".to_string())
);
println!(
"Last Commit Hash (short): {}",
info.head
.last_commit_hash_short
.unwrap_or("Unknown".to_string())
);
Thanks, will give it a go this evening.
Sent with GitHawk
Works a treat - thank you!
Sent with GitHawk
thanks for validating
| gharchive/issue | 2020-04-28T21:23:21 | 2025-04-01T06:45:42.746550 | {
"authors": [
"gilescope",
"sagiegurari"
],
"repo": "sagiegurari/git_info",
"url": "https://github.com/sagiegurari/git_info/issues/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2580545138 | 🛑 Harikrusna International is down
In c49018b, Harikrusna International (https://shreeharikrushnaintl.com) was down:
HTTP code: 503
Response time: 452 ms
Resolved: Harikrusna International is back up in 6dc222a after 17 minutes.
| gharchive/issue | 2024-10-11T06:28:03 | 2025-04-01T06:45:42.749106 | {
"authors": [
"vanpariyar"
],
"repo": "sahajananddigital/status",
"url": "https://github.com/sahajananddigital/status/issues/910",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
137343663 | /auth/facebook not found
Sorry for my ignorance. i have included Satellizer for login purposes and when clicked any of the Social networks setup all run fine only that when receive the response, it tries to get access to an individual URL (http://www.mydomain.com/auth/facebook) in my case.
I do not know what should i do there? What kind of script?
I only need to get Email and basic data from the user. I thought that once response was received I could call to the Account Factory to get such data.
When you successfully logged in, you need to store user's information on your database, right? The script is to get the data from provider (in your case, Facebook) and then store it to your database.
Check server implementations here
https://github.com/sahat/satellizer/tree/master/examples/server
Hi ibrahimyu, thanks. I understand now.
I've been having a look within such directory, in /php, and I do not find anything.
How it's supposed Facebook or any other Social Network send you the data?
Maybe in $_GET method?
@mavillar Are you familiar with Laravel framework?
All OAuth handlers are located in the AuthController.php: https://github.com/sahat/satellizer/blob/master/examples/server/php/app/Http/Controllers/AuthController.php
One line 100 is where you will find implementation of the URL handler for POST /auth/facebook.
Satellizer (client-side library) only helps you with launching the popup window and obtaining code (authorization code) parameter, the rest of the OAuth flow happens on the back-end. That's where you exchange that authorization code for access token, then use that access token to obtain user's profile information, save it to database, and return JSON web token back to the client -side that keeps you authenticated. JSON Web Token is specific to your app, it keeps users logged in. Access token you get from Facebook on the other hand is a unique time-sensitive token used for Facebook's Graph API to obtain profile information and perform Facebook actions on user's behalf.
@sahat I've never used such framework, even it's the first time heard. I'll have a look but too late to be used in my actual project.
I just need to get the basic data from the user and nothing else.
Closing this issue since it's not a problem with Satellizer.
@mavillar the key is at line 100 in this file: https://github.com/sahat/satellizer/blob/master/examples/server/php/app/Http/Controllers/AuthController.php
Basically:
Satellizer gives you the code parameter to /auth/facebook. Use it to send GET request to facebook (line 112).
From that, facebook gives you an access token. Use this access token to get whatever you want. For example, take user's basic data (line 119).
Save that data to database (line 140).
After gathering user data, make token based on your user's data.
Return that token to satellizer. You're logged in!
Please adapt it to your needs.
Cheers!
@ibrahimyu I have been battling with this for long seems Satellizer is not posting to /auth/facebook after authentication and access is granted from facebook.com
Any suggestion?
When the popup closes I get this
access_token
:
"EAAJsVunI6ecBAFd8S9BFr7E0X85EBp7buSP1ZCeggjmpCJnvppXdRlZBqUo59fVGWNKoZClAg55qiY2jOJC1uRg5qZCChqWPFKKuLO9ivNQmvte2ZC3jlk5EndOL0j8g8PrGbeOuJtpZCt09p9gD1hAUMKc4ZBoC268aw422YbAG8uwmZBR21N9m"
expires_in
:
"5220"
from $auth.authenticate('facebook') promise
| gharchive/issue | 2016-02-29T19:14:58 | 2025-04-01T06:45:42.759460 | {
"authors": [
"digitlimit",
"ibrahimyu",
"mavillar",
"sahat"
],
"repo": "sahat/satellizer",
"url": "https://github.com/sahat/satellizer/issues/752",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
166097522 | Bower install throws ENORESTARGET error
Hello,
I was trying to use this module in my application but I failed to install it with bower.
I tried : bower install satellizer --save
Then I got this
bower satellizer#* cached https://github.com/sahat/satellizer.git#0.15.0
bower satellizer#* validate 0.15.0 against https://github.com/sahat/satellizer.git#*
bower angular#>=1.3.x && <= 1.5.x not-cached https://github.com/angular/bower-angular.git#>=1.3.x && <= 1.5.x
bower angular#>=1.3.x && <= 1.5.x resolve https://github.com/angular/bower-angular.git#>=1.3.x && <= 1.5.x
bower angular#>=1.3.x && <= 1.5.x ENORESTARGET Tag/branch >=1.3.x && <= 1.5.x does not exist
Any way to resolve this ?
Additional information :
Node version : 4.4.7
Npm version : 3.8.3
Bower version : 1.7.9
the same problem
Available branches: master, unstable
This looks like an issue with the semver in bower.json in the latest update. It should be 1.3.x - 1.5.x or >=1.3.x <= 1.5.x, not >=1.3.x && <= 1.5.x. I'm using bower install satellizer#0.14.1 for now.
Fixed in #875.
| gharchive/issue | 2016-07-18T13:48:28 | 2025-04-01T06:45:42.763399 | {
"authors": [
"TablesawTablesawsen",
"jhumigas",
"kkaabbaa",
"sahat"
],
"repo": "sahat/satellizer",
"url": "https://github.com/sahat/satellizer/issues/873",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1289613265 | 🛑 Apart Hotel Adelboden is down
In d163c88, Apart Hotel Adelboden (https://www.aparthotel-adelboden.ch) was down:
HTTP code: 403
Response time: 121 ms
Resolved: Apart Hotel Adelboden is back up in 90d4ea6.
| gharchive/issue | 2022-06-30T05:40:34 | 2025-04-01T06:45:42.766100 | {
"authors": [
"m43nu"
],
"repo": "sahli-interactive/status.sahli-interactive.ch",
"url": "https://github.com/sahli-interactive/status.sahli-interactive.ch/issues/257",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
930941195 | AOSP-10 vbmeta.zip link in flashing.md is wrong
I believe it should be: https://github.com/sailfishos-sony-tama/main/releases/download/4.1.0.24-aosp10-alpha1/vbmeta.zip
Thank you very much, link fixed inflashing.md
| gharchive/issue | 2021-06-27T14:13:14 | 2025-04-01T06:45:42.775275 | {
"authors": [
"pagism",
"rinigus"
],
"repo": "sailfishos-sony-tama/main",
"url": "https://github.com/sailfishos-sony-tama/main/issues/148",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
26919452 | Autorotate gallery images
sigal seems smart enough to orient portrait images when it is within its own HTML framing. However, the actual jpgs (thumbnail and gallery images) are not rotated. I'd prefer the actual images be rotated so that if I want to use the image or thumbnail elsewhere on my website, it's in the correct orientation.
I'm wondering if there's been any movement on this one?
Not really, except another project which could be useful to save modified EXIF data (and based on Pillow): https://pypi.python.org/pypi/piexif
I haven't seen any progress at pillow issue 520, any fixes available here?
Nothing new here ...
Is there a workaround for this issue?
@jfleach - If you set only one of 'autorotate_images' and 'copy_exif_data', it should be ok: either rotate images and don't copy exif data, or copy exif and rely on browsers to do the rotation.
Pillow 6.0 added the possibility to modify the EXIF metadata and added ImageOps.exif_transpose :)
https://pillow.readthedocs.io/en/stable/releasenotes/6.0.0.html#added-imageops-exif-transpose
Any news?
Not really, the feature is in Pillow but time is lacking to update the code to use it.
I just looked at the code in this area. For me the images themselves get rotated correctly, but the thumbnails not. It looks like Transpose is missing from the thumbnail generation, is that correct?
I just looked at the code in this area. For me the images themselves get rotated correctly, but the thumbnails not. It looks like Transpose is missing from the thumbnail generation, is that correct?
That seems to be the case, I'm seeing the same thing over here.
That seems to be the case, I'm seeing the same thing over here.
Indeed. I added a parameter autorotate to generate_thumbnail and a passage to call Transpose in the same way as in generate_image and it seems to work.
Indeed. I added a parameter autorotate to generate_thumbnail and a passage to call Transpose in the same way as in generate_image and it seems to work.
This is another issue, but I have seen that recently too and committed a fix (a77b2e3f2e9d5ab2d79f6e2bd845dfc652f8aaf6).
That should happen only with use_orig = True because otherwise the thumbnails are created from the resized images which are rotated.
This is another issue, but I have seen that recently too and committed a fix (a77b2e3f2e9d5ab2d79f6e2bd845dfc652f8aaf6).
That should happen only with use_orig = True because otherwise the thumbnails are created from the resized images which are rotated.
| gharchive/issue | 2014-02-04T22:00:10 | 2025-04-01T06:45:42.791042 | {
"authors": [
"hjbaader",
"jfleach",
"reagle",
"saimn",
"virtadpt"
],
"repo": "saimn/sigal",
"url": "https://github.com/saimn/sigal/issues/72",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
821888885 | Create Unit Converter Calculator
Create a Unit Converter Calculator to convert Units
Length
Millimeter
Centimeter
Meter
Kilometer
Inch
Foot
Yard
Mile
Nautical mile
Mil
Area
Acre
Are
Hectare
Square centimeter
Square foot
Square inch
Square meter
Volume
Gallon (US)
Gallon (UK)
Liter
Milliliter
Cubic centimeter (cc)
Cubic meter
Cubic inch
Cubic foot
Mass
Ton
Ton (UK)
Ton (US)
Pound
Ounce
Kilogram
Gram
Temperature
Celsius
Fahrenheit
Kelvin
I would like to work on this issue as a part of GSSOC 2021 @sairish2001 please assign it to me
Feel Free to ask your doubts, if any
| gharchive/issue | 2021-03-04T08:26:24 | 2025-04-01T06:45:42.802947 | {
"authors": [
"sairish2001",
"satyampgt4"
],
"repo": "sairish2001/makesmatheasy.github.io",
"url": "https://github.com/sairish2001/makesmatheasy.github.io/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
842555831 | Adding Euler representation of complex number.
I would like to add Euler representation of complex number in complex number calculator.
Please assign this to me.
Go ahead
Deadline 1 April,2021 (11:59 PM IST)
| gharchive/issue | 2021-03-27T16:52:27 | 2025-04-01T06:45:42.804115 | {
"authors": [
"kanishka8276",
"sairish2001"
],
"repo": "sairish2001/makesmatheasy",
"url": "https://github.com/sairish2001/makesmatheasy/issues/565",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
151551553 | 怎么取消WheelView和ScrollView滑动冲突?
怎么解决WheelView和ScrollView滑动冲突?
重写Scrollview的分发机制
| gharchive/issue | 2016-04-28T05:18:30 | 2025-04-01T06:45:42.804908 | {
"authors": [
"jaychou2012",
"saiwu-bigkoo"
],
"repo": "saiwu-bigkoo/Android-PickerView",
"url": "https://github.com/saiwu-bigkoo/Android-PickerView/issues/82",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
54856596 | KNL-1304 Add a property to control refresh authzgroup call.
Sorry, I've tried to change my branch and squash into one commit but something was wrong (I have to improve my git skills :cry:), so I had to create a new branch/pull request.
The problem was that the original code has spaces, something like:
<tab><tab><tab><spaces>// KNL-1183
And my code only has tabs:
<tab><tab><tab><tab>// KNL-1304
But not for the first line that was using spaces, sorry :cry:
Now the new code is using tabs only :+1: :tada:
Wow, we need this like water. We have sites with more than 2K users and we use to refresh at night. If we can't disable the autorefresh our users will complain a lot. :+1:
I don't think this approach is ideal, turning off refresh on AuthzGroups would have an adverse effect on many things, for example a system that has Course Management enabled would not see updates to users in sites even though CM has accurate information on enrollments.
However this method can cause an extreme amount of load which is also problematic.
We have another approach that addresses both of these issues, see
https://jira.sakaiproject.org/browse/KNL-1325
Hi, In our case we need this feature in the code. We have two kinds of provided sites. Some with a regular CM implementation and others that have another approximation. The realms provided by CM would be fine with KNL- 1325 but the others not. Those realms must be never refreshed. People when access to sakai are automatically joined to the realm but if you refresh the realm you empty the list. Is something like the sample provider implementation. Those providers uses our user types and other ldap attributes to determine if they must be in a site or not, but we don't extract the inverse query (who must be in this site). We use them to get big sites like intranet sites, faculty, etc.
So, we really need that refreshAuth don't be automatically launched, just when admin wants.
| gharchive/pull-request | 2015-01-20T08:40:57 | 2025-04-01T06:45:42.810664 | {
"authors": [
"ern",
"juanjmerono",
"sinmsinm"
],
"repo": "sakaiproject/sakai",
"url": "https://github.com/sakaiproject/sakai/pull/56",
"license": "ECL-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
277819286 | Class not found Exception while running HBASE Mini Cluster (org.apache.hadoop.hbase.backup.impl.BackupException)
I have an existing project running storm that writes into HBASE. As a result i use HBASE client libs. When i run this i get a class not found exception.
I am pretty sure it is a class conflict with the HBASE client libraries , as i can get this to run clean on a project that has no HBASE client dependencies.
I use an older version of HBASE libs
org.apache.hbase
hbase-client
1.1.2
and i tried multiple versions of the mini cluster to see if it works (0.1.6 to 0.1.14). Would anyone know which versions work well together ? or rather for 1.1.2 is there a compatible version
Error.log
Sorry for the delay here. The hbase version leveraged in the mini cluster project is 1.1.2. Can you post your pom.xml so I can try to find a fix?
Hello Shane , attached is the POM. I have tried combinations including the hbase-master and backup jars in the pom in various scopes. I think the issue is with the hbase-client jars and phoenix-jars that i already use
pom.zip
How did you fix this?
delete the dependency of hbase-client in your pom
hope this will help you
| gharchive/issue | 2017-11-29T16:27:14 | 2025-04-01T06:45:42.815275 | {
"authors": [
"batramanuj",
"cxiiiiiii",
"sakserv",
"the-ram"
],
"repo": "sakserv/hadoop-mini-clusters",
"url": "https://github.com/sakserv/hadoop-mini-clusters/issues/52",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1505976916 | Fix typo in intro.rst
accomodate -> accommodate
signed CLA.
| gharchive/pull-request | 2022-12-21T09:22:12 | 2025-04-01T06:45:42.858417 | {
"authors": [
"eltociear"
],
"repo": "salesforce/LAVIS",
"url": "https://github.com/salesforce/LAVIS/pull/70",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
307464278 | Add support to defaultValue on <Input> component.
We are using DSR on trailhead and we are having issues using the <Input> component because it does not accept defaultValue currently you can pass value but doing that will mean much more complexity than required.
Please can you guys add support for defaultValue? or allow passing inputProps?
@interactivellama ^
I would like to keep DSR components "controlled" as much as we can. See Codebase Overview. It is the recommendation from the React team, too. Uncontrolled components that use the DOM to manage state are an anti-pattern. The other thing to worry about is switching from controlled to uncontrolled inputs.
Is this mostly an issue with having a non-redux/flux app where you want the input uncontrolled and are reading the DOM for state via inputRef on form submit? Is that an accurate assessment?
That said, it might be a good idea to allow folks transitioning from other frameworks to allow the DOM to manage state and just add comments and URLs on the prop that say its use is not recommended.
@interactivellama thanks for your quick feedback, I agree that managing the form with state is better than using refs, but both ways required a similar amount of boilerplate code to be added to the forms.
We actually don't use neither to on our forms coz we built a Form component that automatically serializes the data so we need the defaultValue only to set the initial value of the forms.
Bellow its an example of our auto serialized Form:
export default class Form extends React.PureComponent {
static propTypes = {
action: PropTypes.string.isRequired,
method: PropTypes.string.isRequired,
onSubmit: PropTypes.func,
className: PropTypes.string,
children: PropTypes.node
};
onSubmit = (e) => {
const {onSubmit} = this.props;
if (!onSubmit) return;
e.preventDefault();
// serialize function builds the form params based on its input names
// just like a regular HTML form would.
const data = serialize(this.form, {hash: true, empty: true});
onSubmit(data);
}
render() {
const {props: {children, ...otherProps}} = this;
return (
<form
{...otherProps}
ref={(form) => { this.form = form; }}
acceptCharset="UTF-8"
onSubmit={this.onSubmit}>
<input name="utf8" type="hidden" value="✓" />
{children}
</form>
);
}
}
The component above can be use on our forms what would need any custom logic to deal with its inputs.
function AboutMeForm({name, email, onSave}) {
return (
<Form action="/something" method="post" onSubmit={(data) => onSave(data)}>
<input name="user[name]" defaultValue={name} />
<input name="user[email]" defaultValue={email} />
...
</Form>
);
}
That said, there are multiple ways to make a cake, in some cases we controlled components are the best option, but that is not always true we don't want the complexity that of managing input state and building the serialized data manually on every single form.
I should have time to review a pull request for defaultValue next Tuesday or Friday--after that it will be a two weeks away. I think a Jest DOM snapshot would the best test, but the input doesn't have those set up right now, so Mocha is fine.
Something like This sets the initial value of the inputon an uncontrolledinput. Use sparingly. This prop is only present to support uncontrolled components in applications that do not maintain a flux- like state.
@interactivellama sounds good I will be glad to send a PR ;)
| gharchive/issue | 2018-03-22T00:27:30 | 2025-04-01T06:45:42.865339 | {
"authors": [
"3den",
"interactivellama"
],
"repo": "salesforce/design-system-react",
"url": "https://github.com/salesforce/design-system-react/issues/1305",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
2395877796 | Configurations of the models
Hello,
Thank you so much for this work! I was wondering if the authors could share what hyperparameters were used to evaluate and generate the summaries. Did you simply use the default OpenAI Chat completion hyperparameters (temperature = 1, top_p = 1) or were other parameters used?
Hey @PastelBelem8, good question.
We used default parameters for all models (temperature = 1, top_p = 1).
It would be interesting to see what effect the generation parameters would have on the results, but we have not explored this yet!
Happy to hear your thoughts,
Philippe
| gharchive/issue | 2024-07-08T14:58:00 | 2025-04-01T06:45:42.867485 | {
"authors": [
"PastelBelem8",
"tingofurro"
],
"repo": "salesforce/summary-of-a-haystack",
"url": "https://github.com/salesforce/summary-of-a-haystack/issues/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2268369252 | 🛑 RAÄ Kulturarvsdata Runor is down
In e3fadf8, RAÄ Kulturarvsdata Runor (https://kulturarvsdata.se/uu/srdb/fab985d7-61cc-4014-9d85-0a5011d40dc8) was down:
HTTP code: 0
Response time: 0 ms
Resolved: RAÄ Kulturarvsdata Runor is back up in 68f22ec after 54 minutes.
| gharchive/issue | 2024-04-29T07:57:14 | 2025-04-01T06:45:42.870633 | {
"authors": [
"salgo60"
],
"repo": "salgo60/DIGG-skuggbacklog",
"url": "https://github.com/salgo60/DIGG-skuggbacklog/issues/771",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
3524412 | feature request: set node temporarily offline/online
As is possible with jenkins-cli.jar, it would be great if jenkinsapi could mark a node offline and online.
The use case is part of an upgrade automation script for several jenkins test/build slaves.
hey,
I know and I had planned on implementing that. Switching jobs got sort of in the way, I was hired by spotify and need to focus on relocation for the short term. I'll see if I can cough something up in the next week.
Would you be willing to test since I don't have access to the testing resources I used to have at my previous job and it will take some time to set them up around here.
Sure. I would test that. Let me know.
| gharchive/issue | 2012-03-06T13:54:30 | 2025-04-01T06:45:42.877818 | {
"authors": [
"LarsNordin",
"ramonvanalteren"
],
"repo": "salimfadhley/jenkinsapi",
"url": "https://github.com/salimfadhley/jenkinsapi/issues/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2034468242 | please provide binary wheels for python 3.12
According to https://pypi.org/project/QCustomPlot-PyQt6/#files the current release of QCustomPlot-PyQt6 (2.1.1.1) is distributed as source package and in binary packages for (c)python 3.8-->3.11.
could you provide also binary packages for python 3.12?
I've uploaded the wheels. Please test if all works as expected.
Hi
thanks for the build.
(It may be a problem in my workstation, I'm on a debian trixie/sid ... but)
I managed to install with pip without issues, but if I launch a project using the library I get
ImportError: /lib/x86_64-linux-gnu/libm.so.6: version `GLIBC_2.38' not found (required by /home/zed/eclipse-workspace/venv/lib/python3.12/site-packages/QCustomPlot_PyQt6.cpython-312-x86_64-linux-gnu.so)
oh, and also the qt version: it seems your package requires Qt_6.6
but I have
# apt-cache show libqt6core6
Package: libqt6core6
Source: qt6-base
Version: 6.4.2+dfsg-19
[...]
So, I've build using Manjaro, and there are newer versions of these libraries. Maybe I can use a virtual machine to build wheels that are more compatible with other distributions.
For now you can try to build it yourself. I guess you need the following packages (or similar): qtbase6-dev, libpython3-dev, pyqt6-dev, python3-pyqt6, python3-pyqtbuild, sip-tools. Then you should be able to execute sip-build --qmake _path_to_qmake6_ and sip-install --qmake _path_to_qmake6_.
Or you can download deb packages here: https://download.opensuse.org/repositories/home:/sergeyopensuse:/gpxviewer/Debian_Testing.
I've built packages in Debian 12 and uploaded to PyPI as version 2.1.1.1.1.1. Please test if it works now.
Hi.
Sorry for the late reply, and thanks for the build.
The latest version works flawlessly, thanks!
| gharchive/issue | 2023-12-10T17:48:27 | 2025-04-01T06:45:42.890332 | {
"authors": [
"ZeeD",
"salsergey"
],
"repo": "salsergey/QCustomPlot-PyQt",
"url": "https://github.com/salsergey/QCustomPlot-PyQt/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1041800522 | aws-sdk module download不了。能否支持goproxy
是否在你的action中支持下Goproxy.cn 完全实现了GOPROXY 协议
这是偶发的吧,重试下呢? gha是运行在国外,不需要goproxy啊
因为我挂了一个self hosted节点,在国内跑
Ambor @.***> 于 2021年11月2日周二 下午2:41写道:
这是偶发的吧,重试下呢? gha是运行在国外,不需要goproxy啊
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/saltbo/uptoc/issues/31#issuecomment-957141041, or
unsubscribe
https://github.com/notifications/unsubscribe-auth/AAAJGLXXFTSRLNUM5BAJ56DUJ6B2BANCNFSM5HFJ2H5A
.
我去 这玩意儿还能自建节点啊
那得支持下设置环境变量了
你玩玩先。时代不同了。
| gharchive/issue | 2021-11-02T01:38:32 | 2025-04-01T06:45:42.901980 | {
"authors": [
"saltbo",
"xiaods"
],
"repo": "saltbo/uptoc",
"url": "https://github.com/saltbo/uptoc/issues/31",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1866371377 | plasma-infra: Enable full bootstrap [Publish canary]
Release Notes
Выключили флаг --since для workflow "Publish Canary"
What/why Changed
Столкнулись с тем что во время публикации версии ловим ошибку.
Если коротко то
Делаем npx lerna bootstrap --since=$(git merge-base --fork-point origin/dev)
1.1 Собрали только то что изменили
Публикация пытается собрать все и падает с ошибкой
Если правильно понял, это issue
Похоже на наш кейс.
📦 Published PR as canary version: Canary Versions
:sparkles: Test out this PR locally via:
npm install @salutejs/plasma-b2c@1.221.1-canary.673.5972477545.0
npm install @salutejs/plasma-colors@0.8.1-canary.673.5972477545.0
npm install @salutejs/plasma-core@1.126.1-canary.673.5972477545.0
npm install @salutejs/plasma-hope@1.221.1-canary.673.5972477545.0
npm install @salutejs/plasma-icons@1.155.1-canary.673.5972477545.0
npm install @salutejs/cra-template-plasma-shop-template@2.3.1-canary.673.5972477545.0
npm install @salutejs/plasma-temple@1.170.1-canary.673.5972477545.0
npm install @salutejs/plasma-tokens-b2b@1.25.1-canary.673.5972477545.0
npm install @salutejs/plasma-tokens-b2c@0.34.1-canary.673.5972477545.0
npm install @salutejs/plasma-tokens-core@0.3.1-canary.673.5972477545.0
npm install @salutejs/plasma-tokens-web@1.40.1-canary.673.5972477545.0
npm install @salutejs/plasma-tokens@1.59.1-canary.673.5972477545.0
npm install @salutejs/plasma-typo@0.36.1-canary.673.5972477545.0
npm install @salutejs/plasma-ui@1.202.1-canary.673.5972477545.0
npm install @salutejs/plasma-web@1.221.1-canary.673.5972477545.0
npm install @salutejs/plasma-cy-utils@0.66.1-canary.673.5972477545.0
npm install @salutejs/plasma-sb-utils@0.124.1-canary.673.5972477545.0
npm install @salutejs/plasma-tokens-utils@0.30.1-canary.673.5972477545.0
# or
yarn add @salutejs/plasma-b2c@1.221.1-canary.673.5972477545.0
yarn add @salutejs/plasma-colors@0.8.1-canary.673.5972477545.0
yarn add @salutejs/plasma-core@1.126.1-canary.673.5972477545.0
yarn add @salutejs/plasma-hope@1.221.1-canary.673.5972477545.0
yarn add @salutejs/plasma-icons@1.155.1-canary.673.5972477545.0
yarn add @salutejs/cra-template-plasma-shop-template@2.3.1-canary.673.5972477545.0
yarn add @salutejs/plasma-temple@1.170.1-canary.673.5972477545.0
yarn add @salutejs/plasma-tokens-b2b@1.25.1-canary.673.5972477545.0
yarn add @salutejs/plasma-tokens-b2c@0.34.1-canary.673.5972477545.0
yarn add @salutejs/plasma-tokens-core@0.3.1-canary.673.5972477545.0
yarn add @salutejs/plasma-tokens-web@1.40.1-canary.673.5972477545.0
yarn add @salutejs/plasma-tokens@1.59.1-canary.673.5972477545.0
yarn add @salutejs/plasma-typo@0.36.1-canary.673.5972477545.0
yarn add @salutejs/plasma-ui@1.202.1-canary.673.5972477545.0
yarn add @salutejs/plasma-web@1.221.1-canary.673.5972477545.0
yarn add @salutejs/plasma-cy-utils@0.66.1-canary.673.5972477545.0
yarn add @salutejs/plasma-sb-utils@0.124.1-canary.673.5972477545.0
yarn add @salutejs/plasma-tokens-utils@0.30.1-canary.673.5972477545.0
Theme Builder app deployed!
http://plasma.sberdevices.ru/pr/plasma-theme-builder-pr-673/
Theme Builder app deployed!
http://plasma.sberdevices.ru/pr/plasma-theme-builder-pr-673/
🚀 This PR is included in version: @salutejs/cra-template-plasma-shop-template@2.4.0-dev.0, @salutejs/plasma-b2c@1.222.0-dev.0, @salutejs/plasma-colors@0.9.0-dev.0, @salutejs/plasma-core@1.127.0-dev.0, @salutejs/plasma-cy-utils@0.67.0-dev.0, @salutejs/plasma-hope@1.222.0-dev.0, @salutejs/plasma-icons@1.156.0-dev.0, @salutejs/plasma-sb-utils@0.125.0-dev.0, @salutejs/plasma-temple@1.171.0-dev.0, @salutejs/plasma-tokens-b2b@1.26.0-dev.0, @salutejs/plasma-tokens-b2c@0.35.0-dev.0, @salutejs/plasma-tokens-core@0.4.0-dev.0, @salutejs/plasma-tokens-utils@0.31.0-dev.0, @salutejs/plasma-tokens-web@1.41.0-dev.0, @salutejs/plasma-tokens@1.60.0-dev.0, @salutejs/plasma-typo@0.37.0-dev.0, @salutejs/plasma-ui@1.203.0-dev.0, @salutejs/plasma-web@1.222.0-dev.0 🚀
🚀 This PR is included in version: @salutejs/cra-template-plasma-shop-template@2.4.0-dev.0, @salutejs/plasma-b2c@1.222.0-dev.0, @salutejs/plasma-colors@0.9.0-dev.0, @salutejs/plasma-core@1.127.0-dev.0, @salutejs/plasma-cy-utils@0.67.0-dev.0, @salutejs/plasma-hope@1.222.0-dev.0, @salutejs/plasma-icons@1.156.0-dev.0, @salutejs/plasma-sb-utils@0.125.0-dev.0, @salutejs/plasma-temple@1.171.0-dev.0, @salutejs/plasma-tokens-b2b@1.26.0-dev.0, @salutejs/plasma-tokens-b2c@0.35.0-dev.0, @salutejs/plasma-tokens-core@0.4.0-dev.0, @salutejs/plasma-tokens-utils@0.31.0-dev.0, @salutejs/plasma-tokens-web@1.41.0-dev.0, @salutejs/plasma-tokens@1.60.0-dev.0, @salutejs/plasma-typo@0.37.0-dev.0, @salutejs/plasma-ui@1.203.0-dev.0, @salutejs/plasma-web@1.222.0-dev.0 🚀
| gharchive/pull-request | 2023-08-25T06:03:30 | 2025-04-01T06:45:43.096588 | {
"authors": [
"Salute-Eva",
"Yakutoc"
],
"repo": "salute-developers/plasma",
"url": "https://github.com/salute-developers/plasma/pull/673",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2360587357 | HTTP/3 realization doesn't work
Describe the bug
Cannot connect to QuinnListener from Firefox/Chromium/curl with an error message like this:
ERROR salvo_core::server: accept connection failed error=Custom { kind: Other, error: "the cryptographic handshake failed: error 120: peer doesn't support any known protocol" }
To Reproduce
Steps to reproduce the behavior:
Clone the repository
Cd into examples/
Execute cargo run --bin example-hello-h3
Run in another console curl --http3-only --tlsv1.3 --cacert certs/cert.pem https://127.0.0.1:5800 and see curl: (7) error:8000006F:system library::Connection refused
In example-hello-h3 console also see an error
Expected behavior
After 4th step see Hello World
Desktop (please complete the following information):
OS: Arch Linux (kernel 6.9.3-arch1-1)
curl: 8.8.0 (x86_64-pc-linux-gnu) libcurl/8.8.0 OpenSSL/3.3.1 zlib/1.3.1 brotli/1.1.0 zstd/1.5.6 libidn2/2.3.7 libpsl/0.21.5 libssh2/1.11.0 nghttp2/1.62.1 nghttp3/1.3.0, Release-Date: 2024-05-22
Note that this error was found in 0.68; in 0.66 QuinnListener works as expected!
| gharchive/issue | 2024-06-18T19:48:43 | 2025-04-01T06:45:43.103663 | {
"authors": [
"markcda"
],
"repo": "salvo-rs/salvo",
"url": "https://github.com/salvo-rs/salvo/issues/814",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2556141952 | Stop all executions before compile and run
Maybe a "stop all" button is also necessary?
I'll add a shortcut for it, but I don't think I'll add a button, because I think the default UI should stay the same right now to reach maximum parity with the Sublime Text version (I might allow users to customize the UI, although that'll take much more planning)
Note to self: this also causes process leak
To clarify: if a test case is running and you're re-running it (compile & run all), it will stop the test case and proceed normally. So running one test case does not affect the status of other test cases.
I know this is not exactly what you asked for (at least to my interpretation), but due to the design of this extension, enforcing this 100% would effectively mean you can only have one test case running at any given moment, and that currently goes against the strength of being able to interchangeably add/edit/run multiple test cases at once. If the concern is avoiding running any processes against an "older" executable, then the situation can easily be avoided with running all the test cases.
| gharchive/issue | 2024-09-30T09:53:07 | 2025-04-01T06:45:43.110044 | {
"authors": [
"YuiHrsw",
"sam20908"
],
"repo": "sam20908/vscode-fastolympiccoding",
"url": "https://github.com/sam20908/vscode-fastolympiccoding/issues/25",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
113001302 | Added a simple Ruby example | #19
My Ruby skills are humble, but parsing a JSON file is far from complicated.
I cannot reference issue #19 by changing the title :(
Done :+1:
| gharchive/pull-request | 2015-10-23T11:22:07 | 2025-04-01T06:45:43.112491 | {
"authors": [
"Ava7",
"samayo"
],
"repo": "samayo/country-data",
"url": "https://github.com/samayo/country-data/pull/49",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
92753385 | how i run it
hi dear
i download sitemap and update with composer update command line
i am amator and need help you for run this script
for example i create index.php and set this code
////////////////
addItem(new SitemapItem(
'http://rmcreative.ru/', // URL
time(), // last modifcation timestamp
Item::DAILY, // update frequency
0.7 // priority
));
// add more pages
foreach ($pages as $page){
$sitemap->addItem(new SitemapItem(
'http://rmcreative.ru/' . $page->url,
$page->updatedOn,
Item::MONTHLY
));
}
// generate sitemap.xml
$sitemap->writeToFile('sitemap.xml');
// or get it as string
$sitemapString = $sitemap->render();
////////////////////////////////////////////
after run have more error
Fatal error: Class 'samdark\sitemap' not found in C:\xampp\htdocs\samdark\index.php on line 9
how i run it????
update ////////
addItem(new SitemapItem(
'http://rmcreative.ru/', // URL
time(), // last modifcation timestamp
Item::DAILY, // update frequency
0.7 // priority
));
// add more pages
foreach ($pages as $page){
$sitemap->addItem(new SitemapItem(
'http://rmcreative.ru/' . $page->url,
$page->updatedOn,
Item::MONTHLY
));
}
// generate sitemap.xml
$sitemap->writeToFile('sitemap.xml');
// or get it as string
$sitemapString = $sitemap->render();
//////////
error
Fatal error: Class 'samdark\sitemap' not found in C:\xampp\htdocs\samdark\index.php on line 9
i now correct it
addItem(new Item(
'http://rmcreative.ru/' . 'sdfsf.htm',
'201',
Item::MONTHLY
));
// generate sitemap.xml
$sitemap->writeToFile('sitemap.xml');
https://github.com/samdark/sitemap/pull/2
| gharchive/issue | 2015-07-02T23:01:55 | 2025-04-01T06:45:43.123205 | {
"authors": [
"lynicidn",
"vahidalvandi"
],
"repo": "samdark/sitemap",
"url": "https://github.com/samdark/sitemap/issues/1",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
112459699 | which: no skype in (/usr/sbin:/usr/bin:/sbin:/bin)
Skype is not launched here.
Also, issue with usermod (adding skype user to group 91, which does not exists in container machine).
This is a laptop with Arch and plain docker installed.
% docker version
Client:
Version: 1.8.3
API version: 1.20
Go version: go1.5.1
Git commit: f4bf5c7-dirty
Built: Wed Oct 14 11:17:02 UTC 2015
OS/Arch: linux/amd64
Server:
Version: 1.8.3
API version: 1.20
Go version: go1.5.1
Git commit: f4bf5c7-dirty
Built: Wed Oct 14 11:17:02 UTC 2015
OS/Arch: linux/amd64
% docker info
Containers: 1
Images: 12
Storage Driver: btrfs
Build Version: Btrfs v4.2.1
Library Version: 101
Execution Driver: native-0.2
Logging Driver: json-file
Kernel Version: 4.2.3-1-ARCH
Operating System: Arch Linux (containerized)
CPUs: 4
Total Memory: 7.696 GiB
Name: suns
ID: AZ47:NNAJ:YSV3:QE2R:CNHP:BSIQ:WNA4:BWXB:YL3X:3BTJ:WJBG:RPC2
(I have removed -d and /dev/null redirect from "docker run" command)
% skype
which: no skype in (/usr/sbin:/usr/bin:/sbin:/bin)
Cleaning up stopped skype instances...
cat: /etc/timezone: No such file or directory
Starting skype...
usermod: group '91' does not exist
it appears no group with GID 91 exists in the container, so the usermod fails. To fix this we will need to create a new group and assign it the GID 91 before doing usermod. Will make this changes soon,
Fix did not work very well:
% skype
which: no skype in (/usr/sbin:/usr/bin:/sbin:/bin)
Cleaning up stopped skype instances...
cat: /etc/timezone: No such file or directory
Starting skype...
groupadd: group 'video' already exists
And skype is still not started (container exits). Do you think it's due to wrong /dev/video0 permissions? Because skype not launching was my main problem.
hmm.. its because a group with the name video already exists. will change the name
please try after a while once the build is ready. let me know
Ok, works properly now.
| gharchive/issue | 2015-10-20T21:01:04 | 2025-04-01T06:45:43.132936 | {
"authors": [
"mikken",
"sameersbn"
],
"repo": "sameersbn/docker-skype",
"url": "https://github.com/sameersbn/docker-skype/issues/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
497355617 | RuntimeError: Assertion `cur_target >= 0 && cur_target < n_classes' failed.
I am trying to predict clothing attributes but getting an error like below. I am very new to using torch. Can anyone tell what is causing the problem.
{'neck': 7, 'sleeve_length': 5, 'pattern': 11}
Processing Attribute: neck
Start Training for: neck
/anaconda3/envs/nlp/lib/python3.6/site-packages/torchvision/transforms/transforms.py:211: UserWarning: The use of the transforms.Scale transform is deprecated, please use transforms.Resize instead.
"please use transforms.Resize instead.")
Before pre-processing... (2238, 4)
After pre-processing... (1776, 3)
Found validation dataset.... preparing
Before pre-processing... (2238, 4)
After pre-processing... (446, 3)
LR is set to 0.01
/Users/akshyasingh/PycharmProjects/predict_deep_fashion_attributes/classifier.py:65: UserWarning: Implicit dimension choice for log_softmax has been deprecated. Change the call to include dim=X as an argument.
classes_softmax = F.log_softmax( average_pool_flatten )
Traceback (most recent call last):
File "multi_label_vgg_nn.py", line 27, in
use_gpu=False )
File "/Users/akshyasingh/PycharmProjects/predict_deep_fashion_attributes/classifier.py", line 247, in create_attributes_model
batch_size, num_workers, num_epochs, use_gpu=use_gpu)
File "/Users/akshyasingh/PycharmProjects/predict_deep_fashion_attributes/classifier.py", line 206, in train_model
flatten_pretrained_out=flatten_pretrained_out)
File "/Users/akshyasingh/PycharmProjects/predict_deep_fashion_attributes/classifier.py", line 144, in train_attribute_model
loss = criterion(outputs, labels)
File "/anaconda3/envs/nlp/lib/python3.6/site-packages/torch/nn/modules/module.py", line 547, in call
result = self.forward(*input, **kwargs)
File "/anaconda3/envs/nlp/lib/python3.6/site-packages/torch/nn/modules/loss.py", line 916, in forward
ignore_index=self.ignore_index, reduction=self.reduction)
File "/anaconda3/envs/nlp/lib/python3.6/site-packages/torch/nn/functional.py", line 1995, in cross_entropy
return nll_loss(log_softmax(input, 1), target, weight, None, ignore_index, None, reduction)
File "/anaconda3/envs/nlp/lib/python3.6/site-packages/torch/nn/functional.py", line 1824, in nll_loss
ret = torch._C._nn.nll_loss(input, target, weight, _Reduction.get_enum(reduction), ignore_index)
RuntimeError: Assertion `cur_target >= 0 && cur_target < n_classes' failed. at /Users/distiller/project/conda/conda-bld/pytorch_1565272679438/work/aten/src/THNN/generic/ClassNLLCriterion.c:94
The issue is resolved. No of classes to predict, I changed the input to load_model( col_dim + 1). Can anyone tell me if this is the correct way. The labels json i have outputs : {'neck': 7, 'sleeve_length': 5, 'pattern': 11}
| gharchive/issue | 2019-09-23T22:16:32 | 2025-04-01T06:45:43.158172 | {
"authors": [
"akshay772"
],
"repo": "sampathweb/apparel-styles",
"url": "https://github.com/sampathweb/apparel-styles/issues/7",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
781069689 | Move scrubber position programmatically
I want to move scrubber position programmatically just like clicking on scrubber and move scrubber left or right of the seek bar. I tried with player.forward(seekTime) but unfortunately its not working as expected. Is there any way to move scrubber position?
Thanks in advance for your help.
Hi,
It seems to be working fine for me on https://plyr.io. Do you have any more info to reproduce this?
Hi,
It seems to be working fine for me on https://plyr.io. Do you have any more info to reproduce this?
| gharchive/issue | 2021-01-07T06:34:12 | 2025-04-01T06:45:43.160712 | {
"authors": [
"abhijitpd",
"sampotts"
],
"repo": "sampotts/plyr",
"url": "https://github.com/sampotts/plyr/issues/2062",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1387713529 | Pass status flag to the api call
Features in this PR:
Simplifies API call to get workflows, to return more relevant results. I added the status flag to the API call so that the workflow runs returned already meet the status criteria required. This way the runner doesn't need to do the extra work.
Added foundation to track unique dates
Added color variable to show colors for each Deployment frequency status
Outputs job summary. A sample is below:
DORA Metric: Deployment Frequency
This is how often you deploy successfully to production.
Results
Workflow(s): 203496, 108084
Deployment Frequency: 1.33 per day
In the last 30 days, you deployed to production on 30 days
This is Elite deployment frequency.
DORA Classification: Elite
Recommend simplifying the output to be:
# DORA Metric: Deployment Frequency
**Defintion:** For the primary application or service, how often is it successfully deployed to production.
**Results:** Deployment frequency over last 30 days, is **1.2 per day**, with a DORA rating of **'Elite'**. Workflow(s) used: CI, CD

| gharchive/pull-request | 2022-09-27T12:54:14 | 2025-04-01T06:45:43.188114 | {
"authors": [
"ayodejiayodele",
"samsmithnz"
],
"repo": "samsmithnz/deployment-frequency",
"url": "https://github.com/samsmithnz/deployment-frequency/pull/21",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
262267856 | Multi-version Jenkins testing
This adds up/down testing for all supported kubernetes versions.
not sure how I feel about running all versions with every commit. maybe something to only do on master?
Yeah, I hate it. In a perfect world, maybe only run it after review approval but before merge to master. So far, this is mostly a PoC.
I don't hate it, this is what I was thinking of doing as part of another PR actually. not sure what the 'occasional' key to run these would be though.
My preference would be that all prior versions pass before something gets merged to master, but if it's too heavy to run on every PR, perhaps we could have a daily staging branch that can be auto-merged to master by a passing nightly.
or maybe nightly master -> release branch? Seems less disruptive to current developer workflows that expect master to have all merged PRs.
this leaked a 1.5 cluster last night. I'm cleaning it up now but there seems to be fragility there.
This is failing because each kubectl config set command creates $HOME/.kube/config.lock. I'm having a lot of difficulty in finding the code that does this and I've had no response from sig-cli on Slack.
@kubernetes/sig-cli-test-failures
This will end up being done as a scheduled job in GitLab. Closing.
| gharchive/pull-request | 2017-10-03T00:17:29 | 2025-04-01T06:45:43.193928 | {
"authors": [
"coffeepac",
"joejulian"
],
"repo": "samsung-cnct/kraken-lib",
"url": "https://github.com/samsung-cnct/kraken-lib/pull/865",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
98909977 | Uncaught TypeError: Cannot read property 'on' of null
[Enter steps to reproduce below:]
Open folder as project, or remove project folder
Atom Version: 1.0.3
System: Mac OS X 10.10.3
Thrown From: git-diff-details package, v0.19.0
Stack Trace
Uncaught TypeError: Cannot read property 'on' of null
At /Applications/Atom.app/Contents/Resources/app.asar/src/git-repository.js:133
TypeError: Cannot read property 'on' of null
at GitRepository.module.exports.GitRepository.onDidChangeStatuses (/Applications/Atom.app/Contents/Resources/app.asar/src/git-repository.js:133:26)
at AtomGitDiffDetailsView.module.exports.Housekeeping.subscribeToRepository (/Users/i-dranichnikov/.atom/packages/git-diff-details/lib/housekeeping.coffee:107:41)
at /Users/i-dranichnikov/.atom/packages/git-diff-details/lib/housekeeping.coffee:42:26
at Emitter.module.exports.Emitter.emit (/Applications/Atom.app/Contents/Resources/app.asar/node_modules/event-kit/lib/emitter.js:82:11)
at Project.module.exports.Project.addPath (/Applications/Atom.app/Contents/Resources/app.asar/src/project.js:274:29)
at /Applications/Atom.app/Contents/Resources/app.asar/src/atom.js:908:41
at EventEmitter.<anonymous> (/Applications/Atom.app/Contents/Resources/app.asar/src/atom.js:444:16)
at emitOne (events.js:77:13)
at EventEmitter.emit (events.js:166:7)
Commands
-4:04.1.0 editor:select-to-beginning-of-word (atom-text-editor.editor.is-focused)
-3:59.7.0 linter:set-bubble-transparent (atom-text-editor.editor.is-focused)
-3:59.4.0 pane:show-next-item (atom-text-editor.editor.is-focused)
-3:55.5.0 linter:set-bubble-transparent (atom-text-editor.editor.is-focused)
-3:55.4.0 pane:show-previous-item (atom-text-editor.editor.is-focused)
-3:54.8.0 linter:set-bubble-transparent (atom-text-editor.editor.is-focused)
-3:54.2.0 pane:show-next-item (atom-text-editor.editor.is-focused)
-3:50.7.0 pane:show-previous-item (atom-text-editor.editor.is-focused)
-3:49.6.0 core:move-left (atom-text-editor.editor.is-focused)
-3:49.4.0 linter:set-bubble-transparent (atom-text-editor.editor.is-focused)
3x -3:49.3.0 editor:select-to-end-of-word (atom-text-editor.editor.is-focused)
2x -3:46.9.0 core:backspace (atom-text-editor.editor.is-focused.autocomplete-active)
3x -3:45.2.0 core:save (atom-text-editor.editor.is-focused)
-3:22.4.0 pane:show-previous-item (atom-text-editor.editor.is-focused)
-3:19.8.0 core:copy (atom-text-editor.editor.is-focused)
2x -0:12.5.0 application:add-project-folder (ol.tree-view.full-menu.list-tree.has-collapsable-children.focusable-panel)
Config
{
"core": {
"themes": [
"isotope-ui",
"atom-dark-syntax"
],
"ignoredNames": [
".git",
".svn",
".DS_Store"
],
"projectHome": "~/rails_projects",
"disabledPackages": [
"open-on-github",
"autocomplete-css",
"goto",
"jslint",
"csslint"
]
}
}
Installed Packages
# User
angularjs, v0.3.0
autocomplete-ruby, v0.1.0
color-picker, v2.0.11
docs-snippets, v0.8.0
emmet, v2.3.12
git-blame, v0.4.5
git-diff-details, v0.19.0
isotope-ui, v2.5.1
jumpy, v2.0.9
language-haml, v0.21.0
lesscompile, v0.5.0
linter, v1.2.4
linter-ruby, v0.1.6
pain-split, v1.4.0
project-manager, v1.15.11
# Dev
No dev packages
+1
Atom Version: 1.0.4
System: Xubuntu 14.04
Thrown From: git-diff-details package, v0.19.0
@abm0 @bisraelsen this should be fixed with v0.20.0
Please let me know if it works for you!
I'm not getting this error anymore, thanks!
| gharchive/issue | 2015-08-04T07:53:53 | 2025-04-01T06:45:43.216287 | {
"authors": [
"abm0",
"bisraelsen",
"samu"
],
"repo": "samu/git-diff-details",
"url": "https://github.com/samu/git-diff-details/issues/27",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
718609598 | Is it possible to use Pydantic to coerce models to valid JSON models?
Checks
[X] I added a descriptive title to this issue
[X] I have searched (google, github) for similar issues and couldn't find anything
[X] I have read and followed the docs and couldn't find an answer
[X] After submitting this, I commit to one of:
Look through open issues and helped at least one other person
Hit the "watch" button on this repo to receive notifications and I commit to help at least 2 people that ask questions in the future
Implement a Pull Request for a confirmed bug
Question
Output of python -c "import pydantic.utils; print(pydantic.utils.version_info())":
pydantic version: 1.6.1
pydantic compiled: True
install path: /Users/gideon/.pyenv/versions/3.8.2/envs/test_env/lib/python3.8/site-packages/pydantic
python version: 3.8.2 (default, Mar 21 2020, 12:42:02) [Clang 11.0.0 (clang-1100.0.33.17)]
platform: macOS-10.15.6-x86_64-i386-64bit
optional deps. installed: ['typing-extensions']
I was wondering if it's possible to coerce Pydantic models into valid JSONs? The following test doesn't pass but I would like to know if there's any equivalent code that achieves the same goal
from typing import Set
import pytest
from pydantic import BaseModel, Json, validate_arguments
class Person(BaseModel):
name: str
scores: Set[int]
@validate_arguments
def convert(payload: Json):
return payload
def test_coerces_inputs_to_valid_json():
p1 = Person(name="John", scores={1, 3, 4, 3})
converted = convert(p1)
scores = dict(converted)["scores"]
assert sorted(scores) == [1, 3, 4]
Note how the type coercion made scores a List[int] instead of a Set[int] so it could be used as a valid JSON
Why is this useful? I've got a piece of pymongo code that can only save valid JSONs and I want to avoid making explicit type conversions if it can be done using Pydantic
Thanks in advance 🙏
Hello @gidgid
Just change converted = convert(p1) into converted = convert(p1.json()).
Is it what you're looking for ?
🤯
I feel embarrassed asking it
Thanks @PrettyWood you're the best
| gharchive/issue | 2020-10-10T12:26:18 | 2025-04-01T06:45:43.222140 | {
"authors": [
"PrettyWood",
"gidgid"
],
"repo": "samuelcolvin/pydantic",
"url": "https://github.com/samuelcolvin/pydantic/issues/1988",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1197095634 | Test appending to file
Just copying a test here from https://github.com/jupyter-server/jupyter_server/pull/783, where we see it failing on MacOS.
I'm afraid this is not going to work.
MacOS filesystem notifications can be significantly delayed, sometimes by seconds. You can also get an "added" or "modified" when you delete a file
The unit tests in test_rust_notify.py have to do a lot to work around this - e.g. instead of using the tmp_path fixture, we have to have a pre-existing path which we reset after tests - otherwise the creation of the directory from tmp_path is received when the watcher starts.
You can see this by enabling debug and running watchfiles on MacOS - events are significantly confused.
We already have some logic to try and clean events which is for macos, but it can't cover all cases.
https://github.com/samuelcolvin/watchfiles/blob/c35fa7997b5503bd067aad87170b342ebebb20b7/src/lib.rs#L67-L76
Thanks for the feedback, I think I'll just simplify tests in jupyter-server for now. We don't want to test watchfiles again there. If the PR is merged we'll just trust whatever watchfiles is doing.
I agree that's the best way. You could even mock the rust code like we do in many of these tests.
#113 will also help a lot with tests - avoiding tests hanging if no change is detected.
I implemented that partly to help with the uvicorn tests for watchfiles, see https://github.com/encode/uvicorn/pull/1437.
Great, thanks!
| gharchive/pull-request | 2022-04-08T09:45:07 | 2025-04-01T06:45:43.227968 | {
"authors": [
"davidbrochart",
"samuelcolvin"
],
"repo": "samuelcolvin/watchfiles",
"url": "https://github.com/samuelcolvin/watchfiles/pull/118",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1191999664 | React v18: Invalid hook call
When using React 18 there is an „Invalid hook call“ showing up in the console.
https://reactjs.org/warnings/invalid-hook-call-warning.html
If you change the version in your sandbox example you can see it as well.
Hi @svenruegg, thank you for noting this.
Hi @svenruegg! Thank you again for issuing the React v18 update out. Clover IIIF should now work in 18.0 and be backwards compatible through 17.0 and to 16.13. See https://codesandbox.io/s/clover-react-v18-y2whrm?file=/src/App.js
| gharchive/issue | 2022-04-04T15:59:32 | 2025-04-01T06:45:43.232748 | {
"authors": [
"mathewjordan",
"svenruegg"
],
"repo": "samvera-labs/clover-iiif",
"url": "https://github.com/samvera-labs/clover-iiif/issues/69",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
194865333 | Generate Pinned Site Metadata for win8/10 compatibility
See https://msdn.microsoft.com/en-us/library/bg183312(v=vs.85).aspx and https://msdn.microsoft.com/en-us/library/gg491732(v=VS.85).aspx
e.g. using meta tags
<meta name="application-name" content="Channel 9 Audio Player">
<meta name="msapplication-tooltip" content="Channel 9 Podcasts">
<meta name="msapplication-window" content="width=1024;height=768">
<meta name="msapplication-navbutton-color" content="#FF3300">
<meta name="msapplication-starturl" content="/">
e.g. using browserconfig.xml
<?xml version="1.0" encoding="utf-8"?>
<browserconfig>
<msapplication>
<tile>
<square70x70logo src="images/smalltile.png"/>
<square150x150logo src="images/mediumtile.png"/>
<wide310x150logo src="images/widetile.png"/>
<square310x310logo src="images/largetile.png"/>
<TileColor>#009900</TileColor>
</tile>
<notification>
<polling-uri src="notifications/contoso1.xml"/>
<polling-uri2 src="notifications/contoso2.xml"/>
<polling-uri3 src="notifications/contoso3.xml"/>
<frequency>30</frequency>
<cycle>1</cycle>
</notification>
</msapplication>
</browserconfig>
It would be nice to have an option to turn off the generation of these metadata. The app might already have a browserconfig.xml in place and we don't want to override it.
@YoranBrondsema did the initial implementation for this feature on #66. We can set the application name, tile background color and icon. Now we can start adding more and more options for this.
<meta name="msapplication-tooltip" content="Channel 9 Podcasts">
<meta name="msapplication-window" content="width=1024;height=768">
<meta name="msapplication-navbutton-color" content="#FF3300">
<meta name="msapplication-starturl" content="/">
We can fill these metadata using the values from the manifest and allow the user to override.
| gharchive/issue | 2016-12-11T23:29:29 | 2025-04-01T06:45:43.243129 | {
"authors": [
"san650"
],
"repo": "san650/ember-web-app",
"url": "https://github.com/san650/ember-web-app/issues/13",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1679978159 | [Bug]Invalid URL 'None': No schema supplied
This issue occurs when I provide a Youtube link. I'm on Windows 11 (Python 3.10.6) using command python app.py
Traceback (most recent call last):
File "/home/ethan/.local/lib/python3.10/site-packages/gradio/routes.py", line 401, in run_predict
output = await app.get_blocks().process_api(
File "/home/ethan/.local/lib/python3.10/site-packages/gradio/blocks.py", line 1302, in process_api
result = await self.call_function(
File "/home/ethan/.local/lib/python3.10/site-packages/gradio/blocks.py", line 1025, in call_function
prediction = await anyio.to_thread.run_sync(
File "/home/ethan/.local/lib/python3.10/site-packages/anyio/to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "/home/ethan/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "/home/ethan/.local/lib/python3.10/site-packages/anyio/_backends/_asyncio.py", line 867, in run
result = context.run(func, *args)
File "/mnt/c/Users/rosha/whisper-jax/app/app.py", line 185, in transcribe_youtube
text, runtime = tqdm_generate(inputs, task=task, return_timestamps=return_timestamps, progress=progress)
File "/mnt/c/Users/rosha/whisper-jax/app/app.py", line 126, in tqdm_generate
model_outputs.append(forward(batch, task=task, return_timestamps=return_timestamps))
File "/mnt/c/Users/rosha/whisper-jax/app/app.py", line 69, in forward
outputs = chunked_query(
File "/mnt/c/Users/rosha/whisper-jax/app/app.py", line 62, in chunked_query
response = requests.post(API_URL_FROM_FEATURES, json=payload)
File "/home/ethan/.local/lib/python3.10/site-packages/requests/api.py", line 119, in post
return request('post', url, data=data, json=json, **kwargs)
File "/home/ethan/.local/lib/python3.10/site-packages/requests/api.py", line 61, in request
return session.request(method=method, url=url, **kwargs)
File "/home/ethan/.local/lib/python3.10/site-packages/requests/sessions.py", line 528, in request
prep = self.prepare_request(req)
File "/home/ethan/.local/lib/python3.10/site-packages/requests/sessions.py", line 456, in prepare_request
p.prepare(
File "/home/ethan/.local/lib/python3.10/site-packages/requests/models.py", line 316, in prepare
self.prepare_url(url, params)
File "/home/ethan/.local/lib/python3.10/site-packages/requests/models.py", line 390, in prepare_url
raise MissingSchema(error)
requests.exceptions.MissingSchema: Invalid URL 'None': No schema supplied. Perhaps you meant http://None?
Hey @mriamnobody, you need to specify the API_URL for this to work when you launch the app. Please refer to the instructions for creating an endpoint: https://github.com/sanchit-gandhi/whisper-jax#creating-an-endpoint
| gharchive/issue | 2023-04-23T10:55:07 | 2025-04-01T06:45:43.245921 | {
"authors": [
"mriamnobody",
"sanchit-gandhi"
],
"repo": "sanchit-gandhi/whisper-jax",
"url": "https://github.com/sanchit-gandhi/whisper-jax/issues/28",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
472689722 | blePeripheral.begin() not executing.
I'm using an nrf8001 module from adafruit to send pressure readings to a device using ble. I'm having an issue getting the code to execute the blePeripheral.begin(); line. I've included the code and the warnings/errors im receiving when verifying the code. Any help in rectifying this issue is appreciated as I can't work it out.
Code
`#include <BLEPeripheral.h>
#include<SPI.h>
#include <Wire.h>
#define BLE_REQ 10
#define BLE_RDY 2
#define BLE_RST 9
BLEPeripheral blePeripheral = BLEPeripheral(BLE_REQ,BLE_RDY,BLE_RST);
BLEService service = BLEService("CCC0");
BLECharacteristic characteristic = BLEFixedLengthCharacteristic("CCC1", BLERead | BLENotify,12);
unsigned long previousMillis = millis(); //Stores the last time sensor was read
unsigned long interval = 1000; //Interval between sensor readings
void setup()
{
Serial.begin(9600);
Serial.println(service.uuid());
Serial.println(characteristic.uuid());
Serial.println(F("Bluetooth Low Energy Sensor Reader"));
blePeripheral.setLocalName("Pressure Reader");
blePeripheral.setDeviceName("Pressure Reader");
Serial.println(F("Local name and device name set to Pressure Reader"));
blePeripheral.setAdvertisedServiceUuid(service.uuid());
blePeripheral.addAttribute(service);
blePeripheral.addAttribute(characteristic);
Serial.println("1234");
void clearData();
blePeripheral.begin();
Serial.println("blePeripheral.begin() worked");
}
void loop()
{
blePeripheral.poll();
if (millis() - previousMillis >= interval)
{
pollSensors();
previousMillis = millis();
}
}
void pollSensors()
{
union {
float f;
byte bytes[4];
} floatUnion;
union {
float f;
byte bytes[4];
} floatUnion2;
union {
float f;
byte bytes[4];
} floatUnion3;
float voltage1 = analogRead(A0);
float pressure1 = voltage1*5/1023; //Change for correct calibration
floatUnion.f = pressure1;
float voltage2 = analogRead(A0);
float pressure2 = voltage225/1023; //Change for correct calibration
floatUnion2.f = pressure2;
float voltage3 = analogRead(A0);
float pressure3 = voltage335/1023; //Change for correct calibration
floatUnion3.f = pressure3;
char byteArr[12];
int ii = 0;
for(ii= 0 ; ii < 12 ; ii ++ ){
if(ii<4){
//pressure1
byteArr[ii] = floatUnion.bytes[ii];
}else if((ii>=4)&&(ii<8)){
//pressure2
byteArr[ii] = floatUnion2.bytes[ii-4];
}else if((ii>=8)&&(ii<12)){
//pressure3
byteArr[ii] = floatUnion3.bytes[ii-8];
}
}
characteristic.setValue(byteArr);
Serial.print(F("Pressure 1: "));
Serial.println(pressure1);
Serial.println(pressure2);
Serial.println(pressure3);
}`
Warning/Error messages
`C:\Users\joshm\Documents\Arduino\libraries\arduino_958432\src/utility/hal_aci_tl.h:86:3: note: type 'struct aci_pins_t' should match type 'struct aci_pins_t' that itself violate one definition rule
} aci_pins_t;
^
C:\Users\joshm\Documents\Arduino\libraries\arduino_958432\src/utility/hal_aci_tl.h:86:3: note: the incompatible type is defined here
} aci_pins_t;
^
C:\Users\joshm\Documents\Arduino\libraries\arduino_958432\src/utility/lib_aci.h:65:16: note: type 'struct aci_state_t' should match type 'struct aci_state_t' that itself violate one definition rule
typedef struct aci_state_t
^
C:\Users\joshm\Documents\Arduino\libraries\arduino_958432\src/utility/lib_aci.h:65:16: note: the incompatible type is defined here
typedef struct aci_state_t
^`
Here is the same. It looks the developer of this library has no clue !
| gharchive/issue | 2019-07-25T06:33:35 | 2025-04-01T06:45:43.311933 | {
"authors": [
"JMAC4790",
"John-Almardeny"
],
"repo": "sandeepmistry/arduino-BLEPeripheral",
"url": "https://github.com/sandeepmistry/arduino-BLEPeripheral/issues/269",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
196298990 | Unused import in EmailService.php
This import of ConfigurationManager is not used.
https://github.com/sandstorm/UserManagement/blob/master/Classes/Sandstorm/UserManagement/Domain/Service/EmailService.php#L23
resolved.
| gharchive/issue | 2016-12-18T21:00:21 | 2025-04-01T06:45:43.344705 | {
"authors": [
"Akii",
"beheist"
],
"repo": "sandstorm/UserManagement",
"url": "https://github.com/sandstorm/UserManagement/issues/9",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
226697122 | Upgrade dependnecies
This change is
@homu r+
:pushpin: Commit f15d523 has been approved by saneyuki
:hourglass: Testing commit f15d523 with merge 2aeb95b...
:sunny: Test successful - status
| gharchive/pull-request | 2017-05-05T21:30:29 | 2025-04-01T06:45:43.347170 | {
"authors": [
"homu",
"saneyuki"
],
"repo": "saneyuki/option-t.js",
"url": "https://github.com/saneyuki/option-t.js/pull/172",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2346672362 | x1211 Add kit costing to probe operation request
For #409
merged to devel
| gharchive/pull-request | 2024-06-11T15:05:43 | 2025-04-01T06:45:43.361080 | {
"authors": [
"khelwood",
"sabrine33"
],
"repo": "sanger/stan-core",
"url": "https://github.com/sanger/stan-core/pull/410",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
741263182 | Use 'document' type for Mux assets
Consider using 'document' for mux video assets so the types show up in GraphQL. Side-effect is that these documents may show up in auto-generated studio menus.
Note: You can hide Mux Assets from the ‘New document’ menu by providing a custom ‘newDocumentStructure’ and filtering out the mux assets.
Reference: https://www.sanity.io/docs/initial-value-templates#configuring-the-new-document-menu-56b4073ca73a
Sample newDocumentStructure.js:
import S from '@sanity/base/structure-builder'
export default [
...S.defaultInitialValueTemplateItems().filter(i => i.spec.id !== 'mux.videoAsset')
]
This definitely needs testing. Now I’m getting reference errors when trying to remove assets. They can’t be deleted due to existing reference (but it’s the reference to the current doc that I’m trying to remove in the first place). Also, the checkbox labels in the delete dialog are now missing? (not sure how that would be related, investigating...)
Don't know when but I see it's already changed to mux.videoAsset :relaxed:
Don't know when but I see it's already changed to mux.videoAsset :relaxed:
| gharchive/pull-request | 2020-11-12T04:14:34 | 2025-04-01T06:45:43.370715 | {
"authors": [
"stipsan",
"tkdave"
],
"repo": "sanity-io/sanity-plugin-mux-input",
"url": "https://github.com/sanity-io/sanity-plugin-mux-input/pull/20",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2646586206 | 🛑 Komga is down
In d85b8de, Komga ($KOMGA_URL) was down:
HTTP code: 502
Response time: 906 ms
Resolved: Komga is back up in b39d977 after 15 minutes.
| gharchive/issue | 2024-11-09T21:54:11 | 2025-04-01T06:45:43.374558 | {
"authors": [
"sannidhyaroy"
],
"repo": "sannidhyaroy/Uptime-Ryuu",
"url": "https://github.com/sannidhyaroy/Uptime-Ryuu/issues/190",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
249201180 | Failed to resolve : com.android.support.appcompat-v7:26.0.0
My dear
When i add compile project , i have receive this error although i have compile appcompat 26.+ library
Hi,
Did you add google repository in your top most build.gradle file?
allprojects {
repositories {
jcenter()
maven { url "https://maven.google.com" }
}
}
| gharchive/issue | 2017-08-10T00:41:21 | 2025-04-01T06:45:43.383073 | {
"authors": [
"eddydn",
"santalu"
],
"repo": "santalu/diagonal-imageview",
"url": "https://github.com/santalu/diagonal-imageview/issues/1",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
230930164 | Skip specific prompts
Sometimes we want to use cli flags as prompt's value:
module.exports = {
prompts: {
typescript: {
type: 'confirm',
default: false,
flag: 'typescript'
}
}
}
Then, if you init a project with sao template --typescript flag, the value of flag --typescript will be assigned to prompt typescript, and that prompt will be skipped.
Any news on this?
@KeitIG you can export a function in the config file and conditionally set what you should prompt 😄 https://sao.js.org/#/create?id=exporting-a-function
Try this recipe
| gharchive/issue | 2017-05-24T06:09:23 | 2025-04-01T06:45:43.385674 | {
"authors": [
"KeitIG",
"egoist",
"kristianmandrup"
],
"repo": "saojs/sao",
"url": "https://github.com/saojs/sao/issues/30",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2487730530 | Update README
Update the README.md to improve visual appeal and readability.
Links and Picture Grid:
Add links to external resources.
Create a grid of pictures in the "Reviews" section using the provided format for YouTube images.
Restrict image width to 128px
Arrange videos in 4 columns, 2 videos per line
Add CTA on adding reviews
reword cta for testimonials
✨✨ Here's an AI-assisted sketch of how you might approach this issue saved by @ajnart using Copilot Workspace v0.23
Topic
Let's make the README look better ! We will add an icon and different sections, we will aslo add links and a grid of pictures
Before
The current README.md includes an icon, sections, links, and a picture grid.
The README.md already has a well-structured format with various sections such as "How it works", "Tools", "Working in Existing Projects", "Only With Your Permission", "Contribution", "Reviews", "License", "Questions?", and "Acknowledgments".
The README.md contains multiple links to external resources, such as the VSCode Extension download link, Discord link, and various review articles and videos.
The README.md includes a grid of pictures in the "Reviews" section, showcasing different reviews and articles about the project.
After
The README.md now includes an icon at the top for better visual appeal.
The README.md now has distinct sections with clear headings for better readability.
The README.md now includes links to external resources and a grid of pictures in the "Reviews" section.
The README.md file now tells a story about the user journey, it will explain how someone can simply make a game using only Claudedev
Can we make a grid for the reviews instead ? Use this format to get the youtube image :
Make all of them inside of a table that contains video / author
Make it 4 columns instead, 2 videos per line and make the images restricted to 128px in width
Open all the links in a new page instead
Make all links in the table open in a new tab
Plan
README.md (CHANGE)
Add an icon at the top for better visual appeal
Add distinct sections with clear headings for better readability
Include links to external resources and a grid of pictures in the "Reviews" section
Tell a story about the user journey, explaining how someone can make a game using only Claude Dev
Make all links in the table open in a new tab
Sketch of implementation
View the changes
Details
Code analyzed at 63f3589f897075bf1fd98a657b0d9fa41df3ce7c
| gharchive/pull-request | 2024-08-26T20:34:29 | 2025-04-01T06:45:43.397104 | {
"authors": [
"ajnart"
],
"repo": "saoudrizwan/claude-dev",
"url": "https://github.com/saoudrizwan/claude-dev/pull/178",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1417072560 | Deploy Your SAPUI5 App
Tutorials: https://developers.sap.com/tutorials/sapui5-101-deploy-application.html
Hi, currently I have an issue with this tutorial any maybe someone could help:
I did the CF sign in but I cant select any org and space
But in general there is an existing org and space
Any suggestions how to solve this?
Hi @bmaciag, looks like you signed in to Cloud Foundry using the wrong API endpoint. The org you are trying to login to sits in an extension landscape (eu10-004) so you will have to change that in the SAP Business Application Studio. You can use the command cf login -a in the terminal to specify another API endpoint. There is probably also a way to do it via the GUI.
| gharchive/issue | 2022-10-20T18:26:22 | 2025-04-01T06:45:43.400549 | {
"authors": [
"bmaciag",
"nicoschoenteich"
],
"repo": "sap-tutorials/Tutorials",
"url": "https://github.com/sap-tutorials/Tutorials/issues/18947",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
465770905 | TypeError: Cannot read property 'hasOwnProperty' of undefined
There is an error in this file while using with sarala Model:
https://github.com/sarala-io/sarala-json-api-data-formatter/blob/master/src/Formatter.js#L119
TypeError: Cannot read property 'hasOwnProperty' of undefined
at eval (Formatter.js?8709:158)
at arrayEach (lodash.js?2ef0:516)
at Function.forEach (lodash.js?2ef0:9344)
at Formatter.mapAndKillProps (Formatter.js?8709:157)
at eval (Formatter.js?8709:207)
at arrayEach (lodash.js?2ef0:516)
at Function.forEach (lodash.js?2ef0:9344)
at Formatter.serializeOne (Formatter.js?8709:205)
at Formatter.serializeRelationship (Formatter.js?8709:255)
at eval (Formatter.js?8709:266)
Hi @ilyaskarim ,
Very sorry for the late late reply. This issue is been fixed in the v2.0.0
Hi @milroyfraser
I am still getting this error in Saralajs package - "sarala": "^0.3.10",
It happens only when you find a resource and try to update any of its fields.
It works fine when you create a new resource. Saraljs might not be on the latest version of this package.
ie.
let address = new Address();
let addressResource = await address.find(addressId)
addressResource.line_1 = "new line 1"
addressResource.save() // throws same error above
| gharchive/issue | 2019-07-09T13:08:10 | 2025-04-01T06:45:43.415096 | {
"authors": [
"ilyaskarim",
"maniyadv",
"milroyfraser"
],
"repo": "sarala-io/sarala-json-api-data-formatter",
"url": "https://github.com/sarala-io/sarala-json-api-data-formatter/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2003967105 | 🛑 Sarp Aykent is down
In 66a7ec5, Sarp Aykent (https://www.sarpaykent.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Sarp Aykent is back up in d043196 after 10 minutes.
| gharchive/issue | 2023-11-21T10:34:20 | 2025-04-01T06:45:43.418521 | {
"authors": [
"sarpaykent"
],
"repo": "sarpaykent/uptime",
"url": "https://github.com/sarpaykent/uptime/issues/793",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1925182969 | 🛑 Tian is down
In 5451187, Tian (https://www.tianaix.com) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Tian is back up in e48e034 after 1 hour, 42 minutes.
| gharchive/issue | 2023-10-04T00:44:57 | 2025-04-01T06:45:43.421089 | {
"authors": [
"sarpaykent"
],
"repo": "sarpaykent/uptime",
"url": "https://github.com/sarpaykent/uptime/issues/89",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2301979920 | Include Github actions to dependabot.yaml
Github actions should be automatically updated by dependabot.yaml, not only Rust software
It should be as easy as:
...
- package-ecosystem: "github-actions"
directory: "/"
schedule:
interval: "weekly"
| gharchive/issue | 2024-05-17T07:05:03 | 2025-04-01T06:45:43.424330 | {
"authors": [
"sarroutbi"
],
"repo": "sarroutbi/koji-retriever",
"url": "https://github.com/sarroutbi/koji-retriever/issues/214",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
685669379 | Broken link in confirm email
Somehow we have one link (the password reset link) in the automated confirm email linking to localhost rather than autism drive.
The instance config on production needs these three lines (it's missing one, so that one link is falling back to the default config which lists localhost):
FRONTEND_AUTH_CALLBACK = SITE_URL + "/#/session"
FRONTEND_EMAIL_RESET = SITE_URL + "/#/reset_password/"
FRONTEND_FORGOT_PASSWORD = SITE_URL + "/#/forgot-password"
| gharchive/issue | 2020-08-25T17:38:50 | 2025-04-01T06:45:43.435711 | {
"authors": [
"eleanorgraham"
],
"repo": "sartography/star-drive",
"url": "https://github.com/sartography/star-drive/issues/466",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
712046655 | Event Resource Details enhancements
By 10/2/2020
Edit form for Event
[ ] Add hero image
[ ] Add registration call-to-action button
[ ] Option to hide related resources
The draft of the landing page: https://autismdrive.virginia.edu/#/event/4127
image for the banner
| gharchive/issue | 2020-09-30T15:24:58 | 2025-04-01T06:45:43.438634 | {
"authors": [
"ajlouie",
"wendyhan2016"
],
"repo": "sartography/star-drive",
"url": "https://github.com/sartography/star-drive/issues/485",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
114812556 | Math operations with comments in between fail
Using
min-width: 20px /* height */ + 2*5px /* padding */ + 2*1px /*border*/;
results in
min-width: 20px 10px 2px;
The comments in between mess it up as
min-width: 20px + 2*5px + 2*1px; /* height + padding + border */
works fine.
Thanks for your time and effort!
Thanks for the report @sepptruetsch. I can confirm this is a bug with LibSass. I've created an issue to track it with them - https://github.com/sass/libsass/issues/1765
This is fixed in the upcoming 3.5.0-beta. Subscribe to #1349 for further updates.
| gharchive/issue | 2015-11-03T13:12:28 | 2025-04-01T06:45:43.465949 | {
"authors": [
"sepptruetsch",
"xzyfer"
],
"repo": "sass/node-sass",
"url": "https://github.com/sass/node-sass/issues/1242",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1809298629 | Bug report: MToon Appears Black on New Version
Hi, Reporting a bug on the MToon shader on the new version
Currently the MToon Doesn't work and Model appear as black as [Screenshot (1)]
from what I have suspect it might be that there is no Material Output Node at the end [Screenshot (2)]
I have been using the Addon in version 2.16 and the shader works fine
Thanks for reporting!
MToon seems to work fine in my Windows11+Blender3.6.0 environment. Please let us know your OS and Blender version so we can investigate.
Also, there is a tutorial on how to use MToon, so check it out https://vrm-addon-for-blender.info/en/material-mtoon/
The shader seems to work great! I just have to manually plug it in MToon, sorry for the misunderstanding. it's just that the older version plug the surface automatically
| gharchive/issue | 2023-07-18T07:15:19 | 2025-04-01T06:45:43.497044 | {
"authors": [
"Nickdnr",
"saturday06"
],
"repo": "saturday06/VRM-Addon-for-Blender",
"url": "https://github.com/saturday06/VRM-Addon-for-Blender/issues/278",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1921992502 | VRM1.0で出力したモデルの視点操作に不具合が起きる件
バグについて
どのようなバグが起きているのか書いてください。
VRM1.0モデルをで出力した場合、モデルの水平方向の視点操作ができなくなり、カメラ目線も機能しなくなる。
(0.xモデルでは上記不具合は起きません)
バグの再現方法
Blender(3.6.4)にVRM-Addon-for-Blender(2.20.7)をインストールし、Vroid studio(1.23.2)で出力した
v1.0モデルのファイルを読み込ませ、VRMファイルを再出力させると上記不具合が起きます。
Blender内で何か操作することに関係なく、単純にv1.0のVRMファイルの読み込みと書き出しをするだけで生じます。
バグが発生する手順を書いてください。
バグが発生するモデルのファイルやそのURL、エラーメッセージを添付してもらえると助かります。
Unityで確認すると、J_Adj_L_FaceEyeおよびJ_Adj_R_FaceEyeのテールが下向きになっていました。
Blenderでバグが発生たファイルを再度読み込むと、当該ボーンは正面を向いており、正常のように見えます。
応急処置として、Blenderのボーン編集モードで、2つのボーンのテールを上向き(ワールドz軸方向)に
してから出力させると、ボーンのテールは、ほぼ正面を向くので、例えば、VRM posing desktopなどの
アプリ上での視点操作やカメラ目線は正常に機能するようになります。しかし、このファイルを再度
Blenderに読み込ませ、上記応急操作をせずに出力すると、また視点操作の不具合が再発します。
期待する動作
どうのようにバグを直してほしいか書いてください。
視点操作において、ver0.xと同じような出力結果を希望します。
スクリーンショット
何か参考になりそうなスクショがあれば添付してください。
PC環境情報
OS: [Windows, macOS, Ubuntu Linux, ]
Windous10
Blenderのバージョン [2.93.5, 3.1.2, ...]
3.6.4
アドオンのバージョン [2.4.7, 2.3.8, ...]
2.20.7
なにか他に書いておきたいことはこちらへ
Vroid studio側に問題が潜んでいる可能性も排除できないので
その可能性があれば、VRoid事務局にも報告したいと思います。
ご報告ありがとうございます。この問題は非常に難しい問題のため、継続して改善中です。過去にもご報告あったのでリンクを貼ります https://github.com/saturday06/VRM-Addon-for-Blender/issues/227
直接の原因は、お使いの視点操作ソフトウェアが、VRMの視点操作に未対応なのだと思います。調査のため、どのような視点操作ソフトをお使いか教えていただけますでしょうか?
--------- 以下状況を解説します ---------
VRMの視点の操作をするためには、VRM0の場合はVRM Lookatコンポーネント、VRM1の場合はVRMInstanceコンポーネントの機能を使って操作する必要があります。これは、VRMの下記の仕様をカバーするするためです。
ブレンドシェイプによる視点操作と、ボーン回転による視点操作の切り替えができる。
VRM 1.0のボーン回転の場合は、ボーンのデフォルトの回転値はどんな値でも良い。
お使いの視点操作ソフトウェアはそれらのコンポーネントを使わず、直接ボーンの回転を操作しているのかなと思っています。それを行うと、今回と同様の事象が発生します。VRM posing desktopで正常なのは、VRM posing desktopはVRMの視点操作に対応しているからだと思います。また、VRMをUnity上のVRM未対応ソフトウェアで再生する場合は、一度何らかの変換ソフトウェアを噛ませる必要がありそうです。
おそらくVRChatかなと想像しています。VRChatはUnity 2019だったのでVRM 0.0からの持ち込みしかできなかったところ、VRChat Open BetaでUnity 2022に対応し、VRM 1.0からの持ち込みがベータ版で可能になったことで問題が顕在化します。
また、BlenderのVRMアドオンでは、子供の無いeyeボーンを上に向けてエクスポートしたとしても、再度インポートすると正面を向いてしまうという問題があります。
この問題が解決するまでの一時的対応として、eyeボーンに子ボーンを追加し、その子ボーンはeyeボーンの上に配置することで、eyeボーンの向きを上向きに固定する方法があります。エンドボーン、終端ボーンなどと呼ばれる手法です。
Blender 4.0の機能を使うことでこの問題が完全解決できそうなので、大変申し訳ないのですがこちらの手法で一時的対応をお願いします。
--------- 以下状況を解説します ---------
VRMには「ボーンの向き」という概念が存在しません。しかしBlenderにはボーンの向きという概念が存在するので、ボーンの向きを自動的に推定する必要があります。ボーンに子ボーンがある場合はその子ボーンの方向を向ければ良いのですが、子ボーンが無いボーンが向いている方向を決めることができません。その場合はどうにか方向をでっちあげる必要が出てきます。
VRMアドオンでは「目は正面向いているのが普通」という考えから、目のボーンに子ボーンが無い場合は正面を向けています。
このようなことから、目のボーンを上に向けてVRMエクスポートした場合「目のボーンが上を向いている」という情報は失われてしまい、再度インポートする際はまた「目は正面向いているのが普通」という推定が発生し、インポート時に目が正面を向いてしまうという現象です。
Blender 4.0から「ボーンの向きを気にしなくて済む設定」が追加されるので、それを用いることでこのような自動変換をしないで良くなる、という解決が可能になるため、完全解決はそれまでお待ちいただけると助かります。
早速のご回答有り難うございます。
過去の同内容の報告を確認できず、失礼しました。
解決の可能性が見通せているとのことですので、一安心しました。
VRMの挙動を確認しているソフトは、最近ver1.0に対応した、VRM Posing Desktop ver 4.09-2 および、VPic ver
2.0.0です。
共にVroid Hubで連携アプリとして紹介されています。
3DCGは、最近手を出したばかりの若輩者ですが
今後ともどうぞ宜しくお願いします。
2023年10月3日(火) 9:04 saturday06 @.***>:
また、BlenderのVRMアドオンでは、子供の無いeyeボーンを上に向けてエクスポートしたとしても、再度インポートすると正面を向いてしまうという問題があります。
この問題が解決するまでの一時的対応として、eyeボーンに子ボーンを追加し、その子ボーンはeyeボーンの上に配置することで、eyeボーンの向きを上向きに固定する方法があります。エンドボーン、終端ボーンなどと呼ばれる手法です。
Blender 4.0の機能を使うことでこの問題が完全解決できそうなので、大変申し訳ないのですがこちらの手法で一時的対応をお願いします。
--------- 以下状況を解説します ---------
VRMには「ボーンの向き」という概念が存在しません。しかしBlenderにはボーンの向きという概念が存在するので、ボーンの向きを自動的に推定する必要があります。ボーンに子ボーンがある場合はその子ボーンの方向を向ければ良いのですが、子ボーンが無いボーンが向いている方向を決めることができません。その場合はどうにか方向をでっちあげる必要が出てきます。
VRMアドオンでは「目は正面向いているのが普通」という考えから、目のボーンに子ボーンが無い場合は正面を向けています。
このようなことから、目のボーンを上に向けてVRMエクスポートした場合「目のボーンが上を向いている」という情報は失われてしまい、再度インポートする際はまた「目は正面向いているのが普通」という推定が発生し、インポート時に目が正面を向いてしまうという現象です。
Blender
4.0から「ボーンの向きを気にしなくて済む設定」が追加されるので、それを用いることでこのような自動変換をしないで良くなる、という解決が可能になるため、完全解決はそれまでお待ちいただけると助かります。
—
Reply to this email directly, view it on GitHub
https://github.com/saturday06/VRM-Addon-for-Blender/issues/329#issuecomment-1743944022,
or unsubscribe
https://github.com/notifications/unsubscribe-auth/BC7Q2ZNGLFE6CKV25OPUZ7DX5NI7LAVCNFSM6AAAAAA5PPN5UCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTONBTHE2DIMBSGI
.
You are receiving this because you authored the thread.Message ID:
@.***>
VRM Posing DesktopやVPicなどのアプリで視点操作に不具合が発生する件は、UniVRMの不具合が原因のようでした。本件UniVRMにバグ報告をしました。
こちらでバグ報告の最新の状況が確認できます → https://github.com/vrm-c/UniVRM/issues/2159
VRMお手軽ポーズ (https://www.vrmwebpose.app/) やVRoid Hubの写真撮影画面など、UniVRMではなくthree-vrmを用いるアプリでは、本BlenderアドオンからエクスポートしたVRM1.0モデルにおいても視点操作は正しく動作していることを確認しました。
UniVRMの視点操作のバグが修正されました (https://github.com/vrm-c/UniVRM/pull/2163)。VRM Posing DesktopやVPicで利用されているUniVRMが最新バージョンにアップデートされた際に、このアドオンで作成されたモデルの視点操作が正しく行われるようになります。
報告いただいたケースは、VRMインポートアプリが視点操作のバグを含む古いUniVRMを利用しているのが原因のバグでした。UniVRM側が修正完了したため、こちらは解決済みとします
| gharchive/issue | 2023-10-02T14:06:09 | 2025-04-01T06:45:43.514024 | {
"authors": [
"saturday06",
"teresayawn"
],
"repo": "saturday06/VRM-Addon-for-Blender",
"url": "https://github.com/saturday06/VRM-Addon-for-Blender/issues/329",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1103893727 | Update .env_deps
:rocket: An automated PR
@wreis weird CI issue here - seen this before?
@forana yeah. it has been fixed in 11.10 branch; considering this is a fork from a open-source project, I haven't got to create the PR to main workflow yet. I will create the PR now.
https://github.com/saturncloud/sshpiper/pull/13
| gharchive/pull-request | 2022-01-14T16:46:58 | 2025-04-01T06:45:43.518389 | {
"authors": [
"forana",
"saturn-automation",
"wreis"
],
"repo": "saturncloud/sshpiper",
"url": "https://github.com/saturncloud/sshpiper/pull/12",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1705567513 | Update test-configuration-options.md
Fix a bug in our docs
@wswebcreation Thank you!
@wswebcreation Thank you!
Thanks you @lysannep for the feedback and approving it
| gharchive/pull-request | 2023-05-11T10:53:51 | 2025-04-01T06:45:43.532220 | {
"authors": [
"lysannep",
"wswebcreation"
],
"repo": "saucelabs/sauce-docs",
"url": "https://github.com/saucelabs/sauce-docs/pull/1986",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
909911030 | Refactoring Best Practices content -> Test Config moved
Thank you for contributing to the sauce-docs project!
A well described pull request helps maintainers quickly review and merge your change
Before submitting your PR, please check our contributing
guidelines, applied for this repository. Avoid large PRs, help reviewers by making them as simple
and short as possible.
Description
Motivation and Context
Types of changes
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to change)
[X] Documentation fix (typos, incorrect content, missing content, etc.)
Checklist
[ ] I have read the contributing document.
[ ] My change requires a change to the documentation.
[ ] I have updated the documentation accordingly.
[ ] I have added tests to cover my changes.
[ ] All new and existing tests passed.
Yes! As part of the BP content re-structuring, I had to move the Test Config content.
Of course it did. I'll fix them now.
@sweeneyskirt-sl Fixed--I think. Do I need to create a new PR or can you go from here?
| gharchive/pull-request | 2021-06-02T21:47:09 | 2025-04-01T06:45:43.538088 | {
"authors": [
"lauramthomas"
],
"repo": "saucelabs/sauce-docs",
"url": "https://github.com/saucelabs/sauce-docs/pull/425",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2754682163 | 🛑 cubeent is down
In bb1ea95, cubeent (http://www.cubeent.co.kr) was down:
HTTP code: 0
Response time: 0 ms
Resolved: cubeent is back up in f748908 after 12 minutes.
| gharchive/issue | 2024-12-22T14:22:08 | 2025-04-01T06:45:43.623761 | {
"authors": [
"savorfamily"
],
"repo": "savorfamily/upptime",
"url": "https://github.com/savorfamily/upptime/issues/732",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
476877483 | Advice about atreugo
Hope to increase the support of mysql and sqlite.
What do you mean?
Thanks, I have written the database method myself, directly using github.com/go-sql-driver/mysql and github.com/mattn/go-sqlite3
But only with this project, I will rewrite it next time.
Sorry, but I don't know what do you need exactly 😥
atreugo is a framework about httpserver,you can add mysql and sqlite module by yourself with some orm such as xorms
Thanks, I have written the database method myself, directly using github.com/go-sql-driver/mysql and github.com/mattn/go-sqlite3
But only with this project, I will rewrite it next time.
| gharchive/issue | 2019-08-05T14:02:16 | 2025-04-01T06:45:43.627056 | {
"authors": [
"imxxiv",
"kael777",
"savsgio"
],
"repo": "savsgio/atreugo",
"url": "https://github.com/savsgio/atreugo/issues/30",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1591777862 | Workshop: Booleans in Cairo
Level: Easy
As Cairo doesn't have a native boolean type, it is not uncommon to see developers using 0 and 1 as a replacement.
E.g.
if (result == 0):
//Do Something
However, for readability purposes we prefer to utilize the FALSE and TRUE constants from the common cairo-lang library:
https://github.com/starkware-libs/cairo-lang/blob/master/src/starkware/cairo/common/bool.cairo
In the following file 1 is still used instead of the TRUE constant.
https://github.com/sayajin-labs/kakarot/blob/f519876a3acb67507ccea05e1f87d82f36093142/src/utils/utils.cairo#L257-L260
To fix this import the relevant cairo-lang library and replace 1 with the TRUE constant.
Hello! Can I take this?
Sure, go ahead :)
| gharchive/issue | 2023-02-20T12:32:52 | 2025-04-01T06:45:43.635205 | {
"authors": [
"TotalPizza",
"weiihann"
],
"repo": "sayajin-labs/kakarot",
"url": "https://github.com/sayajin-labs/kakarot/issues/517",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
34272802 | Empty canvas?
I did just the following reference for the ng-controller:
function FilesController($scope, $http) {
$scope.pdfUrl = '/doc.pdf';
}
And the view ist just as simple as:
<div ng-controller="FilesController">
<nav ng-class="{'pdf-controls fixed': scroll > 100, 'pdf-controls': scroll <= 100}">
<button ng-click="goPrevious()"><</span></button>
<button ng-click="goNext()">></span></button>
<button ng-click="zoomIn()">+</span></button>
<button ng-click="zoomOut()">-</span></button>
<button ng-click="rotate()">90</span></button>
<span>Page: </span>
<input type="text" min="1" ng-model="pageNum">
<span> / {{ pageCount }}</span> (is empty)
</nav>
<hr>
<canvas id="pdf-canvas" class="rotate0"></canvas>
</div>
This should work, right? Strange it doesn't.. No console errors.. What am I missing?
Hi!
First of all thanks for this directive!
I have the same problem: an empty canvas!
My directive is in a view which is routed to by a stateprovider:
<ion-view view-title="{{galleryTitle | uppercase}}">
<ion-content>
<div class="wrapper" ng-controller="pdfViewerController">
<ng-pdf template-url="templates/views/viewer.html" canvasid="pdf"></ng-pdf>
</div>
</ion-content>
</ion-view>
This is the viewer.html with the canvas:
<canvas id="pdf"></canvas>
My controller:
app.controller('pdfViewerController', function($scope, PdfUriService) {
console.log(PdfUriService.getPdfUri());
$scope.pdfUrl = PdfUriService.getPdfUri();
$scope.onError = function(error) {
// handle the error
console.log(error);
}
$scope.onProgress = function(progress) {
// handle a progress bar
//progress% = progress.loaded / progress.total
console.log(progress);
}
});
When i check the console i can see the correct path to the pdf. Any guess what might be wrong?
Thanks a lot!
| gharchive/issue | 2014-05-25T21:00:41 | 2025-04-01T06:45:43.639185 | {
"authors": [
"raenzen",
"renege"
],
"repo": "sayanee/angularjs-pdf",
"url": "https://github.com/sayanee/angularjs-pdf/issues/9",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2496468615 | feat(sbb-header): introduce size s
Draft - to display the logo without text a new property has been added to the sbb-logo component.
If the API change is fine, the final visual result should be validated by @mcilurzo, otherwise a different solution (new SVG? new component?) should be find.
EDIT:
API change is fine (signetOnly is the property proposed name)
no need of a size prop on the header-button/link
ux review:
https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=43777-147301#935958908
https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=43777-147442#935950489
https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=43910-264561#936057771
ux review: https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=43777-147301#935958908 https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=43777-147442#935950489 https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=43910-264561#936057771
@mcilurzo
I added the application name/version in stories size=s, applying some base styling class
may I ask to you to recheck the height? i see it correctly
16px
20px from medium
As discussed via MS Teams:
no changes on the sbb-logo component
add a new padded prop to the sbb-signet to display the logo without text
use the full logo or the signet for the header in size=s is a consumer responsibility, so no changes on header component but only in stories
ux review:
https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=44464-179452#946527963
https://www.figma.com/design/9r6xSfNmEfCFxl1yFYedrj?node-id=44464-179452#946820512
LGTM. Thank you!
| gharchive/pull-request | 2024-08-30T07:33:57 | 2025-04-01T06:45:43.697146 | {
"authors": [
"DavideMininni-Fincons",
"mcilurzo"
],
"repo": "sbb-design-systems/lyne-components",
"url": "https://github.com/sbb-design-systems/lyne-components/pull/3047",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1980501112 | What happened?
OneDrive个人版:无法鉴权
app.js:1 Obsidian Developer Console
plugin:obsidian-markdown-furigana:93 loading Markdown Furigana plugin
app.js:1 Uncaught (in promise) TypeError: Cannot read properties of undefined (reading 'length')
at e.printHotkeyForCommand (app.js:1:1096737)
at t. (app.js:1:1295494)
at app.js:1:237027
at Object.next (app.js:1:237132)
at app.js:1:236048
at new Promise ()
at v (app.js:1:235793)
at t.onOpen (app.js:1:1295291)
at t. (app.js:1:1287590)
at app.js:1:237027
plugin:longform:37647 [Longform] Starting Longform 2.0.4…
plugin:obsidian-outliner:3317 Loading obsidian-outliner
plugin:obsidian-weread-plugin:2 load weread plugin
plugin:obsidian-weread-plugin:2 --------init cookie------ Array(10)
plugin:obsidian-weread-plugin:2 [weread plugin] setting user vid=> 8671170
plugin:obsidian-weread-plugin:2 [weread plugin] setting user name=> 考槃
plugin:table-editor-obsidian:28663 loading markdown-table-editor plugin
plugin:dataview:20020 Dataview: version 0.5.64 (requires obsidian 0.13.11)
plugin:note-refactor-obsidian:6572 Loading Note Refactor plugin
plugin:quickadd:19504 Loading QuickAdd
plugin:quickadd:17504 QuickAdd: (LOG) No migrations to run.
plugin:obsidian42-brat:1970 loading Obsidian42 - BRAT
plugin:attachment-management:1840 Plugin loading: Attachment Management v.0.8.4
plugin:card-board:1 loading CardBoard plugin
plugin:webpage-html-export:57369 loading webpage-html-export plugin
raw.githubusercontent.com/KosmosisDire/obsidian-webpage-export/master/manifest.json?cache=1699329055057:1 Failed to load resource: net::ERR_ADDRESS_INVALID
plugin:webpage-html-export:57451 Could not check for update
plugin:webpage-html-export:56733 Error while minifying CSS file.:
Error: Parse Error: <"]:checked,
li[data-task="<"]>input:checked,
li[data-task="<"]>p>input:checked,
input[data-task=">"]:checked,
li[data-task=">"]>input:checked,
li[data-task=">"]>p>input:checked,
input[data-task="d"]:checked,
li[data-task="d"]>input:checked,
li[data-task="d"]>p>input:checked,
input[data-task="u"]:checked,
li[data-task="u"]>input:checked,
li[data-task="u"]>p>input:checked,
input[data-task="S"]:checked,
li[data-task="S"]>input:checked,
li[data-task="S"]>p>input:checked,
input[data-task="“"]:checked,
li[data-task="“"]>input:checked,
li[data-task="“"]>p>input:checked,
input[data-task="c"]:checked,
li[data-task="c"]>input:checked,
li[data-task="c"]>p>input:checked,
input[data-task="p"]:checked,
li[data-task="p"]>input:checked,
li[data-task="p"]>p>input:checked,
input[data-task="n"]:checked,
li[data-task="n"]>input:checked,
li[data-task="n"]>p>input:checked,
input[data-task="b"]:checked,
li[data-task="b"]>input:checked,
li[data-task="b"]>p>input:checked,
input[data-task="-"]:checked,
li[data-task="-"]>input:checked,
li[data-task="-"]>p>input:checked,
input[data-task="l"]:checked,
li[data-task="l"]>input:checked,
li[data-task="l"]>p>input:checked,
input[data-task="I"]:checked,
li[data-task="I"]>input:checked,
li[data-task="I"]>p>input:checked,
input[data-task="/"]:checked,
li[data-task="/"]>input:checked,
li[data-task="/"]>p>input:checked,
input[data-task="i"]:checked,
li[data-task="i"]>input:checked,
li[data-task="i"]>p>input:checked,
input[data-task=""]:checked,
li[data-task=""]>input:checked,
li[data-task="*"]>p>input:checked,
input[data-task="?"]:checked,
li[data-task="?"]>input:checked,
li[data-task="?"]>p>input:checked,
input[data-task="!"]:checked,
li[data-task="!"]>input:checked,
li[data-task="!"]>p>input:checked {
--checkbox-marker-color: transparent;
border: none;
border-radius: 0;
background-image: none;
background-color: currentColor;
-webkit-mask-size: var(--checkbox-icon);
-webkit-mask-position: 50% 50%
}
input[data-task="!"]:checked,
li[data-task="!"]>input:checked,
li[data-task="!"]>p>input:checked {
--checkbox-color-hover: var(--color-orange);
color: var(--color-orange);
-webkit-mask-image: url('data:image/svg+xml;utf8, vwzkzukp0zp12vwzkzukp0zp ');
}
input[data-task="?"]:checked,
li[data-task="?"]>input:checked,
li[data-task="?"]>p>input:checked {
--checkbox-color-hover: var(--color-pink);
color: var(--color-pink);
-webkit-mask-image: url('data:image/svg+xml;utf8, vwzkzukp0zp13vwzkzukp0zp ');
}
input[data-task=""]:checked,
li[data-task=""]>input:checked,
li[data-task="*"]>p>input:checked {
--checkbox-color-hover: var(--color-yellow);
color: var(--color-yellow);
-webkit-mask-image: url('data:image/svg+xml;utf8, vwzkzukp0zp14vwzkzukp0zp <path d="M8.91,.58c-.08-.17-.21-.32-.37-.42C8.38,.05,8.19,0,8,0s-.38,.05-.54,.16c-.16,.1-.29,.25-.37,.42l-1.93,4.12L.85,5.36c-.18,.03-.35,.1-.49,.22-.14,.12-.25,.27-.3,.45-.06,.17-.07,.36-.03,.54,.04,.18,.13,.34,.26,.48l3.15,3.23-.75,4.57c-.03,.19,0,.38,.06,.55,.07,.17,.19,.32,.35,.43,.15,.11,.33,.17,.52,.18,.19,0,.37-.03,.54-.12l3.84-2.13,3.84,2.13c.16,.09,.35,.13,.54,.12,.19-.01,.37-.07,.52-.18,.15-.11,.27-.26,.35-.43,.07-.17,.09-.36,.06-.55l-.75-4.57,3.15-3.23c.13-.13,.22-.3,.26-.48,.04-.18,.03-.37-.03-.54-.06-.17-.16-.33-.3
error @ plugin:webpage-html-export:56733
plugin:webpage-html-export:56640 Uncaught (in promise) TypeError: Cannot read properties of undefined (reading 'view')
at Object._reportError (plugin:webpage-html-export:56640:23)
at Object.error (plugin:webpage-html-export:56734:22)
at _AssetHandler.minifyJSorCSS (plugin:webpage-html-export:55766:17)
at process.processTicksAndRejections (node:internal/process/task_queues:95:5)
at async _AssetHandler.updateAssetCache (plugin:webpage-html-export:55822:26)
plugin:cm-chs-patch:7 Jieba loaded
plugin:cm-chs-patch:7 editor word splitting patched
plugin:nl-fast-image-cleaner:567 Fast file Cleaner plugin loaded...
plugin:obsidian-tasks-plugin:289 loading plugin "tasks"
plugin:obsidian-excalidraw-plugin:79 Initialized Excalidraw Image Cache
plugin:dataview:12759 Dataview: all 823 files have been indexed in 1.913s (823 cached, 0 skipped).
plugin:commentator:8 commentator: database load: 2139.44677734375 ms
plugin:commentator:8 commentator: syncing database: 5.7939453125 ms
plugin:longform:35996 [Longform] Loaded and watching projects. Found 2 drafts in 9.431s.
plugin:excalibrain:1 ExcaliBrain indexed 1721 URLs from 165 unique hosts in 120 of 823 markdown files in 15.1 seconds
7VM268 plugin:remotely-secure:53 Uncaught (in promise) ClientConfigurationError: empty_url_error: URL was empty or null.
at e [as constructor] (plugin:remotely-secure:53:52965)
at e [as constructor] (plugin:remotely-secure:53:62298)
at new e (plugin:remotely-secure:53:76448)
at e.createUrlEmptyError (plugin:remotely-secure:53:77222)
at new t (plugin:remotely-secure:53:122232)
at t.transformCIAMAuthority (plugin:remotely-secure:57:2266)
at Function.eval (plugin:remotely-secure:57:2833)
at h (plugin:remotely-secure:53:45310)
at Object.eval [as next] (plugin:remotely-secure:53:44614)
at eval (plugin:remotely-secure:53:44340)
at new Promise ()
at lt (plugin:remotely-secure:53:44137)
at t.createDiscoveredInstance (plugin:remotely-secure:57:2665)
at EE.createAuthority (plugin:remotely-secure:68:21968)
at EE.buildOauthClientConfiguration (plugin:remotely-secure:68:19557)
at EE.getAuthCodeUrl (plugin:remotely-secure:68:16518)
at async qie (plugin:remotely-secure:68:26998)
at async gT.onOpen (plugin:remotely-secure:111:1741)
2plugin:remotely-secure:30 Uncaught (in promise) Error: A client id is required. You can set the client id using .setClientId().
at F.value (plugin:remotely-secure:30:44023)
at Cee (plugin:remotely-secure:53:37127)
at mT.onOpen (plugin:remotely-secure:109:1624)
at e.open (app.js:1:1309979)
at eval (plugin:remotely-secure:114:7333)
at HTMLButtonElement. (app.js:1:1318112)
5plugin:remotely-secure:53 Uncaught (in promise) ClientConfigurationError: empty_url_error: URL was empty or null.
at e [as constructor] (plugin:remotely-secure:53:52965)
at e [as constructor] (plugin:remotely-secure:53:62298)
at new e (plugin:remotely-secure:53:76448)
at e.createUrlEmptyError (plugin:remotely-secure:53:77222)
at new t (plugin:remotely-secure:53:122232)
at t.transformCIAMAuthority (plugin:remotely-secure:57:2266)
at Function.eval (plugin:remotely-secure:57:2833)
at h (plugin:remotely-secure:53:45310)
at Object.eval [as next] (plugin:remotely-secure:53:44614)
at eval (plugin:remotely-secure:53:44340)
at new Promise ()
at lt (plugin:remotely-secure:53:44137)
at t.createDiscoveredInstance (plugin:remotely-secure:57:2665)
at EE.createAuthority (plugin:remotely-secure:68:21968)
at EE.buildOauthClientConfiguration (plugin:remotely-secure:68:19557)
at EE.getAuthCodeUrl (plugin:remotely-secure:68:16518)
at async qie (plugin:remotely-secure:68:26998)
at async gT.onOpen (plugin:remotely-secure:111:1741)
What OS are you using?
Windows
What remote cloud services are you using?
OneDrive for personal
Version of the plugin
0.4.1
Version of Obsidian
1.4.16
Using password or not
[ ] Yes.
Ensure no sensitive information
[X] I ensure that no sensitive information is submitted in the issue.
Hello, thank you for the bug report.
Looks like I need to register a client application with onedrive as this uses oauth code flow. Will update this for the next official release, sorry about that.
Hello, thank you for the bug report.
Looks like I need to register a client application with onedrive as this uses oauth code flow. Will update this for the next official release, sorry about that.
Thank you for your contribution to Obsidian.
| gharchive/issue | 2023-11-07T04:05:45 | 2025-04-01T06:45:43.743178 | {
"authors": [
"NewTypo",
"sboesen"
],
"repo": "sboesen/remotely-secure",
"url": "https://github.com/sboesen/remotely-secure/issues/2",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1290241728 | Include ticket system in website
Replace the "Feedback to Website" in Footer with the new ticket system
is now added under "Internes > Tools"
| gharchive/issue | 2022-06-30T14:44:57 | 2025-04-01T06:45:43.749554 | {
"authors": [
"dominikgeissler"
],
"repo": "sbsev/site",
"url": "https://github.com/sbsev/site/issues/62",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
341851241 | multiple range versions don't resolve correctly
[x] I've read the CONTRIBUTING guidelines
steps
Create new project: sbt new scala/scala-seed.g8
Add following dependency: libraryDependencies += "org.webjars.npm" % "loose-envify" % "1.4.0"
Compile the project: sbt compile
problem
The library is not resolved:
[info] Updating ...
[warn] module not found: org.webjars.npm#js-tokens;[3.0.0,4),[4.0.0,5)
[warn] ==== local: tried
[warn] C:\Users\razvan.rotaru\.ivy2\local\org.webjars.npm\js-tokens\[3.0.0,4),[4.0.0,5)\ivys\ivy.xml
[warn] ==== public: tried
[warn] https://repo1.maven.org/maven2/org/webjars/npm/js-tokens/[3.0.0,4),[4.0.0,5)/js-tokens-[3.0.0,4),[4.0.0,5).pom
[warn] ==== local-preloaded-ivy: tried
[warn] C:\Users\razvan.rotaru\.sbt\preloaded\org.webjars.npm\js-tokens\[3.0.0,4),[4.0.0,5)\ivys\ivy.xml
[warn] ==== local-preloaded: tried
[warn] file:/C:/Users/razvan.rotaru/.sbt/preloaded/org/webjars/npm/js-tokens/[3.0.0,4),[4.0.0,5)/js-tokens-[3.0.0,4),[4.0.0,5).pom
[info] downloading https://repo1.maven.org/maven2/org/scalatest/scalatest_2.12/3.0.5/scalatest_2.12-3.0.5.jar ...
[info] downloading https://repo1.maven.org/maven2/org/scalactic/scalactic_2.12/3.0.5/scalactic_2.12-3.0.5.jar ...
[info] [SUCCESSFUL ] org.scalactic#scalactic_2.12;3.0.5!scalactic_2.12.jar(bundle) (1309ms)
[info] [SUCCESSFUL ] org.scalatest#scalatest_2.12;3.0.5!scalatest_2.12.jar(bundle) (3490ms)
[warn] ::::::::::::::::::::::::::::::::::::::::::::::
[warn] :: UNRESOLVED DEPENDENCIES ::
[warn] ::::::::::::::::::::::::::::::::::::::::::::::
[warn] :: org.webjars.npm#js-tokens;[3.0.0,4),[4.0.0,5): not found
[warn] ::::::::::::::::::::::::::::::::::::::::::::::
[warn]
[warn] Note: Unresolved dependencies path:
[warn] org.webjars.npm:js-tokens:[3.0.0,4),[4.0.0,5)
[warn] +- org.webjars.npm:loose-envify:1.4.0 (D:\imis\sbttest\sbttest\build.sbt#L13)
[warn] +- com.example:sbttest_2.12:0.1.0-SNAPSHOT
The problem seems to be the presence of two version ranges in the dependency list of loose-envify:
<dependencies>
<dependency org="org.webjars.npm" name="js-tokens" rev="[3.0.0,4),[4.0.0,5)" force="true" conf="compile->compile(*),master(compile);runtime->runtime(*)"/>
</dependencies>
expectation
The project is compiled and a version of js-tokens is downloaded.
https://repo1.maven.org/maven2/org/webjars/npm/js-tokens/3.0.2/js-tokens-3.0.2.pom
notes
sbt version: 1.1.6
Thanks for the report with repo steps!
It was reported before: https://github.com/sbt/sbt/issues/2647
@tarsa Thanks for that list. I am going through a bunch of issues, and this is helpful.
This works using sbt 1.3.0-RC3, which uses Coursier internally.
| gharchive/issue | 2018-07-17T10:04:06 | 2025-04-01T06:45:43.755045 | {
"authors": [
"eed3si9n",
"roti",
"tarsa"
],
"repo": "sbt/sbt",
"url": "https://github.com/sbt/sbt/issues/4271",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
727307969 | Undefined Setting Key externalDepedencyClasspath if JVM Plugin is disabled and products is set to nil
steps
Given the following build.sbt with an implicit root project:
val foo = project
.disablePlugins(sbt.plugins.JvmPlugin)
.settings(
Compile / products := Nil
)
lazy val myTask = taskKey[Unit]("Should work")
myTask := (foo / Compile / externalDependencyClasspath).value foreach println
problem
when trying to execute the task with sbt myTask I get an error that the setting externalDependencyClasspath is not defined:
[error] Reference to undefined setting:
[error]
[error] foo / Compile / externalDependencyClasspath from myTask
Once I either remove Compile / products := Nil or enable the JvmPlugin the task works as expected. I can't seem to find the relationship between products and externalDependencyClasspath in the sbt source code that could explain this behavior.
expectation
The task should not fail.
notes
My use case comes from a plugin that tries to get all dependencies from aggregate projects when executed on the root project and it fails for this specifically configured sub-project.
I tested this with sbt 1.4.0 and 1.2.8.
For the record, externalDependencyClasspath is introduced as part of configSettings, so I don't think it should work if you've disabled JvmPlugin.
https://github.com/sbt/sbt/blob/3536959c311d4922bc0803f51a5a0bca57440719/main/src/main/scala/sbt/Defaults.scala#L2464-L2478
Once I either remove Compile / products := Nil .... the task works as expected.
I can't reproduce this. With or without Compile / products := Nil, I correctly get
[error] Reference to undefined setting:
[error]
[error] foo / Compile / externalDependencyClasspath from myTask (/private/tmp/foo/build.sbt:8)
[error]
[error] Use 'last' for the full log.
[warn] Project loading failed: (r)etry, (q)uit, (l)ast, or (i)gnore? (default: r)
| gharchive/issue | 2020-10-22T11:37:22 | 2025-04-01T06:45:43.760076 | {
"authors": [
"albuch",
"eed3si9n"
],
"repo": "sbt/sbt",
"url": "https://github.com/sbt/sbt/issues/6009",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
999150843 | JDK 17
Fixes #6661
Some test failures, but generally sbt seems to be functioning ok -- sans the warning about security manager.
[error] (scripted) Failed tests:
[error] classloader-cache/spark
[error] compiler-project/run-test
[error] dependency-graph/ignoreScalaLibrary
[error] dependency-graph/toFileSubTask
[error] project/unified
| gharchive/pull-request | 2021-09-17T09:34:54 | 2025-04-01T06:45:43.761228 | {
"authors": [
"eed3si9n"
],
"repo": "sbt/sbt",
"url": "https://github.com/sbt/sbt/pull/6662",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
261975086 | implement Snippet mechanism like Paradox
This implements low-tech solution for compiled documentation, inspired by Lightbend Paradox's snippet feature. It's low tech because this feature itself just grabs the referenced file as a code block and includes it, and assumes compile or other validation is done elsewhere. (In our case we are using scripted) This low tech approach is nice because it's fast.
// This includes the entire file as Scala code snippet
@@snip [build.sbt]($root$/src/sbt-test/ref/basic/build.sbt) {}
or
// This includes snippet between a line containing #example another line with #example
@@snip [build.sbt]($root$/src/sbt-test/ref/basic/build.sbt) { #example }
or
// This specifies syntax highlight
@@snip [build.sbt]($root$/src/sbt-test/ref/basic/build.sbt) { #example type=text }
There's a hard-coded treatment for $root$, which is interpreted to be the root of the build. Otherwise, the path is treated to the relative path from the markdown file.
Oh the dilemma.. not having these code blocks inline is a major pain, but enforcing that they're absolutely correct is a great benefit..
Any chance we can refine this to leave them inline?
@dwijnand Well, if we are compatible with Paradox behavior this is how it works.
If we leave anything in but not sync it, then it's confusing.
So a way it work might be:
@@snip [build.sbt]($root$/src/sbt-test/ref/basic/build.sbt) {}
<!-- SNIIPPET BEGIN. DO NOT EDIT MANUALLY -->
```
lazy val root ...
```
<!-- END SNIIPPET -->
and during the "compile"-time we just remove "@@snip" line and HTML-comment (Pamflet unfortunately doesn't support HTML-comment https://github.com/foundweekends/pamflet/issues/70).
I was hoping to have the same behaviour as tut, where you can just write:
Here is how you add numbers:
```tut
1 + 1
```
So I guess ideally we'd have:
Use the following build definition:
```sbt
val root = project in file(".")
```
I've tried that approach with sbt-pamflet https://github.com/sbt/sbt-pamflet, and it's super slow. Besides, I am not sure if that makes sense for builds.
If you are suggesting that we go the opposite direction by generating scripted out of snippet, I don't think it would always work out since we often want part of the code, not the whole thing.
I think we should merge this as is for now, also for the Paradox compatibility.
If you were considering Paradox, then we should keep the snippet syntax the same.
I want to move on and focus my energy on actually documenting things using this feature.
| gharchive/pull-request | 2017-10-02T03:32:07 | 2025-04-01T06:45:43.767336 | {
"authors": [
"dwijnand",
"eed3si9n"
],
"repo": "sbt/website",
"url": "https://github.com/sbt/website/pull/508",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1850432252 | Package Registration
@JuliaRegistrator register()
Registration pull request created: JuliaRegistries/General/89632
After the above pull request is merged, it is recommended that a tag is created on this repository for the registered package version.
This will be done automatically if the Julia TagBot GitHub Action is installed, or can be done manually through the github interface, or via:
git tag -a v1.0.0 -m "<description of version>" d5bb861e3c7c4df3a27c7a7ecc5a0d84f388d7ab
git push origin v1.0.0
| gharchive/issue | 2023-08-14T19:44:07 | 2025-04-01T06:45:43.769510 | {
"authors": [
"JuliaRegistrator",
"sc-dyer"
],
"repo": "sc-dyer/PetroBase.jl",
"url": "https://github.com/sc-dyer/PetroBase.jl/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1989300919 | merge: latest main into cli (including inherited functions for hardhat)
Description
I'm going to start with inherited functions for Foundry and wanted to have the UI merged first.
Additional Information
[x] I have read the contributing docs (if this is your first contribution)
[x] This is not a duplicate of any existing pull request
Related Issues
Your ENS/address: harmont.eth
I've tested it locally now as well.
yarn cli --dev works like a charm @dgrcode
Hey @FilipHarald, thanks for this!
I think you branched from an old commit (#592 was the last commit merged on the CLI branch: Check https://github.com/scaffold-eth/scaffold-eth-2/pull/595)
Check this section of the guide we wrote for CLI dev: https://github.com/scaffold-eth/scaffold-eth-2/blob/cli/contributors/DEVELOPER-GUIDE.md#back-merging-main-branch--publishing-to-npm
Could you re-do this PR branching from the latest changes of CLI + commit the changeset? hmu if you need help.
Thanks! <3
| gharchive/pull-request | 2023-11-12T08:18:56 | 2025-04-01T06:45:43.782670 | {
"authors": [
"FilipHarald",
"carletex"
],
"repo": "scaffold-eth/scaffold-eth-2",
"url": "https://github.com/scaffold-eth/scaffold-eth-2/pull/601",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
116986631 | run/t6246.scala is broken on master
Nightly has been failing consistently over run/t6246.scala, both in 2.11 and 2.12.
It's a linking error:
$ cat t6246-run.log
Referring to non-existent method jl_Class.getInterfaces__Ajl_Class
called from s_reflect_ClassManifestDeprecatedApis$class.loop$1__p0__s_reflect_ClassTag__sci_Set__sci_Set__jl_Class__Z
called from s_reflect_ClassManifestDeprecatedApis$class.subtype__p0__s_reflect_ClassTag__jl_Class__jl_Class__Z
called from s_reflect_ClassManifestDeprecatedApis$class.$$less$colon$less__s_reflect_ClassTag__s_reflect_ClassTag__Z
called from s_reflect_ClassTag$ClassClassTag.$$less$colon$less__s_reflect_ClassTag__Z
called from LTest$.testValueClass__s_reflect_ClassTag__V
called from LTest$.delayedEndpoint$Test$1__V
called from LTest$delayedInit$body.apply__O
called from s_Function0$class.apply$mcV$sp__F0__V
called from sr_AbstractFunction0.apply$mcV$sp__V
called from s_App$class.$$anonfun$1__p0__s_App__F0__V
called from s_App$class.main__s_App__AT__V
called from LTest$.main__AT__V
called from LPartestLauncher$.launch
exported to JavaScript with @JSExport
involving instantiated classes:
s_reflect_ClassTag$ClassClassTag
LTest$
(already seen, not repeating call stack)
LTest$delayedInit$body
(already seen, not repeating call stack)
sjsr_AnonFunction0
(already seen, not repeating call stack)
LPartestLauncher$
(already seen, not repeating call stack)
Exception in thread "main" java.lang.RuntimeException: There were linking errors
at scala.sys.package$.error(package.scala:27)
at org.scalajs.core.tools.optimizer.Linker.link(Linker.scala:102)
at org.scalajs.core.tools.optimizer.Linker.link(Linker.scala:72)
at org.scalajs.core.tools.optimizer.ScalaJSOptimizer$$anonfun$1.apply(ScalaJSOptimizer.scala:119)
at org.scalajs.core.tools.optimizer.ScalaJSOptimizer$$anonfun$1.apply(ScalaJSOptimizer.scala:119)
at org.scalajs.core.tools.optimizer.package$.logTime(package.scala:19)
at org.scalajs.core.tools.optimizer.ScalaJSOptimizer.optimizeIR(ScalaJSOptimizer.scala:118)
at org.scalajs.core.tools.optimizer.ScalaJSOptimizer.optimizeIR(ScalaJSOptimizer.scala:99)
at org.scalajs.core.tools.optimizer.ScalaJSOptimizer$$anonfun$optimizeCP$1.apply$mcV$sp(ScalaJSOptimizer.scala:76)
at org.scalajs.core.tools.io.CacheUtils$.cached(CacheUtils.scala:41)
at org.scalajs.core.tools.optimizer.ScalaJSOptimizer.optimizeCP(ScalaJSOptimizer.scala:74)
at scala.tools.nsc.MainGenericRunner.fastOptimize(MainGenericRunner.scala:93)
at scala.tools.nsc.MainGenericRunner.process(MainGenericRunner.scala:77)
at scala.tools.nsc.MainGenericRunner$.main(MainGenericRunner.scala:200)
at scala.tools.nsc.MainGenericRunner.main(MainGenericRunner.scala)
@sjrd could you please ping Fabien that he fixes the Jenkins emails? These test that fail and we don't notice really shouldn't be that way...
The update they did when I was still at LAMP somehow broke the emails.
This failure stopped last night. A git bisect showed that the first "bad" commit (the commit that actually fixed the issue) was 7b79765633e54608872ae1aa4772fde8a8e5c2b7
7b79765633e54608872ae1aa4772fde8a8e5c2b7 is the first bad commit
commit 7b79765633e54608872ae1aa4772fde8a8e5c2b7
Author: Sébastien Doeraene <sjrdoeraene@gmail.com>
Date: Sun Oct 11 16:21:58 2015 +0200
Remove the dependency on j.u.Formatter from Range.description.
This cuts loose j.u.Formatter entirely from the Reversi example.
:040000 040000 fcad56c8ded8ad9056d060dcb47a87eb877d5e27 209f14c806d494d5545238e91e15dc2a402e9207 M ci
:040000 040000 c25ee2ce9cbddb2d849dfc248b2b2634aa2d0043 d1c179f3e2c3afeba733eb816050e3f68d641e0c M scalalib
Not sure whether to close this or leave it open for further investigation. Because obviously I have no idea why it started to fail nor why that commit fixed the issue.
In any case we should close #2029
After careful consideration, I we should blacklist this test anyway. The fact that it passes is an accident. <:< is not supposed to be supported on ClassTags, as it uses getInterfaces(). It happens to pass, I guess, because the set of things reachable from this tiny codebase allows the anlyzer to dce away the portion that needs getInterfaces(). But depending on the rest of the codebase, this might not link.
| gharchive/issue | 2015-11-15T10:55:37 | 2025-04-01T06:45:43.788217 | {
"authors": [
"gzm0",
"sjrd"
],
"repo": "scala-js/scala-js",
"url": "https://github.com/scala-js/scala-js/issues/2021",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
207980753 | fix argonaut and upickle-pprint
just the usual sorts of issues with scalaVersion and
unmanagedSourceDirectories
tested locally
| gharchive/pull-request | 2017-02-16T01:02:29 | 2025-04-01T06:45:43.832888 | {
"authors": [
"SethTisue"
],
"repo": "scala/community-builds",
"url": "https://github.com/scala/community-builds/pull/479",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
176394196 | Update scala to 2.12.0-RC1
Addresses #14
if you're able to publish for RC1, please add it to https://github.com/scala/make-release-notes/blob/2.12.x/projects-2.12.md
@SethTisue surely!
| gharchive/pull-request | 2016-09-12T14:44:03 | 2025-04-01T06:45:43.847441 | {
"authors": [
"SethTisue",
"danslapman"
],
"repo": "scala/scala-continuations",
"url": "https://github.com/scala/scala-continuations/pull/19",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
297357052 | Consider better separation of public CI validation and release infrastructure
While we're beefing up the security of our Jenkins install, it occurred to me that we should probably run two Jenkins installs: one that does all the stuff that only needs the "public" nodes, and another that does the releases and needs the "publish" node, which has some secrets to protect. We could place that private Jenkins behind a VPN, rather than expose it directly to the internet.
This would protect us better from vulnerabilities in Jenkins (or our configuration of it).
most of our publishing is now Travis-CI based
the remaining exception is publishing of PR validation snapshots. we're taking that up at https://github.com/scala/scala-dev/issues/507
we're also using jenkins-windows-worker-publish for Windows testing, moving to AppVeyor is https://github.com/scala/scala-dev/issues/508
once those tickets are completed, we could decommission the publishing nodes and closed down this ticket.
this remains blocked on figuring out how to securely allow pull requests from forks to have publishing secrets
but in any case any new solution will not be Jenkins-based
| gharchive/issue | 2018-02-15T07:58:31 | 2025-04-01T06:45:43.850368 | {
"authors": [
"SethTisue",
"retronym"
],
"repo": "scala/scala-jenkins-infra",
"url": "https://github.com/scala/scala-jenkins-infra/issues/243",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2221224083 | Local inference reports cycle, suspiciously
Compiler version
3.4.1
Minimized code
class Test {
def local: Unit = {
val timeout = 5.seconds
implicit val askTimeout = Timeout(timeout)
}
}
with supporting code via Akka
import java.util.concurrent.TimeUnit
import java.time.{Duration => JDuration}
import scala.concurrent.duration._
import language.implicitConversions
case class Timeout(duration: FiniteDuration) {
/**
* Construct a Timeout from the given time unit and factor.
*/
def this(length: Long, unit: TimeUnit) = this(Duration(length, unit))
}
/**
* A Timeout is a wrapper on top of Duration to be more specific about what the duration means.
*/
object Timeout {
/**
* A timeout with zero duration, will cause most requests to always timeout.
*/
val zero: Timeout = new Timeout(Duration.Zero)
/**
* Construct a Timeout from the given time unit and factor.
*/
def apply(length: Long, unit: TimeUnit): Timeout = new Timeout(length, unit)
/**
* Create a Timeout from java.time.Duration.
*/
def create(duration: java.time.Duration): Timeout = {
import JavaDurationConverters._
new Timeout(duration.asScala)
}
implicit def durationToTimeout(duration: FiniteDuration): Timeout = new Timeout(duration)
}
object JavaDurationConverters {
def asFiniteDuration(duration: JDuration): FiniteDuration = duration.asScala
final implicit class JavaDurationOps(val self: JDuration) extends AnyVal {
def asScala: FiniteDuration = Duration.fromNanos(self.toNanos)
}
final implicit class ScalaDurationOps(val self: Duration) extends AnyVal {
def asJava: JDuration = JDuration.ofNanos(self.toNanos)
}
}
Edit: the minimization is trivial
class C {
import C.*
def local = {
val x = 42.z
implicit val y = f(x)
}
def f(s: String): String = s
}
object C {
implicit class Z(val v: Int) {
def z: String = "z"
}
}
Output
-- [E045] Cyclic Error: timeout.scala:60:38 ----------------------------------------------------------------------------
60 | implicit val askTimeout = Timeout(timeout)
| ^
| Recursive value timeout needs type
|
| The error occurred while trying to compute the signature of value timeout
| which required to compute the signature of value askTimeout
| which required to compute the signature of value timeout
Expectation
Although there are many implicits in the environment, local inference should just work, especially since the implicit value is defined after the first "site of inference".
Observed at https://github.com/akka/alpakka-kafka/pull/1727
Maybe it's just complaining that the implicit is in local scope while inferring the first definition. Edit: see minimal example.
Since we will remove old style implicits I don't think we should spend cycles tweaking them.
Agreed.
Seems like we will not be working on this, so it can be closed.
Thanks again for the explanation. I'm sure it's tiresome to field settled issues. For Scala 2, Seth added "questions are not bugs" to the template to make everyone think twice before posting.
| gharchive/issue | 2024-04-02T19:00:29 | 2025-04-01T06:45:43.887443 | {
"authors": [
"Gedochao",
"som-snytt"
],
"repo": "scala/scala3",
"url": "https://github.com/scala/scala3/issues/20074",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1575738096 | Be even more tolerant when typing polymorphic Apply(New, ...).
I was not able to reproduce the issue in a standalone test, but apparently it is possible to have both a New with an AppliedType (hence, monomorphic) wrapped in a TypeApply. This would attempt to reapply type parameters to an already monomorphic type, which crashed.
We are now a bit more tolerant, and apply type parameter lists to the type of the New as long as we have a polymorphic type. We stop applying further type parameter lists when the type is fully monomorphic.
The first commit contains more tests that were an attempt at reproducing the above issue. They don't, but the tests are valuable anyway.
it'd be interesting to document how this crashed originally?
Things like this:
tastyquery.Exceptions$InvalidProgramStructureException: found type constructor
AppliedType(
TypeRef(PackageRef(langoustine.lsp), symbol[class lsp.StringEnum]),
List(TypeRef(ThisType(TypeRef(PackageRef(langoustine.lsp), enumerations)), MonikerKind))
)
without type params in AppliedType
| gharchive/pull-request | 2023-02-08T09:02:22 | 2025-04-01T06:45:43.911404 | {
"authors": [
"sjrd"
],
"repo": "scalacenter/tasty-query",
"url": "https://github.com/scalacenter/tasty-query/pull/249",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
298044746 | Service issue? "No suitable compiler available"
Currently, any edits to an existing fiddle cause "No suitable compiler available"
Example: https://scalafiddle.io/sf/9Xs6tIU/6 (compare to the previous https://scalafiddle.io/sf/9Xs6tIU/5)
Thx for noticing. The compiler instances had died, now restarted!
thank you!
On Sun, Feb 18, 2018 at 1:54 PM, Otto Chrons notifications@github.com
wrote:
Thx for noticing. The compiler instances had died, now restarted!
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
https://github.com/scalafiddle/scalafiddle-editor/issues/76#issuecomment-366538023,
or mute the thread
https://github.com/notifications/unsubscribe-auth/AA37jnEgRPMeqWmDyhS-7q7FyeO1_SbEks5tWHHJgaJpZM4SJgKB
.
--
Brandon Barker
brandon.barker@gmail.com
| gharchive/issue | 2018-02-17T23:18:56 | 2025-04-01T06:45:43.916085 | {
"authors": [
"bbarker",
"ochrons"
],
"repo": "scalafiddle/scalafiddle-editor",
"url": "https://github.com/scalafiddle/scalafiddle-editor/issues/76",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1831570637 | SemanticDB: allow silencing info/warning/error
Capture them only in semanticdb and not report further.
The reason why this is useful is because scalac's -Xreporter option is not yet supported in scala3 -- and it's also ignored by sbt, which uses the compilerReporter setting instead (which isn't readily documented).
I am not sure about the exact use case here. Why would we want to silence messages coming from the compiler? Shouldn't we silence them differently via -Wconf?
Or the case is that because we are using our own reporter users are getting messages which should not pop up?
honestly, I'm not sure if there's a different way. i was simply a bit unnerved by all the warnings coming from the compiler when building the integration examples in this build.
and it turns out, that those warnings are needed by semanticdb itself (it intercepts them) so we can't silence them completely but if, other than semanticdb, we wouldn't have wanted to see them, it's not clear how to silence them after.
thoughts?
honestly, I'm not sure if there's a different way. i was simply a bit unnerved by all the warnings coming from the compiler when building the integration examples in this build.
and it turns out, that those warnings are needed by semanticdb itself (it intercepts them) so we can't silence them completely but if, other than semanticdb, we wouldn't have wanted to see them, it's not clear how to silence them after.
thoughts?
This would be useful if we decided to always add things like unused warnings and hide them from the user, but still be able to run scalafix rules etc. :thinking: But honestly I don't see a lot of scenarios beside the one in our tests that we wouldn't want the warning to show up for the user. We want them to know something can be fixed.
Maybe it's fine to just filter out those warnings in the IDE for your use case?
| gharchive/pull-request | 2023-08-01T15:49:37 | 2025-04-01T06:45:43.925434 | {
"authors": [
"kitbellew",
"tgodzik"
],
"repo": "scalameta/scalameta",
"url": "https://github.com/scalameta/scalameta/pull/3313",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
188290161 | [ObjectCopy][XMLResponse] bad Date format in LastModified field
Testing hadoop against the S3 server, the java code fails when attempting to parse a Date. It looks like it's attemtping to parse it as Iso8601 when it's not formatted this way.
Here is the S3 Logs (trace level) from s3server:
{"name":"S3","bucketName":"mybucket","objectKey":"derby93","accessKey":"accessKey1","xml":"<?xml version=\"1.0\" encoding=\"UTF-8\"?><CopyObjectResult><LastModified>Wed, 09 Nov 2016 16:16:12 GMT</LastModified><ETag>"d88d58e233cfba321424c8452ba2a314"</ETag></CopyObjectResult>","time":1478708172757,"req_id":"162a10e21335c36b3f42","level":"trace","message":"xml response","hostname":"c314ca580bc8","pid":57}
Here is the output from a s3cmd ObjectCopy
DEBUG: Response: {'status': 200, 'headers': {'content-length': '234', 'via': '1.1 ubnt1 (squid/3.4.8)', 'x-cache': 'MISS from ubnt1', 'x-amz-id-2': 'NGblUtou6K8lvXGvmQW4tcZEsnJxRm046qLBhMJTR8cpHRrMaLGcFheFWHSrAdO0wK81auZHH4E=', 'x-cache-lookup': 'MISS from ubnt1:0', 'server': 'AmazonS3', 'connection': 'keep-alive', 'x-amz-request-id': '568CD32D9EF08BD1', 'date': 'Wed, 09 Nov 2016 16:24:52 GMT', 'content-type': 'application/xml'}, 'reason': 'OK', 'data': '<?xml version="1.0" encoding="UTF-8"?>\n<CopyObjectResult xmlns="http://s3.amazonaws.com/doc/2006-03-01/"><LastModified>2016-11-09T16:24:52.000Z</LastModified><ETag>"fac2a668031954664ec5febf2f301c27"</ETag></CopyObjectResult>'}
2016-11-09T16:24:52.000Z should be 2016-11-09T16:24:52Z refering to https://en.wikipedia.org/wiki/ISO_8601
No matter, the thing is:
We're returning Wed, 09 Nov 2016 16:16:12 GMT
S3 actually returns 2016-11-09T16:16:12.000Z
oh right ^^ don't see the second output is from Amazon :)
@LaurenSpiegel ?
Here is the fix https://github.com/scality/S3/commit/4bf8d5add6d88089893e9a1be8cdfcc35d03ef12
I will add an s3curl test because the node sdk transforms the format (so we didn't see the difference).
Thanks, I'll notify the original reporter of the ongoing fix. I guess this should be part of the 6.2.4 (if there is one) ?
Looks good. Thanks !
fixed by: https://github.com/scality/S3/commit/351619eec868799ef285e95accc1fc1a2f1bdff8
| gharchive/issue | 2016-11-09T16:28:44 | 2025-04-01T06:45:43.995608 | {
"authors": [
"DavidPineauScality",
"LaurenSpiegel",
"alexandre-merle",
"pgueant",
"vrancurel"
],
"repo": "scality/S3",
"url": "https://github.com/scality/S3/issues/372",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2345106405 | Issue-Release 0.2.2.
Automated Review Comments
Issues in package.json:
(3, 'Hardcoded string found')
(4, 'Hardcoded string found')
(5, 'Hardcoded string found')
(6, 'Line exceeds 80 characters')
(6, 'Hardcoded string found')
(7, 'Hardcoded string found')
(8, 'Hardcoded string found')
I got it.
| gharchive/issue | 2024-06-11T00:55:39 | 2025-04-01T06:45:44.026478 | {
"authors": [
"satokihanada",
"scdnj"
],
"repo": "scdnj/micro-signer-eth",
"url": "https://github.com/scdnj/micro-signer-eth/issues/75",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
302785602 | LICENSE Copyright Heydon Pickering
Can this be changed to Mozilla, or to a dual copyright?
I would say Cupper is licensed under MIT but, the content built on top of it can either be Creative Commons or MPL-2
I see a few options:
Move LICENSE to LICENSE-cupper. Create a LICENSE that says "This part is covered by this license, this part by this license", and add the license statement for the content. I think GitHub will claim the license isn't an open source license, since it will just look at LICENSE
Amend the statement in license to state that some content is covered by the MIT License, and then add the text of a second license below. Manually merge the file as upstream changes. This is the route suggested by https://softwareengineering.stackexchange.com/questions/177768/how-do-i-correctly-sub-license-a-library-that-is-under-the-mit-license.
Accept the MIT License for our stuff as well. Amend the copyright notice to add a second copyright, "Copyright (c) 2018 Mozilla", and manually merge just that line in the future.
We may want to include something in the README, CONTRIBUTING explaining the situation, which will probably be confusing to outside contributors or later staff. An upstream issue might also be needed.
@jwhitlock I have made some updates. Not sure this covers all of our basis, but let me know your thoughts. Thanks!
I think this works for licensing, thanks!
| gharchive/issue | 2018-03-06T17:05:22 | 2025-04-01T06:45:44.043603 | {
"authors": [
"jwhitlock",
"schalkneethling"
],
"repo": "schalkneethling/mdn-fiori",
"url": "https://github.com/schalkneethling/mdn-fiori/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.