id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
98973380 | About how to security the view with no blocking.
What about this:
controller
val dataFuture: Future[Data] = ???
val subjectWithPermissionsFromCacheOrDBFuture: Future[CaseSubject] = ???
for {
data <- data
subject <- subjectWithPermissionFromCacheOrDBFuture
} yield {
Ok(view(data)(SubjectWrap(subject)))
}
view.scala.html
@(data: Data)(wrap: SubjectWrap)
@wrap.require("superAdmin") {
@data.name
}
I mean get the subject informations with future in the controller first and render the view with no db connection.
Oh, when I review this issue, I think I need to say more specific.
model.scala
case class Subject(name: String, permissions: List[Permission]) {
def require(permission: String)(content: => T) =
if(permissions.exists(_.getName == permission)) {
Option(content)
} else {
None
}
}
controller.scala
val dataFuture = Future[Data]
val subjectFuture = Future[Subject]
for {
data <- dataFuture
subject <- subjectFuture
} yield {
Ok(views.html.index(data)(subject))
}
view.html
...
@subject.require("admin"){
@data.name
}
...
That's an interesting idea. However, if you fetch the Subject (or DynamicResourceHandler, or whatever) in a non-blocking controller call, you can make it immediately accessible via the DeadboltHandler and set the blocking duration to -1.
| gharchive/issue | 2015-08-04T13:43:27 | 2025-04-01T06:45:44.047058 | {
"authors": [
"djx314",
"schaloner"
],
"repo": "schaloner/deadbolt-2-scala",
"url": "https://github.com/schaloner/deadbolt-2-scala/issues/14",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
102640377 | Remove ExplicitUnicodeString and add NonUnicodeString
As I explained here, code generatiion for SQL Server should default to Unicode string literals, so I removed ExplicitUnicodeString and added NonUnicodeString instead.
This change affected SQL Server CE too, which I think defaults to Unicode, so 'caș' and N'caș' are equivalent there, but there shouldn't be any harm in using N'.
Hana was also using ExplicitUnicodeString, so I went ahead and changed it too. @FabioNascimento, is it ok if the Hana generator dafaults to Unicode string literals?
I removed the ExplicitUnicodeString class. This means people who were using it will have to change their code. I can add it back and deprecate it if anyone will miss it, but I think migration code is cleaner when you use normal .NET strings.
Fixes #636.
Hello @vgrigoriu ,
Sorry for the late answer, but I just see now. I test your changes in HANA db and everything work fine...
Fabio
Wow, that's a blast from the past 😃
Who is taking care of this repo now? @fubar-coder ? Any chance of merging this in if I solve the conflicts?
This seems to be a breaking change for the end user due to the removal of ExplicitUnicodeString. Can you please keep ExplicitUnicodeString as obsolete (with explanation)? Can it cause problems when we switch to unicode by default?
All right, I'll put ExplicitUnicodeString back in as obsolete.
I explained [here|https://github.com/schambers/fluentmigrator/issues/636#issuecomment-129391177] what I think are the consequences of switching to Unicode.
Also, please don't forget to handle ExplicitUnicodeString instances 😉 .
@schambers, I guess, that this change is OK for a minor library update when the only user visible change is a deprecated class?
Here you are.
Thank you
| gharchive/pull-request | 2015-08-23T17:45:53 | 2025-04-01T06:45:44.053427 | {
"authors": [
"FabioNascimento",
"fubar-coder",
"vgrigoriu"
],
"repo": "schambers/fluentmigrator",
"url": "https://github.com/schambers/fluentmigrator/pull/642",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1404978108 | забанен незнаю за что
почему меня забанили? моды стоят только с вашего сервера больше не где не играл!!!
This is the source code for the chat mod. Bans are determined by the servers that use it, not the mod itself.
а как тогда играть если банит?
| gharchive/issue | 2022-10-11T17:17:04 | 2025-04-01T06:45:44.055402 | {
"authors": [
"dimooon08101997",
"schana"
],
"repo": "schana/dayz-mod-global-chat",
"url": "https://github.com/schana/dayz-mod-global-chat/issues/10",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
384618412 | best practice to make use of pageviewcontroller in layout
I'd like to implement a tabbar controller positioned at top, so I think of using UIPageViewController, but I did not find the document about this? Is it possible to use the controller? any example? Thanks in advance
thanks for clarification, closing this issue
| gharchive/issue | 2018-11-27T04:33:51 | 2025-04-01T06:45:44.122863 | {
"authors": [
"wpcfan"
],
"repo": "schibsted/layout",
"url": "https://github.com/schibsted/layout/issues/163",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
132000641 | Use websockets for updating client from server (instead of polling) [need help]
Basically I think need to keep track of the connections, as in conn, err := wsupgrader.Upgrade(w, r, nil) in websockets.go...
However, this is only really needed if people are going to use this like a chatroom - i.e. realtime viewing of the same document.
| gharchive/issue | 2016-02-07T19:20:54 | 2025-04-01T06:45:44.312744 | {
"authors": [
"schollz"
],
"repo": "schollz/AwwKoala",
"url": "https://github.com/schollz/AwwKoala/issues/5",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1637966907 | Maintenance of installation module
Module to maintain
Website: https://school-brainhack.github.io/modules/installation/
Code: https://github.com/school-brainhack/school-brainhack.github.io/tree/main/content/en/modules/installation
Work to be done
Please ensure:
[ ] links, resources and exercises reflect the state-of-the-art
[ ] videos are of good quality and reflect the materials.
If any update / re-write of the module is needed, feel free to suggest alternative material in a comment!
If you think major changes are needed, we suggest you open an issue for new modules.
Hi, this is Yu-Shiang Su from Taiwan hub.
It seems that the installation instructions for Windows users are outdated. Microsoft has updated their instructions to install WSL, and the steps are much easier now. Importantly, Docker Toolbox was deprecated, and students should install Docker Desktop which now have better support for WSL.
Dear @yushiangsu thank you for letting us know, that's amazing you caught up this. I wonder would you like to help us with updating the instructions based on new improvements please?
All you need to do is to fork the repo and make the changes and PR, then we will review the changes accordingly. Please have a look at the instructions for the module maintenance here and do not hesitate to reach us with any questions you might have.
Yes, I glad to help. I will prepare a PR later this week.
Closing this since @yushiangsu completed a PR and I have merged
Re-opening since we decided to keep these open until modules have been 'piloted' before school starts.
It seems that new macOS versions (after Monterey 12.3) dropped support for python 2.x. Details here.
This then results in an error when checking installation using the following command from the Installation module:
https://github.com/school-brainhack/school-brainhack.github.io/blob/57f93c5f21bd350a39c7cf3e895029f4033dfa81/content/en/modules/installation/index.md?plain=1#L85
Terminal output
valosek@macbook-pro:~$ bash <( curl -s https://raw.githubusercontent.com/brainhackorg/school/master/content/en/modules/installation/nds_check_install.sh )
Missing software program: python. Check installation instructions
The proper command here, in this case, should be hash python3 instead of hash python:
valosek@macbook-pro:~$ hash python
hash: no such command: python
valosek@macbook-pro:~$ hash python3
Thanks for that @valosekj, please do let us know if you spot any other bugs!
| gharchive/issue | 2023-03-23T17:25:41 | 2025-04-01T06:45:44.339490 | {
"authors": [
"clarkenj",
"complexbrains",
"valosekj",
"yushiangsu"
],
"repo": "school-brainhack/school-brainhack.github.io",
"url": "https://github.com/school-brainhack/school-brainhack.github.io/issues/175",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
646942953 | the bit unpacking of the raw Bayer data is wrong for V1 and V2 camera
Line 179 and following of core.py does the bit unpacking for V1 and V2 camera:
data = data.astype(np.uint16) << 2
for byte in range(4):
data[:, byte::5] |= ((data[:, 4::5] >> ((byte+1) * 2)) & 0b11)
data = np.delete(data, np.s_[4::5], 1)
This does an unpacking for the scheme:
AAAAAAAA BBBBBBBB CCCCCCCC DDDDDDDD DDCCBBAA
But the documentation in https://picamera.readthedocs.io/en/release-1.12/recipes2.html#raw-bayer-data-captures states this:
AAAAAAAA BBBBBBBB CCCCCCCC DDDDDDDD AABBCCDD
and the code:
data = data.astype(np.uint16) << 2
for byte in range(4):
data[:, byte::5] |= ((data[:, 4::5] >> ((4 - byte) * 2)) & 0b11)
data = np.delete(data, np.s_[4::5], 1)
Good catch, not sure how I let that one slip by. Will implement in next release.
| gharchive/issue | 2020-06-28T15:10:30 | 2025-04-01T06:45:44.342640 | {
"authors": [
"schoolpost",
"scriptorron"
],
"repo": "schoolpost/PyDNG",
"url": "https://github.com/schoolpost/PyDNG/issues/14",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2518058503 | Question about examples
In some of the examples you just run (eg. here)
compute.run(...)
and in some you do all the steps manually (eg. here)
try compute.task { task in
try task { dispatch in
try dispatch(....)
}
}
Is there any reason to prefer the second version before the first?
They're both equivalent. run is just more convenient at the expensive of less flexibility..
compute.run(A)
compute.run(B)
compute.run(C)
compute.task { task in
task { dispatch in
dispatch(A)
dispatch(B)
dispatch(C)
}
}
The first example will wait for the first shader to complete before running the second shader… and so on.
The second example submits them all to Metal in the same pass and allows metal to optimise if needed. It only waits for completion once.
I am thinking of deprecating run possibly.
Thanks. I see... so if I want to pass a texture for example through multiple shaders without hitting the CPU in between I would use the second one. I think the variant of run that is used in the example in the readme is actually deprecated.
| gharchive/issue | 2024-09-10T22:26:09 | 2025-04-01T06:45:44.355103 | {
"authors": [
"potmo",
"schwa"
],
"repo": "schwa/Compute",
"url": "https://github.com/schwa/Compute/issues/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
222827773 | Clean brew install error with punic init - AttributeError: 'set' object has no attribute 'keys'
Call stack:
Traceback (most recent call last):
File "/usr/local/bin/punic", line 11, in <module>
load_entry_point('punic==0.2.9', 'console_scripts', 'punic')()
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/punic_cli.py", line 397, in main
punic_cli()
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 722, in __call__
return self.main(*args, **kwargs)
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 697, in main
rv = self.invoke(ctx)
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 1066, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 895, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 535, in invoke
return callback(*args, **kwargs)
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/decorators.py", line 17, in new_func
return f(get_current_context(), *args, **kwargs)
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/punic_cli.py", line 341, in init
config_init(**kwargs)
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/config_init.py", line 74, in config_init
xcode_version = _prompt("Xcode Version", _xcode_versions())
File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/config_init.py", line 41, in _xcode_versions
return sorted([six.text_type(version) for version in Xcode._all_xcodes.keys()])
AttributeError: 'set' object has no attribute 'keys'
I had punic installed a while back, not sure if that makes a difference, so I uninstalled punic first, ie. pip uninstall punic.
Did a fresh install after updating/upgrading brew. Running punic init yields the following callstack. Is it a problem with my Python version (not versed in Python)? It just seems to be a problem with the object type of_all_xcodes.
Thanks!
How did you install? Are you running from a non-release branch?
I'm doing a straight brew install punic, which installed 0.2.9, so I don't believe so.
I'm using Xcode 8.3.1 only (ie. no beta/old versions alongside).
Ah ok - that's bad. Will look into this.
OK fixed and the home-brew tap is updated as well. Can you verify with a brew update and brew upgrade punic?
Seems to have done the trick. Thanks!
| gharchive/issue | 2017-04-19T18:25:24 | 2025-04-01T06:45:44.359559 | {
"authors": [
"mmmilo",
"schwa"
],
"repo": "schwa/punic",
"url": "https://github.com/schwa/punic/issues/67",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
309528578 | Add ability to cancel execution (I.e. stop button)
...
aca33d494a8775737a99f8dba393391f80a0a492
| gharchive/issue | 2018-03-28T20:56:15 | 2025-04-01T06:45:44.360531 | {
"authors": [
"Lucaszw"
],
"repo": "sci-bots/microdrop-3",
"url": "https://github.com/sci-bots/microdrop-3/issues/57",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1722976753 | [Experiement] virtual simple index.
Having to publish all wheels in the same repository is complicated from the size of the storage and token regenration. Any changes to the index URL needs update to all the CI.
So the idea is to have a virtual index that just point to the original one(s) that we can manipulate the way we want:
Not sure this is the best place, but that may affect spec 4 and 5
it does not support the top /simple/ that list all the packages, nor the data-python-requires.
It appears also the Pypi now expose wheel metadata w/o downloading them, but i'm unsure how.
| gharchive/pull-request | 2023-05-24T00:30:18 | 2025-04-01T06:45:44.386548 | {
"authors": [
"Carreau"
],
"repo": "scientific-python/specs",
"url": "https://github.com/scientific-python/specs/pull/194",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1979776241 | ERROR: Could not build wheels for ninja
The problem:
I'm trying to install ninja in a Docker container on an arm32v7/debian image which doesn't have a corresponding wheel on PyPi and I am getting an error during the build process.
Currently I'm using python 3.9 and pip 23.3.1
I tried different versions of either of them and got the same result
Dockerfile:
FROM arm32v7/debian:bullseye
ENV DEBIAN_FRONTEND=noninteractive
RUN \
apt-get update --yes && \
apt-get install --no-install-recommends --yes \
build-essential cmake bison flex re2c binutils \
fakeroot file ca-certificates \
subversion git \
binutils-arm-linux-gnueabihf \
pkg-config libtool automake autoconf \
python python3-pip zlib1g-dev libxrandr-dev \
libdrm-dev libxdamage-dev libxcb-glx0-dev libxcb-shm0-dev \
libx11-xcb-dev libxcb-dri2-0-dev libxcb-dri3-dev \
libxcb-present-dev libxshmfence-dev libxxf86vm-dev
RUN python3 -m pip install --upgrade pip
RUN python3 -m pip install --upgrade setuptools
RUN python3 -m pip install --upgrade wheel
RUN python3 -m pip install --upgrade scikit-build
RUN python3 -m pip install ninja --verbose
Console output:
[+] Building 439.2s (11/11) FINISHED docker:desktop-linux
=> [internal] load build definition from Dockerfile 0.0s
=> => transferring dockerfile: 867B 0.0s
=> [internal] load .dockerignore 0.0s
=> => transferring context: 2B 0.0s
=> [internal] load metadata for docker.io/arm32v7/debian:bullseye 1.6s
=> [auth] arm32v7/debian:pull token for registry-1.docker.io 0.0s
=> [1/7] FROM docker.io/arm32v7/debian:bullseye@sha256:bede2623dae269454c5b6dd4af15a10810a5f4ef75963d4eb6531628f98bd633 0.0s
=> CACHED [2/7] RUN apt-get update --yes && apt-get install --no-install-recommends --yes build-essential cmake bison flex re2c binutils fakeroot file ca-certificates subversion git binutils-arm-linux-gnueabihf pkg 0.0s
=> CACHED [3/7] RUN python3 -m pip install --upgrade pip 0.0s
=> CACHED [4/7] RUN python3 -m pip install --upgrade setuptools 0.0s
=> CACHED [5/7] RUN python3 -m pip install --upgrade wheel 0.0s
=> CACHED [6/7] RUN python3 -m pip install --upgrade scikit-build 0.0s
=> ERROR [7/7] RUN python3 -m pip install ninja --verbose 437.6s
------
> [7/7] RUN python3 -m pip install ninja --verbose:
1.604 Using pip 23.3.1 from /usr/local/lib/python3.9/dist-packages/pip (python 3.9)
2.285 Collecting ninja
2.584 Downloading ninja-1.11.1.1.tar.gz (132 kB)
2.897 ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 132.4/132.4 kB 458.9 kB/s eta 0:00:00
2.977 Installing build dependencies: started
2.978 Running command pip subprocess to install build dependencies
5.822 Collecting setuptools>=42
5.831 Using cached setuptools-68.2.2-py3-none-any.whl.metadata (6.3 kB)
5.975 Collecting scikit-build
5.983 Using cached scikit_build-0.17.6-py3-none-any.whl.metadata (14 kB)
6.333 Collecting setuptools-scm[toml]
6.543 Downloading setuptools_scm-8.0.4-py3-none-any.whl.metadata (6.4 kB)
6.973 Collecting packaging>=20 (from setuptools-scm[toml])
6.981 Using cached packaging-23.2-py3-none-any.whl.metadata (3.2 kB)
7.119 Collecting typing-extensions (from setuptools-scm[toml])
7.185 Downloading typing_extensions-4.8.0-py3-none-any.whl.metadata (3.0 kB)
7.318 Collecting tomli>=1 (from setuptools-scm[toml])
7.324 Using cached tomli-2.0.1-py3-none-any.whl (12 kB)
7.583 Collecting distro (from scikit-build)
7.590 Using cached distro-1.8.0-py3-none-any.whl (20 kB)
7.806 Collecting wheel>=0.32.0 (from scikit-build)
7.813 Using cached wheel-0.41.3-py3-none-any.whl.metadata (2.2 kB)
7.880 Using cached setuptools-68.2.2-py3-none-any.whl (807 kB)
7.893 Using cached scikit_build-0.17.6-py3-none-any.whl (84 kB)
7.900 Using cached packaging-23.2-py3-none-any.whl (53 kB)
7.908 Using cached wheel-0.41.3-py3-none-any.whl (65 kB)
7.958 Downloading setuptools_scm-8.0.4-py3-none-any.whl (42 kB)
8.026 ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.1/42.1 kB 686.9 kB/s eta 0:00:00
8.084 Downloading typing_extensions-4.8.0-py3-none-any.whl (31 kB)
8.813 Installing collected packages: wheel, typing-extensions, tomli, setuptools, packaging, distro, setuptools-scm, scikit-build
11.24 Successfully installed distro-1.8.0 packaging-23.2 scikit-build-0.17.6 setuptools-68.2.2 setuptools-scm-8.0.4 tomli-2.0.1 typing-extensions-4.8.0 wheel-0.41.3
11.24 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv
11.36 Installing build dependencies: finished with status 'done'
11.37 Getting requirements to build wheel: started
11.37 Running command Getting requirements to build wheel
12.27 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:308: UserWarning: git archive did not support describe output
12.27 warnings.warn("git archive did not support describe output")
12.27 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:327: UserWarning: unprocessed git archival found (no export subst applied)
12.28 warnings.warn("unprocessed git archival found (no export subst applied)")
12.41 running egg_info
12.41 writing src/ninja.egg-info/PKG-INFO
12.42 writing dependency_links to src/ninja.egg-info/dependency_links.txt
12.42 writing entry points to src/ninja.egg-info/entry_points.txt
12.43 writing requirements to src/ninja.egg-info/requires.txt
12.43 writing top-level names to src/ninja.egg-info/top_level.txt
12.49 ERROR setuptools_scm._file_finders.git listing git files failed - pretending there aren't any
12.49 reading manifest file 'src/ninja.egg-info/SOURCES.txt'
12.50 reading manifest template 'MANIFEST.in'
12.50 warning: no previously-included files matching '*' found under directory '_skbuild'
12.50 warning: no previously-included files matching '*' found under directory 'Ninja-src'
12.50 warning: no files found matching 'versioneer.py'
12.50 adding license file 'LICENSE_Apache_20'
12.50 adding license file 'AUTHORS.rst'
12.51 writing manifest file 'src/ninja.egg-info/SOURCES.txt'
12.56 Getting requirements to build wheel: finished with status 'done'
12.57 Preparing metadata (pyproject.toml): started
12.57 Running command Preparing metadata (pyproject.toml)
13.48 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:308: UserWarning: git archive did not support describe output
13.48 warnings.warn("git archive did not support describe output")
13.48 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:327: UserWarning: unprocessed git archival found (no export subst applied)
13.48 warnings.warn("unprocessed git archival found (no export subst applied)")
13.59 running dist_info
13.60 creating /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info
13.60 writing /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/PKG-INFO
13.61 writing dependency_links to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/dependency_links.txt
13.61 writing entry points to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/entry_points.txt
13.62 writing requirements to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/requires.txt
13.62 writing top-level names to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/top_level.txt
13.62 writing manifest file '/tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/SOURCES.txt'
13.67 ERROR setuptools_scm._file_finders.git listing git files failed - pretending there aren't any
13.68 reading manifest file '/tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/SOURCES.txt'
13.68 reading manifest template 'MANIFEST.in'
13.68 warning: no previously-included files matching '*' found under directory '_skbuild'
13.68 warning: no previously-included files matching '*' found under directory 'Ninja-src'
13.68 warning: no files found matching 'versioneer.py'
13.68 adding license file 'LICENSE_Apache_20'
13.69 adding license file 'AUTHORS.rst'
13.69 writing manifest file '/tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/SOURCES.txt'
13.69 creating '/tmp/pip-modern-metadata-j639jt5h/ninja-1.11.1.1.dist-info'
13.75 Preparing metadata (pyproject.toml): finished with status 'done'
13.79 Building wheels for collected packages: ninja
13.79 Building wheel for ninja (pyproject.toml): started
13.80 Running command Building wheel for ninja (pyproject.toml)
14.69 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:308: UserWarning: git archive did not support describe output
14.69 warnings.warn("git archive did not support describe output")
14.69 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:327: UserWarning: unprocessed git archival found (no export subst applied)
14.69 warnings.warn("unprocessed git archival found (no export subst applied)")
14.84
14.84
14.84 --------------------------------------------------------------------------------
14.85 -- Trying 'Ninja' generator
14.85 --------------------------------
14.85 ---------------------------
14.85 ----------------------
14.85 -----------------
14.85 ------------
14.85 -------
14.85 --
14.93 Not searching for unused variables given on the command line.
15.02 CMake Error: CMake was unable to find a build program corresponding to "Ninja". CMAKE_MAKE_PROGRAM is not set. You probably need to select a different build tool.
15.02 -- Configuring incomplete, errors occurred!
15.02 See also "/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_cmake_test_compile/build/CMakeFiles/CMakeOutput.log".
15.04 --
15.04 -------
15.04 ------------
15.04 -----------------
15.04 ----------------------
15.04 ---------------------------
15.04 --------------------------------
15.04 -- Trying 'Ninja' generator - failure
15.04 --------------------------------------------------------------------------------
15.04
15.04
15.04
15.05 --------------------------------------------------------------------------------
15.05 -- Trying 'Unix Makefiles' generator
15.05 --------------------------------
15.05 ---------------------------
15.05 ----------------------
15.05 -----------------
15.05 ------------
15.05 -------
15.05 --
15.13 Not searching for unused variables given on the command line.
15.85 -- The C compiler identification is GNU 10.2.1
15.92 -- Detecting C compiler ABI info
16.74 -- Detecting C compiler ABI info - done
16.84 -- Check for working C compiler: /usr/bin/cc - skipped
16.84 -- Detecting C compile features
16.84 -- Detecting C compile features - done
17.50 -- The CXX compiler identification is GNU 10.2.1
17.56 -- Detecting CXX compiler ABI info
18.37 -- Detecting CXX compiler ABI info - done
18.46 -- Check for working CXX compiler: /usr/bin/c++ - skipped
18.47 -- Detecting CXX compile features
18.47 -- Detecting CXX compile features - done
18.48 -- Configuring done
18.48 -- Generating done
18.49 -- Build files have been written to: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_cmake_test_compile/build
18.50 --
18.50 -------
18.50 ------------
18.50 -----------------
18.50 ----------------------
18.50 ---------------------------
18.50 --------------------------------
18.50 -- Trying 'Unix Makefiles' generator - success
18.50 --------------------------------------------------------------------------------
18.50
18.51 Configuring Project
18.51 Working directory:
18.51 /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build
18.51 Command:
18.51 /usr/bin/cmake /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce -G 'Unix Makefiles' --no-warn-unused-cli -DCMAKE_INSTALL_PREFIX:PATH=/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-install -DPYTHON_VERSION_STRING:STRING=3.9.2 -DSKBUILD:INTERNAL=TRUE -DCMAKE_MODULE_PATH:PATH=/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/resources/cmake -DPYTHON_EXECUTABLE:PATH=/usr/bin/python3 -DPYTHON_INCLUDE_DIR:PATH=/usr/include/python3.9 -DPython_EXECUTABLE:PATH=/usr/bin/python3 -DPython_ROOT_DIR:PATH=/usr -DPython_FIND_REGISTRY:STRING=NEVER -DPython_INCLUDE_DIR:PATH=/usr/include/python3.9 -DPython3_EXECUTABLE:PATH=/usr/bin/python3 -DPython3_ROOT_DIR:PATH=/usr -DPython3_FIND_REGISTRY:STRING=NEVER -DPython3_INCLUDE_DIR:PATH=/usr/include/python3.9 -DCMAKE_BUILD_TYPE:STRING=Release
18.51
18.59 Not searching for unused variables given on the command line.
19.19 -- The C compiler identification is GNU 10.2.1
19.80 -- The CXX compiler identification is GNU 10.2.1
19.86 -- Detecting C compiler ABI info
20.64 -- Detecting C compiler ABI info - done
20.73 -- Check for working C compiler: /usr/bin/cc - skipped
20.73 -- Detecting C compile features
20.73 -- Detecting C compile features - done
20.75 -- Detecting CXX compiler ABI info
21.54 -- Detecting CXX compiler ABI info - done
21.63 -- Check for working CXX compiler: /usr/bin/c++ - skipped
21.63 -- Detecting CXX compile features
21.63 -- Detecting CXX compile features - done
21.64 -- *********************************************
21.64 -- Ninja Python Distribution
21.64 --
21.64 -- BUILD_VERBOSE : OFF
21.64 -- RUN_NINJA_TEST : ON
21.64 --
21.64 -- ARCHIVE_DOWNLOAD_DIR : /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build
21.64 --
21.64 -- src_archive : unix_source
21.64 -- <src_archive>_url : https://github.com/Kitware/ninja/archive/v1.11.1.g95dee.kitware.jobserver-1.tar.gz
21.65 -- <src_archive>_sha256 : 7ba84551f5b315b4270dc7c51adef5dff83a2154a3665a6c9744245c122dd0db
21.65 -- *********************************************
21.80 -- download_ninja_source - URL: https://github.com/Kitware/ninja/archive/v1.11.1.g95dee.kitware.jobserver-1.tar.gz
21.80 -- SuperBuild - CMAKE_BUILD_TYPE: Release
21.90 -- Configuring done
21.93 -- Generating done
21.94 -- Build files have been written to: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build
22.32 Scanning dependencies of target download_ninja_source
22.41 [ 5%] Creating directories for 'download_ninja_source'
22.87 [ 11%] Performing download step (download, verify and extract) for 'download_ninja_source'
22.95 -- Downloading...
22.96 dst='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/v1.11.1.g95dee.kitware.jobserver-1.tar.gz'
22.96 timeout='none'
22.96 -- Using src='https://github.com/Kitware/ninja/archive/v1.11.1.g95dee.kitware.jobserver-1.tar.gz'
26.44 -- verifying file...
26.44 file='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/v1.11.1.g95dee.kitware.jobserver-1.tar.gz'
26.45 -- Downloading... done
26.65 -- extracting...
26.65 src='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/v1.11.1.g95dee.kitware.jobserver-1.tar.gz'
26.65 dst='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/Ninja-src'
26.65 -- extracting... [tar xfz]
26.77 -- extracting... [analysis]
26.77 -- extracting... [rename]
26.77 -- extracting... [clean up]
26.77 -- extracting... done
26.92 [ 16%] No update step for 'download_ninja_source'
27.10 [ 22%] No patch step for 'download_ninja_source'
27.29 [ 27%] No configure step for 'download_ninja_source'
27.50 [ 33%] No build step for 'download_ninja_source'
27.72 [ 38%] No install step for 'download_ninja_source'
27.93 [ 44%] Completed 'download_ninja_source'
28.17 [ 44%] Built target download_ninja_source
28.30 Scanning dependencies of target build_ninja
28.41 [ 50%] Creating directories for 'build_ninja'
28.90 [ 55%] No download step for 'build_ninja'
29.12 [ 61%] No update step for 'build_ninja'
29.33 [ 66%] No patch step for 'build_ninja'
29.51 [ 72%] Performing configure step for 'build_ninja'
29.60 loading initial cache file /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/build_ninja-prefix/tmp/build_ninja-cache-Release.cmake
30.32 -- The C compiler identification is GNU 10.2.1
30.95 -- The CXX compiler identification is GNU 10.2.1
31.02 -- Detecting C compiler ABI info
31.83 -- Detecting C compiler ABI info - done
31.92 -- Check for working C compiler: /usr/bin/cc - skipped
31.92 -- Detecting C compile features
31.93 -- Detecting C compile features - done
31.95 -- Detecting CXX compiler ABI info
32.84 -- Detecting CXX compiler ABI info - done
32.93 -- Check for working CXX compiler: /usr/bin/c++ - skipped
32.93 -- Detecting CXX compile features
32.94 -- Detecting CXX compile features - done
37.96 -- IPO / LTO enabled
37.97 -- Performing Test flag_no_deprecated
38.72 -- Performing Test flag_no_deprecated - Success
38.72 -- Performing Test flag_color_diag
39.47 -- Performing Test flag_color_diag - Success
39.57 -- Looking for fork
40.34 -- Looking for fork - found
40.34 -- Looking for pipe
41.10 -- Looking for pipe - found
41.15 -- Configuring done
41.35 -- Generating done
41.36 -- Build files have been written to: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/Ninja-build
41.51 [ 77%] Performing build step for 'build_ninja'
41.82 [ 1%] Generating lexer.cc
41.95 [ 2%] Generating depfile_parser.cc
42.08 Scanning dependencies of target libninja-re2c
42.18 [ 4%] Building CXX object CMakeFiles/libninja-re2c.dir/depfile_parser.cc.o
44.05 [ 5%] Building CXX object CMakeFiles/libninja-re2c.dir/lexer.cc.o
46.17 [ 5%] Built target libninja-re2c
46.31 Scanning dependencies of target libninja
46.42 [ 7%] Building CXX object CMakeFiles/libninja.dir/src/build_log.cc.o
50.01 [ 8%] Building CXX object CMakeFiles/libninja.dir/src/build.cc.o
55.84 [ 10%] Building CXX object CMakeFiles/libninja.dir/src/clean.cc.o
59.49 [ 11%] Building CXX object CMakeFiles/libninja.dir/src/clparser.cc.o
61.63 [ 13%] Building CXX object CMakeFiles/libninja.dir/src/dyndep.cc.o
64.66 [ 14%] Building CXX object CMakeFiles/libninja.dir/src/dyndep_parser.cc.o
67.94 [ 15%] Building CXX object CMakeFiles/libninja.dir/src/debug_flags.cc.o
68.23 [ 17%] Building CXX object CMakeFiles/libninja.dir/src/deps_log.cc.o
71.21 [ 18%] Building CXX object CMakeFiles/libninja.dir/src/disk_interface.cc.o
73.30 [ 20%] Building CXX object CMakeFiles/libninja.dir/src/edit_distance.cc.o
74.99 [ 21%] Building CXX object CMakeFiles/libninja.dir/src/eval_env.cc.o
77.91 [ 23%] Building CXX object CMakeFiles/libninja.dir/src/graph.cc.o
82.09 [ 24%] Building CXX object CMakeFiles/libninja.dir/src/graphviz.cc.o
84.75 [ 26%] Building CXX object CMakeFiles/libninja.dir/src/json.cc.o
86.13 [ 27%] Building CXX object CMakeFiles/libninja.dir/src/line_printer.cc.o
87.81 [ 28%] Building CXX object CMakeFiles/libninja.dir/src/manifest_parser.cc.o
91.32 [ 30%] Building CXX object CMakeFiles/libninja.dir/src/metrics.cc.o
93.08 [ 31%] Building CXX object CMakeFiles/libninja.dir/src/missing_deps.cc.o
97.68 [ 33%] Building CXX object CMakeFiles/libninja.dir/src/parser.cc.o
99.55 [ 34%] Building CXX object CMakeFiles/libninja.dir/src/state.cc.o
103.4 [ 36%] Building CXX object CMakeFiles/libninja.dir/src/status.cc.o
106.4 [ 37%] Building CXX object CMakeFiles/libninja.dir/src/string_piece_util.cc.o
108.3 [ 39%] Building CXX object CMakeFiles/libninja.dir/src/tokenpool-gnu-make.cc.o
109.8 [ 40%] Building CXX object CMakeFiles/libninja.dir/src/util.cc.o
113.6 [ 42%] Building CXX object CMakeFiles/libninja.dir/src/version.cc.o
115.0 [ 43%] Building CXX object CMakeFiles/libninja.dir/src/subprocess-posix.cc.o
117.2 [ 44%] Building CXX object CMakeFiles/libninja.dir/src/tokenpool-gnu-make-posix.cc.o
117.9 [ 44%] Built target libninja
118.0 Scanning dependencies of target manifest_parser_perftest
118.1 [ 46%] Building CXX object CMakeFiles/manifest_parser_perftest.dir/src/manifest_parser_perftest.cc.o
121.1 [ 47%] Linking CXX executable manifest_parser_perftest
133.3 [ 47%] Built target manifest_parser_perftest
133.4 Scanning dependencies of target hash_collision_bench
133.5 [ 49%] Building CXX object CMakeFiles/hash_collision_bench.dir/src/hash_collision_bench.cc.o
135.6 [ 50%] Linking CXX executable hash_collision_bench
138.3 [ 50%] Built target hash_collision_bench
138.4 Scanning dependencies of target depfile_parser_perftest
138.5 [ 52%] Building CXX object CMakeFiles/depfile_parser_perftest.dir/src/depfile_parser_perftest.cc.o
140.3 [ 53%] Linking CXX executable depfile_parser_perftest
143.2 [ 53%] Built target depfile_parser_perftest
143.4 Scanning dependencies of target clparser_perftest
143.5 [ 55%] Building CXX object CMakeFiles/clparser_perftest.dir/src/clparser_perftest.cc.o
145.2 [ 56%] Linking CXX executable clparser_perftest
148.5 [ 56%] Built target clparser_perftest
148.6 Scanning dependencies of target canon_perftest
148.7 [ 57%] Building CXX object CMakeFiles/canon_perftest.dir/src/canon_perftest.cc.o
150.3 [ 59%] Linking CXX executable canon_perftest
152.7 [ 59%] Built target canon_perftest
152.8 Scanning dependencies of target ninja_test
152.9 [ 60%] Building CXX object CMakeFiles/ninja_test.dir/src/build_log_test.cc.o
156.3 [ 62%] Building CXX object CMakeFiles/ninja_test.dir/src/build_test.cc.o
166.6 [ 63%] Building CXX object CMakeFiles/ninja_test.dir/src/clean_test.cc.o
170.9 [ 65%] Building CXX object CMakeFiles/ninja_test.dir/src/clparser_test.cc.o
173.8 [ 66%] Building CXX object CMakeFiles/ninja_test.dir/src/depfile_parser_test.cc.o
177.3 [ 68%] Building CXX object CMakeFiles/ninja_test.dir/src/deps_log_test.cc.o
180.8 [ 69%] Building CXX object CMakeFiles/ninja_test.dir/src/disk_interface_test.cc.o
184.5 [ 71%] Building CXX object CMakeFiles/ninja_test.dir/src/dyndep_parser_test.cc.o
188.3 [ 72%] Building CXX object CMakeFiles/ninja_test.dir/src/edit_distance_test.cc.o
190.9 [ 73%] Building CXX object CMakeFiles/ninja_test.dir/src/graph_test.cc.o
195.9 [ 75%] Building CXX object CMakeFiles/ninja_test.dir/src/json_test.cc.o
198.6 [ 76%] Building CXX object CMakeFiles/ninja_test.dir/src/lexer_test.cc.o
201.3 [ 78%] Building CXX object CMakeFiles/ninja_test.dir/src/manifest_parser_test.cc.o
205.6 [ 79%] Building CXX object CMakeFiles/ninja_test.dir/src/missing_deps_test.cc.o
209.4 [ 81%] Building CXX object CMakeFiles/ninja_test.dir/src/ninja_test.cc.o
212.1 [ 82%] Building CXX object CMakeFiles/ninja_test.dir/src/state_test.cc.o
215.1 [ 84%] Building CXX object CMakeFiles/ninja_test.dir/src/string_piece_util_test.cc.o
218.0 [ 85%] Building CXX object CMakeFiles/ninja_test.dir/src/subprocess_test.cc.o
221.1 [ 86%] Building CXX object CMakeFiles/ninja_test.dir/src/test.cc.o
225.2 [ 88%] Building CXX object CMakeFiles/ninja_test.dir/src/tokenpool_test.cc.o
227.9 [ 89%] Building CXX object CMakeFiles/ninja_test.dir/src/util_test.cc.o
230.8 [ 91%] Linking CXX executable ninja_test
375.9 [ 91%] Built target ninja_test
376.0 [ 92%] Generating build/browse_py.h
376.3 Scanning dependencies of target ninja
376.4 [ 94%] Building CXX object CMakeFiles/ninja.dir/src/ninja.cc.o
382.0 [ 95%] Building CXX object CMakeFiles/ninja.dir/src/browse.cc.o
383.2 [ 97%] Linking CXX executable ninja
420.5 [ 97%] Built target ninja
420.6 Scanning dependencies of target build_log_perftest
420.7 [ 98%] Building CXX object CMakeFiles/build_log_perftest.dir/src/build_log_perftest.cc.o
423.6 [100%] Linking CXX executable build_log_perftest
435.9 [100%] Built target build_log_perftest
436.0 [ 83%] Stripping CMake executables
436.1 [ 88%] Running Ninja test suite
437.3 gmake[2]: *** [CMakeFiles/build_ninja.dir/build.make:143: build_ninja-prefix/src/build_ninja-stamp/build_ninja-run_ninja_test_suite] Error 130
437.3 gmake[1]: *** [CMakeFiles/Makefile2:97: CMakeFiles/build_ninja.dir/all] Error 2
437.3 gmake: *** [Makefile:149: all] Error 2
437.3 Traceback (most recent call last):
437.3 File "/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/setuptools_wrap.py", line 674, in setup
437.3 cmkr.make(make_args, install_target=cmake_install_target, env=env)
437.3 File "/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/cmaker.py", line 697, in make
437.3 self.make_impl(clargs=clargs, config=config, source_dir=source_dir, install_target=install_target, env=env)
437.3 File "/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/cmaker.py", line 742, in make_impl
437.3 raise SKBuildError(msg)
437.3
437.3 An error occurred while building with CMake.
437.3 Command:
437.3 /usr/bin/cmake --build . --target install --config Release --
437.3 Install target:
437.3 install
437.3 Source directory:
437.3 /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce
437.3 Working directory:
437.3 /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build
437.3 Please check the install target is valid and see CMake's output for more information.
437.3
437.4 error: subprocess-exited-with-error
437.4
437.4 × Building wheel for ninja (pyproject.toml) did not run successfully.
437.4 │ exit code: 1
437.4 ╰─> See above for output.
437.4
437.4 note: This error originates from a subprocess, and is likely not a problem with pip.
437.4 full command: /usr/bin/python3 /usr/local/lib/python3.9/dist-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py build_wheel /tmp/tmpdw_4oh6w
437.4 cwd: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce
437.4 Building wheel for ninja (pyproject.toml): finished with status 'error'
437.4 ERROR: Failed building wheel for ninja
437.4 Failed to build ninja
437.4 ERROR: Could not build wheels for ninja, which is required to install pyproject.toml-based projects
------
Dockerfile:23
--------------------
21 | RUN python3 -m pip install --upgrade scikit-build
22 |
23 | >>> RUN python3 -m pip install ninja --verbose
--------------------
ERROR: failed to solve: process "/bin/sh -c python3 -m pip install ninja --verbose" did not complete successfully: exit code: 1
I’m building ninja in a RISC-V riscv64 container and it outputs the exactly same error.
By my testing the latest version on ninja that can be successfully built is ** 1.10.2.3**.
Errors start at 1.10.2.4.
By the way, apt install ninja-build is also an option.
I have the same issue on Raspberry Pi 4, with Python 3.10 in a pipenv. For me, 1.10.2.3 doesn't work either.
Same issue appear when I try to install streamlit on RiscV
Same issue when building for armhf / linux/arm32/v7
I don't know if building from sources got fixed on any of those platforms with 1.11.1.2 but binary armv7l wheels should now be retrieved instead of trying to build from sources.
1.11.1.2 is still broken on riscv on the NumPy CI with python3.10.
@mattip, I have no trouble building in a riscv64 ghcr.io/mayeut/manylinux_2_35:2024.11.23-1 container.
It seems NumPy CI is doing some clever things to do cross-compilation instead of native compilation through QEMU, I'll try to reproduce this workflow locally if I can. There's a first error that's related to setuptools_scm which suggest one of our build dependency isn't doing the right thing here - but the error does not seem to cause the build process to stop -, I guess from either one of those (or the combination), @henryiii should probably know what's going on under the hood in scikit-build-core:
metadata.version.provider = "scikit_build_core.metadata.setuptools_scm"
metadata.readme.provider = "scikit_build_core.metadata.fancy_pypi_readme"
The other error suggests something wrong with the cross-compilation environment:
/host/usr/riscv64-linux-gnu/include/wchar.h:35:10: fatal error: stddef.h: No such file or directory
35 | #include <stddef.h>
That being said, for Numpy QEMU CI , you might want to filter out ninja from the build requirements as you're overriding it with the host ninja right after installation.
Closing this issue as we could verify building from sources on armhf (where there's now binary wheels which should also reduce the number of users building from sources) & riscv (tested on ubuntu 22.04).
Please re-open or open a new issue if the last version did not solve the issue for your use-case.
| gharchive/issue | 2023-11-06T18:15:55 | 2025-04-01T06:45:44.416508 | {
"authors": [
"AnatolySh",
"KumaTea",
"fberndl",
"gdmrw",
"guerda",
"mattip",
"mayeut"
],
"repo": "scikit-build/ninja-python-distributions",
"url": "https://github.com/scikit-build/ninja-python-distributions/issues/226",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
548810719 | Instructions to set up a development environment
This is a tiny commit which adds a requirements-dev.txt containing the packages needed to run the suite and a short description how to set up a working environment.
I always forget which packages are needed for the test suite and I run pytest a few times. This will make it a bit easier :wink:
Yes, that I understood. My instructions are indeed aimed for what you have described in your last paragraph, I just skipped that it's cloned via git as I assumed it is anyways.
Following that, maybe a tiny change of the line cd /path/to/... to git clone ... && cd .. is maybe enough, or what do you think?
The main idea was to make it a bit simpler to get started with a green test suite.
I just updated the setup.py to use extras_require with a "testing" entry. Not sure though if the tests_require is used somewhere in your CI or so.
The tests_require is not explicitly relied upon in CI.
I like this solution using extras_require! It's good to know that Python has a built-in concept of "optional dependencies."
Alright, ready for merge then. Thanks!
| gharchive/pull-request | 2020-01-13T09:41:53 | 2025-04-01T06:45:44.437437 | {
"authors": [
"jpivarski",
"tamasgal"
],
"repo": "scikit-hep/uproot",
"url": "https://github.com/scikit-hep/uproot/pull/436",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
842855848 | Some more memberwise stuff
I've been trying to read some CMS edm collections, and played around a bit, finding at least one memberwise interpretation successful, with a header like >h: version >I: length >I: values_num_bytes, e.g.
--+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+-
0 10 0 0 16 1 64 22 127 198 0 9 0 0 0 0 0 0 0 0
--- --- --- --- --- --- @ --- --- --- --- --- --- --- --- --- --- --- --- ---
for a class that indeed was version 10 and had the right amount of bytes. This class had a very simple set of members (just one):
_stl_container0 = uproot.containers.AsVector(False, numpy.dtype("u1"))
base_names_versions = []
member_names = ['data_']
class_flags = {}
@kratsg might be interested.
I think we can close this for now. Perhaps in the future we can revisit. At least some work has been documented.
Yeah, for now let's close this... Jim knows very well that I probably broke my brain trying to understand parts of this back then... and I'm happy to try to stab this again in the future when I have more spare cycles, or if we want to get a student working on this... I have a perfect project for them.
Yeah, that's where it would be nice to get a student/fellow to jump in. I have ROOT files they can use with RooFit/RooStats stuff which are always memberwise-serialized and it's a good test bench to try and get uproot to read those out correctly.
I agree; we'd need some concerted effort from someone who can concentrate on just this for at least a month. Knowing that you have a lot of sample files is a good thing.
Even pedantically worse -- ROOT generates them in different formats pre 6.18 and post 6.20 -- where they're not backward-compatible readable even within ROOT... so it will be interesting to see if uproot can bridge this gap.
| gharchive/pull-request | 2021-03-28T23:18:18 | 2025-04-01T06:45:44.441462 | {
"authors": [
"jpivarski",
"kratsg",
"nsmith-"
],
"repo": "scikit-hep/uproot4",
"url": "https://github.com/scikit-hep/uproot4/pull/314",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1116071174 | Set baseurl to "/" to make CSS / JS loading work under https://blog.scikit-learn.org/
As discussed in #23.
That should fix browsing from https://blog.scikit-learn.org/ .
But that will break browsing from https://scikit-learn.org/blog/ (but maybe we don't care?)
Edit: now https://scikit-learn.org/blog/ automatically redirects to https://blog.scikit-learn.org/.
Since the redirect is active, let's merge to see if that fixes the rendering.
| gharchive/pull-request | 2022-01-27T11:02:51 | 2025-04-01T06:45:44.461323 | {
"authors": [
"ogrisel"
],
"repo": "scikit-learn/blog",
"url": "https://github.com/scikit-learn/blog/pull/63",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
310845467 | 'NoneType' object is not iterable
This uninformative error comes from LeveragingBagging, under for example, covtype.csv. See _test_file_stream_multiple_cfier.py ;the error is quite subtle, in fact it is not even explicitly an error:
root - INFO - Prequential Evaluation
root - INFO - Generating 7 targets.
root - INFO - Pre-training on 100 samples.
root - INFO - Evaluating...
'NoneType' object is not iterable
root - INFO - Evaluation time: 0.000 s
root - INFO - Total instances: 100
root - INFO - Global performance:
root - INFO - Learner 0 - Accuracy : 0.000
The root cause seems to be in the KNN (used as default model in leveraging bag) when working with multi-label data.
Good, that can probably be fixed. The thing is that sometimes the problems are very silent, .. and its difficult to see the problem. There is another one:
"classes should include all valid labels that can be in y"
which probably comes from sklearn -- maybe we should break the stream and return an error when these things happen, rather than continuing and producing a summary at the end as if all was well?
| gharchive/issue | 2018-04-03T13:50:51 | 2025-04-01T06:45:44.605601 | {
"authors": [
"jacobmontiel",
"jmread"
],
"repo": "scikit-multiflow/scikit-multiflow",
"url": "https://github.com/scikit-multiflow/scikit-multiflow/issues/19",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
489837421 | PS: improve metrics
[x] Segment registrations by AS and type (a given AS will only register a single type of segment, but this allows querying along either axis). (done by #3152)
[x] Segment registrations by update (i.e. it already exists in pathdb) vs new. (done by #3152)
[x] Segment requests by type (up/down/core) and number of segments and revocations returned. Have label to indicate status of CacheOnly flag. (done by #3232)
[x] Segment requests errors by type (up/down/core) and error type. (details https://github.com/scionproto/scion/issues/3106#issuecomment-528821888) (done by #3232)
[x] segsyncer metrics. (done by #3241)
[x] Revocations received. (details https://github.com/scionproto/scion/issues/3106#issuecomment-531679988)
[x] ~SCMP errors received.~ Replaced by #3107, for now. Eventually it would be nice to know what actions got an scmp error reply.
[x] Crypto issues:
[x] cert/trc request failures: timeouts vs not available
[x] Signature verification failed (e.g. possible forgery)
[x] Signature validation failed (e.g. signature from expired cert)
[x] ~Dispatcher errors/reconnects.~ Replaced by #3107
[x] Implement https://github.com/scionproto/scion/issues/3121 for all periodic tasks
Currently we have request counts handled generally in the messenger. We could expand the Handler interface to provide more information about a request and additional metric labels, so that we can still keep this in the messenger.
@lukedirtwalker : that will works for some stuff, but not others. E.g. knowing whether a registration is updating an existing segment or not requires the business logic of the PS.
@kormat that can only be decided ~after the handler is done, so it could be in a map in the handler result (maybe?)
It gets hard to handle this in a generic way. E.g. a path registration should indicate the type (up/down/core) in the "result", plus whether it's an update or new. A path request should indicate the type (up/down/core) and cache_only status in the "result". It seems likely to end up a real mess.
As an example of what i'm thinking of:
# Number of path registrations. "result" can be one of: new, updated, err_crypto, err_db
pathsrv_registrations_total{elem="ps1-1-1", type="up", src_ia="1-34", result="updated"} 443
# Number of path requests. Incremented by one for every request. "result" can be one of: cached, fetched, err_crypto, err_db, err_timeout.
pathsrv_requests_total{elem="ps1-1-1", type="core", cache_only="false", dst_isd="3", result="fetched"} 11
# Number of segments in replies to path requests. Could be a histogram later if we really wanted more insight. Incremented by the number of segments in a reply.
pathsrv_requests_reply_segs_total(elem="ps1-1-1", type="core", cache_only="false", dst_isd="3"} 32
@kormat question:
Revocations received. should that also include revocations received as part of a SegReply ?
Yeah, i think we want a metric for revocations received, with a label to say whether it's:
part of a SegReply
a top-level PathMgmt.sRevInfo
an SCMP revocation
and another label to say what the result is (expired, crypto not available, invalid, ok, etc).
segsyncer metrics:
# Number of segments received from segment syncing "result" can be one of: new, updated, err_crypto, err_db
ps_sync_regs_total{elem="ps1-1-1", type="up", src_ia="1-34", result="updated"} 93
# Number of segment syncer executions
ps_segsyncer_pushes_total{elem="ps1-1-1", dst_ia="1-32", result="ok_succsess"} 31
All metrics should be in by now. If there is something missing we should create a new issue.
| gharchive/issue | 2019-09-05T15:27:06 | 2025-04-01T06:45:44.636090 | {
"authors": [
"kormat",
"lukedirtwalker"
],
"repo": "scionproto/scion",
"url": "https://github.com/scionproto/scion/issues/3106",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
3141275 | Added a Sylvester equation solver to scipy.linalg
I've added a function to provide a solution to the Sylvester equation (AX + XB = Q) to the linear algebra module. The code is currently grouped with "basic" linear algebra routines as this seems to be the only applicable place to include such a solver. The routine relies on some Schur decomposition calls and a LAPACK call. If including the code in this location is sub-optimal, maybe another location in SciPy could be suggested.
I did a quick read-through: the code is very clear, well-written, and well documented, and the test-coverage is great. Nice work!
As far as including this in basic.py: I think it's fine. If we end up adding more specialized solvers in the future, it may be better to factor them out into a "solvers" submodule, but I don't think that's necessary right now.
One more thing: you should add a brief description of the routine to the doc string in scipy/linalg/info.py.
I added a doc string, but I think I tripped up adding a merge from scipy/master. Hopefully we're still ok.
Looks pretty good to me too.
The committer probably should rebase to get rid of the merge, and move the "Computes a solution to the Sylvester matrix equation ..." part of the docstring to its Notes section.
@jakevdp: Note that you can also use git commit --amend --author="..." to change the author field of the commit (the committer field retains your name).
| gharchive/issue | 2012-02-08T14:17:14 | 2025-04-01T06:45:44.644983 | {
"authors": [
"ArmstrongJ",
"jakevdp",
"pv"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/154",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1110250610 | BUG: Documentation of stft misses the import statement for numpy
Describe your issue.
The sample code given for stft in the signal module misses the following line:
import numpy as np
Reproducing Code Example
from scipy import signal
import matplotlib.pyplot as plt
rng = np.random.default_rng()
Error message
NameError: name 'np' is not defined
SciPy/NumPy/Python version information
1.7.3 1.21.4 sys.version_info(major=3, minor=9, micro=9, releaselevel='final', serial=0)
Have a look at https://github.com/scipy/scipy/issues/13049
@ilayn, thanks for the very quick response. Sorry I missed that issue.
I was just looking for bugs related to stft and didn't expect it is a
general issue. One can then probably also close my related PR.
| gharchive/issue | 2022-01-21T09:33:21 | 2025-04-01T06:45:44.648424 | {
"authors": [
"hoehnp",
"ilayn"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/15446",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1316022151 | BUG: malloc(0) in NI_InitFilterOffsets
Describe your issue.
In NI_InitFilterOffsets, it's possible to end up with a calculation where malloc(0) would be called, due to one of the multiplication parameters being 0 (e.g., footprint).
Passing 0 to malloc is implementation-defined and may return some pointer (not necessary NULL), but any attempts to dereference that pointer would be invalid.
Here's an example patch that detects this issue and a test that fails (from my WIP dev branch):
https://github.com/nkaretnikov/scipy/commit/c44143f16b527d21dc865ea026ccddddd2698fd7
https://github.com/nkaretnikov/scipy/commit/a99cbbdd62570bc61d52da6e4f2dce2943e0b31f
Reproducing Code Example
see above, note that the code is slightly different than in main. but here's an inlined version:
diff --git a/scipy/ndimage/src/ni_support.c b/scipy/ndimage/src/ni_support.c
index de7bace3a..481a7ffbb 100644
--- a/scipy/ndimage/src/ni_support.c
+++ b/scipy/ndimage/src/ni_support.c
@@ -583,14 +583,24 @@ int NI_InitFilterOffsets(PyArrayObject *array, npy_bool *footprint,
offsets_size *= (ashape[ii] < fshape[ii] ? ashape[ii] : fshape[ii]);
/* allocate offsets data: */
*offsets_bytes = offsets_size * footprint_size * sizeof(npy_intp);
+ if (*offsets_bytes == 0) {
+ PyErr_SetString(PyExc_RuntimeError, "*offsets_bytes is 0");
+ goto exit;
+ }
*offsets = malloc(*offsets_bytes);
if (!*offsets) {
PyErr_NoMemory();
goto exit;
}
+
if (coordinate_offsets) {
*coordinate_offsets_bytes = offsets_size * rank
* footprint_size * sizeof(npy_intp);
+ if (*coordinate_offsets_bytes == 0) {
+ PyErr_SetString(
+ PyExc_RuntimeError, "*coordinate_offsets_bytes is 0");
+ goto exit;
+ }
*coordinate_offsets = malloc(*coordinate_offsets_bytes);
if (!*coordinate_offsets) {
PyErr_NoMemory();
# XXX: This ends up calling NI_BinaryErosion, which calls NI_InitFilterOffsets.
# In the latter, footprint_size is computed and set to 0, so malloc(0) is called
# for offsets. Then in NI_BinaryErosion, the offsets pointer is passed to
# NI_FilterNext2 where it would be adjusted, but this is prevented by the size
# check at the start of NI_FilterNext2, so the whole thing errors out.
@pytest.mark.skip(reason="footprint_size equals 0")
def test_binary_hit_or_miss_input_as_output():
rstate = numpy.random.RandomState(123)
data = rstate.randint(low=0, high=2, size=100).astype(bool)
# input data is not modified
data_orig = data.copy()
expected = ndimage.binary_hit_or_miss(data)
assert_array_equal(data, data_orig)
# data should now contain the expected result
ndimage.binary_hit_or_miss(data, output=data)
assert_array_equal(expected, data)
commit 4277f939b1d63b2212a3903eb500b3ab16bc33a8 (upstream-https/main, upstream-https/HEAD)
Merge: 7d9301f7c 33ea45b44
Author: Ralf Gommers <ralf.gommers@gmail.com>
Date: Sat Jul 2 13:16:40 2022 +0200
Merge pull request #16444 from sabonerune/fix/messagestream_tempfile
BUG: fix fail to open tempfile in messagestream.pyx (#8850)
Thanks @nkaretnikov. This reproducer doesn't fail for me without your patch. Is it supposed to?
Also, is the shown commit for messagestream.pyx relevant?
@rgommers i've created a pr here: https://github.com/scipy/scipy/pull/16697. it's based on top of the current main, so should be easy to repro. i've also adjusted the comment there to better match the behavior in main.
about the bug: passing 0 to malloc is valid as long as you pass the resulting pointer to free iiuc. that's the only guarantee you have. dereferencing this malloc pointer is not valid (since it would be either NULL or would be pointing to some address backed by 0 bytes, this is implementation-defined). the test doesn't seem to crash with asan if i try without these checks. however, it might be just because asan fails to catch this. i discovered this on my branch which implements extensive size checking. without this patch on that branch, you should get an error later due to accessing out of bounds memory. iirc, the pointer is adjusted (incremented or decremented) and then dereferenced. but it doesn't make sense to do with a pointer pointing to 0 bytes.
now, as i mention above, using malloc like this seems problematic, so better to just fail early instead of relying on other code to notice. it's all due to one of the arguments used in multiplication being 0 (the footprint).
here's a SO discussion on this: https://stackoverflow.com/questions/2022335/whats-the-point-of-malloc0
| gharchive/issue | 2022-07-24T22:14:48 | 2025-04-01T06:45:44.655617 | {
"authors": [
"nkaretnikov",
"rgommers"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/16694",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2090727004 | ENH: Adding COBYQA to scipy.optimize?
Is your feature request related to a problem? Please describe.
As discussed in https://github.com/scipy/scipy/issues/18118, COBYLA is the only derivative-free optimization solver available in SciPy. COBYQA is a solver we developed with my colleague Zaikun Zhang from The Hong Kong Polytechnic University. It is designed to supersede COBYLA as a general solver.
Describe the solution you'd like.
I want to propose the inclusion of the COBYQA solver in scipy.optimize. A Python implementation of COBYQA is available at https://github.com/cobyqa/cobyqa, and the related Python package is available on PyPI. Numerical experiments we conducted show the clear superiority of COBYQA over COBYLA in general.
I would be happy to share our thoughts on this inclusion.
Describe alternatives you've considered.
No response
Additional context (e.g. screenshots, GIFs)
No response
Thanks @ragonneau, can you post this on the mailing list as this is required for new features?
That being said I think I am +1
Pinging also @zaikunzhang
Thanks @ragonneau for the proposal. Is there sign of general adoption yet?
We have strict academic criteria for inclusion in SciPy. Usually we don't follow trends and add new things but wait until there is a large consensus around adoption of new methods.
https://docs.scipy.org/doc/scipy/dev/core-dev/index.html#core-dev-guide
Based on that alone I would be -1.
I'd like to understand the lineage and the status of this method. If it's completely new, I'd second Pamphile's opinion: it should mature as a separate package. However, is it a completely new method?
There is COBYLA by Powell; then there's PRIMA by Zhang & Powell, and now here's COBYQA by Ragonneau & Zhang.
So PRIMA is a streamlined and modernized implementation of COBYLA, and we sort of planned to replace the latter by the former. Now PRIMA is still in FORTRAN so https://github.com/scipy/scipy/issues/18566 applies.
So is COBYQA separate if related to COBYLA/PRIMA, or can it be thought of an improved version of the same method?
If it is, we can I guess think of formulating a plan to skip PRIMA and go straight to COBYQA maybe.
Also cross-ref gh-1477 and gh-18118.
I'd like to understand the lineage and the status of this method. If it's completely new, I'd second Pamphile's opinion: it should mature as a separate package. However, is it a completely new method? There is COBYLA by Powell; then there's PRIMA by Zhang & Powell, and now here's COBYQA by Ragonneau & Zhang.
So PRIMA is a streamlined and modernized implementation of COBYLA, and we sort of planned to replace the latter by the former. Now PRIMA is still in FORTRAN so #18566 applies.
No, PRIMA is not in FORTRAN. PRIMA is in Fortran, which I regard as a different language that conveniently benefits from the abundant libraries in FORTRAN without bearing the historical burden.
So is COBYQA separate if related to COBYLA/PRIMA, or can it be thought of an improved version of the same method? If it is, we can I guess think of formulating a plan to skip PRIMA and go straight to COBYQA maybe.
COBYQA is not COBYLA. COBYQA is a completely new and different algorithm, using more advanced techniques than COBYLA. This particularly includes SQP and derivative-free symmetric Broyden update.
Thank you @zaikunzhang
It is in Fortran, which I regard as a different language that conveniently benefits from the abundant libraries in FORTRAN without bearing the historical burden.
OK, let's keep the F77 vs modern Fortran discussion in gh-18118 or elsewhere.
COBYQA is not COBYLA. COBYQA is a completely new and different algorithm, using more advanced techniques than COBYLA. This particularly includes SQP and derivative-free symmetric Broyden update. Tom @ragonneau and I (Zaikun Zhang) developed COBYQA as part of Tom's Ph.D. thesis.
Thanks. So is COBYQA in some sense a superset of COBYLA? As in, can it be expected to be strictly better than COBYLA, would domain experts (you) recommend it over cobyla/prima?
shgo was also a very new solver when it was added to scipy so I would not directly reject COBYQA. How does COBYQA perform for example on the CUTEST benchmark compared to what scipy currently has to offer? Not sure about the quality of our benchmark suite there though.
Regarding COBYLA I would strongly argue against removing it for COBYLA. In my experience, no optimizer is strictly better than another (no free lunch theorem etc. ..) and COBYLA is a famous and widely used algorithm.
Thanks. So is COBYQA in some sense a superset of COBYLA? As in, can it be expected to be strictly better than COBYLA, would domain experts (you) recommend it over cobyla/prima?
No, COBYQA is not a superset of COBYLA. They are different algorithms, just like GMRES is not a superset of CG or vice versa.
From a mathematical perspective, the techniques used in COBYQA are more advanced. We have developed COBYQA as a successor of COBYLA. COBYQA is a package under development, but COBYLA has been stabilized for almost 30 years, the last major change being the modernization I made. Note that "under development" does not mean that it is not ready for production use, but means that new features and new improvements will continue to be made (recall that the Linux kernel is under development).
No algorithm can always outperform another one on all problems. We, as algorithm developers, try to develop algorithms that perform better than others on most test problems we have. This does not mean that our algorithm will outperform others on a particular problem. However, a particular user normally only cares about a single particular problem.
If a user asks my advice for solving a nonlinearly constrained optimization problem, I would suggest trying both COBYLA and COBYQA. Both have their advantages.
Note that it is improper to compare PRIMA with COBYQA, the former being a package of five solvers while the latter being a particular solver. For unconstrained, bound-constrained, and linearly-constrained problems, the solvers in PRIMA are the state of the art, and the performance of COBYQA is comparable when the number of variables is at most 50. We have not tested larger problems systematically due to the speed limitation of Python, but PRIMA algorithms are tested on larger problems every day.
COBYQA has its own advantages. For example, it respects the bound constraints (if any), meaning that it never evaluates the objective/constraint functions at points outside the bounds. This is the next improvement I will make to LINCOA and COBYLA.
COBYQA still has much potential to be fulfilled and explored. I believe it can still be improved to obtain high-precision solutions.
Derivative-free optimization methods (aka, zeroth-order methods) is attracting more and more attention due to new applications arising from machine learning and AI. My suggestion is to make both PRIMA solvers (including COBYLA) and COBYLA available in SciPy --- this will be of great value to both SciPy and the community.
Thank you.
Regarding COBYLA I would strongly argue against removing it for COBYQA. In my experience, no optimizer is strictly better than another (no free lunch theorem etc. ..) and COBYLA is a famous and widely used algorithm.
I hold the same opinion.
shgo was also a very new solver when it was added to scipy so I would not directly reject COBYQA. How does COBYQA perform for example on the CUTEST benchmark compared to what scipy currently has to offer? Not sure about the quality of our benchmark suite there though.
We have our own benchmark suite that one can run (see in the benchmark folder). If, as shgo, the method proves itself to be competitive, then by all means my -1 would become +1 😃
Running the benchmark would in any case be the next step for me.
Running the benchmark would in any case be the next step for me.
Agreed, it would be great to see the result of those. @ragonneau you can find those in benchmarks/benchmarks/optimize.py. It would be very useful to see the results of those benchmarks that include COBYLA now (and other solvers), and add COBYQA to that.
We have our own benchmark suite that one can run (see in the benchmark folder).
Those include benchmarks from CUTEST, which may be what @dschmitz89 meant.
A Python implementation of COBYQA is available at https://github.com/cobyqa/cobyqa,
I had a quick browse of this package; the code generally looks in good shape. It's not so easy to immediately see how much of it would be included in SciPy, since there's models, problem sets, and some duplicated code from SciPy. But it looks like it'd be a few thousand lines of code. Those will presumably land in a single file in the SciPy implementation (e.g., scipy/optimize/_cobyla.py). You'll probably have to put that together anyway to be able to run the benchmarks. It'll be nice to look at that once it's ready.
Those will presumably land in a single file in the SciPy implementation (e.g., scipy/optimize/_cobyla.py).
Thank you @rgommers, I suppose you meant scipy/optimize/_cobyqa.py.
Yes indeed - that was a typo
Agreed, it would be great to see the result of those. @ragonneau you can find those in benchmarks/benchmarks/optimize.py. It would be very useful to see the results of those benchmarks that include COBYLA now (and other solvers), and add COBYQA to that.
I ran SciPy benchmarks locally after including COBYQA.
The results for BenchDFO:
[ 2.27%] ··· ============================== ============== =========== ========================
-- result type
--------------------------------------------- ------------------------------------
DFO benchmark problem number solver mean_nfev min_obj
============================== ============== =========== ========================
0 COBYLA 182.0 36.00000001629636
0 COBYQA 38.0 36.0
0 SLSQP 21.0 35.99999999999999
0 Powell 213.0 35.99999999999998
0 nelder-mead 1390.0 36.000000004916636
0 L-BFGS-B 30.0 36.00000000029089
0 BFGS 40.0 36.00000000000747
0 trust-constr 30.0 36.000000000065775
1 COBYLA 239.0 36.00000002161123
1 COBYQA 43.0 36.00000000000001
1 SLSQP 21.0 36.0
1 Powell 356.0 35.99999999999997
1 nelder-mead 1096.0 36.00000001672908
1 L-BFGS-B 50.0 36.00000000000111
1 BFGS 70.0 36.00000000020062
1 trust-constr 40.0 36.00000000004839
2 COBYLA 93.0 8.38460155098156
2 COBYQA 102.0 8.380281696617237
2 SLSQP 23.0 8.380281690140846
2 Powell 203.0 8.380281690140844
2 nelder-mead 276.0 8.380314322959702
2 L-BFGS-B 32.0 8.380281690152536
2 BFGS 252.0 8.380281698131483
2 trust-constr 128.0 8.380281690145555
3 COBYLA 107.0 8.400780446868142
3 COBYQA 121.0 8.380281704183599
3 SLSQP 23.0 8.380284602260867
3 Powell 182.0 8.380281690140844
3 nelder-mead 320.0 8.380314069260285
3 L-BFGS-B 40.0 8.380281690152602
3 BFGS 372.0 8.380281690224777
3 trust-constr 168.0 8.380281690143562
4 COBYLA 103.0 9.89172925972701
4 COBYQA 116.0 9.880597101883852
4 SLSQP 23.0 9.880597014925373
4 Powell 209.0 9.880597014925371
4 nelder-mead 267.0 9.880630904843933
4 L-BFGS-B 112.0 9.880597014925657
4 BFGS 341.0 9.880597015302904
4 trust-constr 136.0 9.880597014929531
5 COBYLA 124.0 9.880673204658704
5 COBYQA 109.0 9.880597794292857
5 SLSQP 23.0 9.880597015430164
5 Powell 139.0 9.880597014925371
5 nelder-mead 310.0 9.880645653536046
5 L-BFGS-B 40.0 9.880597014932862
5 BFGS 380.0 9.880597014940303
5 trust-constr 200.0 9.880597014935532
6 COBYLA 1000.0 0.2761343670249867
6 COBYQA 136.0 3.4577885926458176e-11
6 SLSQP 109.0 3.534301293924672e-05
6 Powell 607.0 1.7404243721438573e-26
6 nelder-mead 159.0 8.177661197416674e-10
6 L-BFGS-B 126.0 4.006329767353906e-07
6 BFGS 117.0 3.403318352733089e-11
6 trust-constr 177.0 8.189074233317919e-09
7 COBYLA 1000.0 18.15444498891401
7 COBYQA 342.0 1.6575428203999043e-08
7 SLSQP 295.0 2.4772818584112955e-05
7 Powell 950.0 1.1093356479670479e-31
7 nelder-mead 325.0 4.3387935016943286e-10
7 L-BFGS-B 33.0 18.35859455687261
7 BFGS 390.0 2.1519670672414852e-11
7 trust-constr 372.0 2.8716495252807545e-11
8 COBYLA 1000.0 0.0063088149736903664
8 COBYQA 87.0 4.72095533231052e-08
8 SLSQP 112.0 2.467522680839185e-06
8 Powell 60.0 0.0
8 nelder-mead 142.0 0.0003575878654431348
8 L-BFGS-B 120.0 1.2458244845337688e-08
8 BFGS 136.0 3.645847385308141e-12
8 trust-constr 128.0 4.664490098037049e-12
9 COBYLA 1000.0 1.302791737849277
9 COBYQA 171.0 2.6707378731301188e-08
9 SLSQP 114.0 3.1523978560448592e-06
9 Powell 54.0 0.0
9 nelder-mead 110.0 0.0001185672468711772
9 L-BFGS-B 140.0 1.0981915721898683e-05
9 BFGS 160.0 4.3547185576347985e-12
9 trust-constr 172.0 5.033947368965333e-11
10 COBYLA 1000.0 0.0019940963089381952
10 COBYQA 312.0 1.2655194279446566e-10
10 SLSQP 99.0 9.231453061284653e-05
10 Powell 908.0 1.692602785645291e-13
10 nelder-mead 305.0 1.3905860499424258e-06
10 L-BFGS-B 100.0 3.5705209681058525e-05
10 BFGS 165.0 3.266335996489636e-07
10 trust-constr 190.0 3.424058956351133e-07
11 COBYLA 1000.0 0.007588035153884722
11 COBYQA 232.0 1.9523309860714916e-08
11 SLSQP 179.0 3.44391715908274e-05
11 Powell 1577.0 3.013054708612935e-22
11 nelder-mead 380.0 1.2207288155937388e-06
11 L-BFGS-B 165.0 2.5692617247172744e-05
11 BFGS 270.0 1.7148717367764208e-07
11 trust-constr 355.0 2.2298540818792096e-07
12 COBYLA 1000.0 63.35600768650094
12 COBYQA 73.0 48.98425367984218
12 SLSQP 28.0 48.9842541402506
12 Powell 118.0 48.98425367924135
12 nelder-mead 120.0 48.98425367981376
12 L-BFGS-B 54.0 48.98425585393734
12 BFGS 27.0 48.98425367924926
12 trust-constr 60.0 48.98425367924055
13 COBYLA 1000.0 52.5180366249726
13 COBYQA 90.0 48.98425367987076
13 SLSQP 47.0 48.984285104912686
13 Powell 232.0 48.984253679248496
13 nelder-mead 119.0 1.7922698821282258e-09
13 L-BFGS-B 90.0 48.984313597884025
13 BFGS 75.0 48.98425367924157
13 trust-constr 87.0 48.98425367924004
14 COBYLA 1000.0 0.010846968025621654
14 COBYQA 104.0 0.008214877370595158
14 SLSQP 36.0 0.010926876593972496
14 Powell 435.0 0.008214880387696202
14 nelder-mead 226.0 0.008214877316410097
14 L-BFGS-B 36.0 0.009974372219725124
14 BFGS 92.0 0.008214877322487959
14 trust-constr 96.0 0.008214931318406598
15 COBYLA 1000.0 8.3527643805625
15 COBYQA 119.0 0.008214877325203946
15 SLSQP 80.0 0.010852444332488186
15 Powell 212.0 1.7749269756254313
15 nelder-mead 346.0 0.008214877318251797
15 L-BFGS-B 92.0 0.010443321667281222
15 BFGS 136.0 0.008214877306852426
15 trust-constr 1924.0 0.008214941243166055
16 COBYLA 1000.0 0.00038308147268715565
16 COBYQA 213.0 0.0003075056071085903
16 SLSQP 22.0 0.0005011638818990982
16 Powell 517.0 0.00030750560950841397
16 nelder-mead 260.0 0.00030750561104950664
16 L-BFGS-B 45.0 0.0004998454317253484
16 BFGS 165.0 0.00030750717208645955
16 trust-constr 170.0 0.000309855462194756
17 COBYLA 49.0 7191620.061242092
17 COBYQA 240.0 125887.6768230767
17 SLSQP 113.0 1417869686.359142
17 Powell 267.0 106966.8745748054
17 nelder-mead 600.0 6409.119264481873
17 L-BFGS-B 16.0 6977375.029202724
17 BFGS 1768.0 120.82601972635028
17 trust-constr 3376.0 22723.636975175406
18 COBYLA 1000.0 0.010975991274795866
18 COBYQA 631.0 0.0022876786037281716
18 SLSQP 145.0 0.0031019739183846718
18 Powell 1742.0 0.0022876970870087574
18 nelder-mead 1200.0 0.004794624522076027
18 L-BFGS-B 133.0 0.009260117102741666
18 BFGS 266.0 0.002287670084190521
18 trust-constr 371.0 0.0022877933510461667
19 COBYLA 1000.0 0.01449258959907267
19 COBYQA 1000.0 0.003041678435983581
19 SLSQP 334.0 0.002387635461912692
19 Powell 1840.0 0.017136978287569753
19 nelder-mead 1200.0 0.11412012392582763
19 L-BFGS-B 224.0 0.01039282675002245
19 BFGS 518.0 0.002287670112758827
19 trust-constr 539.0 0.0022876807830514287
20 COBYLA 1000.0 0.012314747076097223
20 COBYQA 1000.0 4.926796490034533e-05
20 SLSQP 435.0 3.7403380991915257e-05
20 Powell 3507.0 0.00013999662219714502
20 nelder-mead 1275.0 0.0037644989112963437
20 L-BFGS-B 270.0 0.0028624144428293146
20 BFGS 410.0 3.7976544143093916e-05
20 trust-constr 540.0 3.9115677318496754e-05
21 COBYLA 1000.0 0.19021806725341522
21 COBYQA 1000.0 0.12473258117608635
21 SLSQP 975.0 0.005234314850263207
21 Powell 5037.0 0.10605836851573695
21 nelder-mead 1800.0 0.5377026960946113
21 L-BFGS-B 500.0 0.0002881772390133559
21 BFGS 1000.0 2.1010434078081007e-06
21 trust-constr 1260.0 3.997821365625117e-05
22 COBYLA 1000.0 0.10003177690766717
22 COBYQA 1000.0 0.00013082934380459935
22 SLSQP 613.0 0.007622482850835413
22 Powell 9020.0 0.0009381552314367413
22 nelder-mead 2400.0 0.008256161684722775
22 L-BFGS-B 715.0 0.00033578796198918936
22 BFGS 936.0 1.8927701687374778e-06
22 trust-constr 1105.0 1.3321166924342236e-05
23 COBYLA 1000.0 0.2362971492708929
23 COBYQA 1000.0 0.10741551690232379
23 SLSQP 1331.0 0.10106796888625674
23 Powell 4778.0 6.567880194260053
23 nelder-mead 2400.0 5.580001849914321
23 L-BFGS-B 793.0 0.011328602404693662
23 BFGS 2041.0 0.000278772211425114
23 trust-constr 1612.0 1.3460798054702056e-05
24 COBYLA 1000.0 0.0027178520030068
24 COBYQA 177.0 6.38970543076227e-14
24 SLSQP 85.0 0.039023423059962446
24 Powell 1077.0 0.07583186991605789
24 nelder-mead 480.0 0.07558874075499751
24 L-BFGS-B 100.0 0.000234110576993369
24 BFGS 108.0 4.1709385349620864e-09
24 trust-constr 132.0 2.214996479723849e-06
25 COBYLA 32.0 256.1647690452771
25 COBYQA 73.0 124.36218407850306
25 SLSQP 16.0 2020.0
25 Powell 303.0 124.3621843304955
25 nelder-mead 72.0 124.3621848396066
25 L-BFGS-B 72.0 214.29894954503868
25 BFGS 147.0 124.36218235561685
25 trust-constr 180.0 124.36218235562204
26 COBYLA 678.0 85822.20465825836
26 COBYQA 161.0 85822.20162724197
26 SLSQP 130.0 85822.20163050466
26 Powell 270.0 85822.24229503315
26 nelder-mead 333.0 85822.20162974804
26 L-BFGS-B 70.0 85822.28170932793
26 BFGS 195.0 85822.20162635628
26 trust-constr 220.0 85822.2016263596
27 COBYLA 928.0 85822.20748496873
27 COBYQA 421.0 85822.20162810793
27 SLSQP 252.0 85822.20162647964
27 Powell 260.0 85822.29989083631
27 nelder-mead 397.0 85822.20164251165
27 L-BFGS-B 155.0 85822.2134844625
27 BFGS 435.0 85822.20162635644
27 trust-constr 390.0 85822.20162635874
28 COBYLA 274.0 4.567933585235919e-06
28 COBYQA 202.0 1.793038448544232e-09
28 SLSQP 73.0 0.0001275291143854763
28 Powell 2017.0 2.62902981284408e-22
28 nelder-mead 630.0 4.168900076322192e-09
28 L-BFGS-B 112.0 9.673157105942231e-07
28 BFGS 140.0 1.3774502648136339e-10
28 trust-constr 7000.0 0.019837220256553954
29 COBYLA 169.0 7.228441475939319e-07
29 COBYQA 173.0 1.2167995824458087e-08
29 SLSQP 58.0 0.00017792262972978623
29 Powell 2050.0 6.554533948298135e-23
29 nelder-mead 472.0 4.392761093121637e-07
29 L-BFGS-B 112.0 2.0258935358234505e-05
29 BFGS 176.0 6.121604659655201e-11
29 trust-constr 152.0 2.653738494450361e-10
30 COBYLA 1000.0 0.0036544588617460697
30 COBYQA 328.0 0.0035168812562413423
30 SLSQP 72.0 0.006287589283329275
30 Powell 2782.0 0.0035168774842111537
30 nelder-mead 813.0 0.0037430185611118747
30 L-BFGS-B 180.0 0.0036331210839248406
30 BFGS 261.0 0.0035168737405634758
30 trust-constr 333.0 0.0035168745891778416
31 COBYLA 507.0 9.563260675050992e-06
31 COBYQA 359.0 9.022403180824982e-08
31 SLSQP 113.0 7.559376195674753e-05
31 Powell 4760.0 1.4405961280325035e-21
31 nelder-mead 1008.0 1.993137661280346e-05
31 L-BFGS-B 190.0 0.0001645285146463783
31 BFGS 290.0 4.2026392909231023e-11
31 trust-constr 5000.0 4.6482458943424287e-10
32 COBYLA 1000.0 0.004996093419091057
32 COBYQA 428.0 0.004772756479272126
32 SLSQP 76.0 0.00748298302440139
32 Powell 2196.0 0.00654899679530417
32 nelder-mead 1142.0 0.004782384226390245
32 L-BFGS-B 143.0 0.007512546418004671
32 BFGS 363.0 0.006503957324546606
32 trust-constr 11000.0 0.03376312998006939
33 COBYLA 1000.0 0.004790522744041185
33 COBYQA 492.0 0.002800054880834453
33 SLSQP 80.0 0.006136919804702198
33 Powell 4493.0 0.0027998978593152275
33 nelder-mead 1799.0 0.0031707681652494613
33 L-BFGS-B 312.0 0.002825355695283585
33 BFGS 504.0 0.0027997619178927627
33 trust-constr 12000.0 0.026740538342589985
34 COBYLA 1000.0 0.002477075965308623
34 COBYQA 765.0 7.087734211593793e-08
34 SLSQP 47.0 1.8202225661529484e-05
34 Powell 2981.0 1.3549012557932144e-22
34 nelder-mead 2000.0 0.3406697389548668
34 L-BFGS-B 55.0 2.991968759920641e-05
34 BFGS 132.0 1.873957153532047e-13
34 trust-constr 154.0 1.7082620401841644e-08
35 COBYLA 1000.0 2.775269864105493
35 COBYQA 386.0 0.0660477478502633
35 SLSQP 74.0 1.1056629223458196
35 Powell 658.0 0.05064107286626231
35 nelder-mead 747.0 5.46489474798272e-05
35 L-BFGS-B 108.0 1.1059653777113296
35 BFGS 125.0 0.6345412274548067
35 trust-constr 198.0 1.105952334795245
36 COBYLA 1000.0 0.3000350186116832
36 COBYQA 1000.0 0.041631120736703206
36 SLSQP 426.0 0.04140440668157865
36 Powell 3143.0 0.040143302727005024
36 nelder-mead 2200.0 0.07609312836138543
36 L-BFGS-B 684.0 0.042042079864747596
36 BFGS 756.0 0.040137738738389564
36 trust-constr 1092.0 0.040137794964173
37 COBYLA 298.0 1.7898166135669258
37 COBYQA 232.0 1.7898135873783405
37 SLSQP 228.0 1.789825242678684
37 Powell 265.0 1.7898135868820382
37 nelder-mead 809.0 1.7898135880745636
37 L-BFGS-B 396.0 1.7898136300054583
37 BFGS 252.0 1.789813586881127
37 trust-constr 960.0 1.7898135869084242
38 COBYLA 289.0 10.238983941422939
38 COBYQA 424.0 10.238973723645511
38 SLSQP 216.0 10.238994699907447
38 Powell 494.0 10.238973850652153
38 nelder-mead 1058.0 10.238973442379468
38 L-BFGS-B 135.0 10.239986312974672
38 BFGS 396.0 10.23897342140286
38 trust-constr 405.0 10.238973421449707
39 COBYLA 704.0 18.281166339083896
39 COBYQA 379.0 18.281162006790357
39 SLSQP 276.0 18.281172707363083
39 Powell 603.0 18.281163008134268
39 nelder-mead 2000.0 18.281162945170095
39 L-BFGS-B 165.0 18.28315259151096
39 BFGS 517.0 18.28116175363231
39 trust-constr 638.0 18.281161753760827
40 COBYLA 864.0 22.26060556377675
40 COBYQA 560.0 22.260591789088142
40 SLSQP 398.0 22.2606981682437
40 Powell 682.0 22.260594329663107
40 nelder-mead 2200.0 22.260916512172354
40 L-BFGS-B 180.0 22.26368109072446
40 BFGS 564.0 22.260591734938853
40 trust-constr 792.0 22.260591734991067
41 COBYLA 807.0 26.2727789431198
41 COBYQA 622.0 26.272766516221093
41 SLSQP 367.0 26.272774470672505
41 Powell 715.0 26.272777250417114
41 nelder-mead 2400.0 26.27279966391945
41 L-BFGS-B 195.0 26.27565349718534
41 BFGS 611.0 26.27276639682053
41 trust-constr 949.0 26.272766396968294
42 COBYLA 1000.0 0.07310066404205529
42 COBYQA 1000.0 0.00014177578856341126
42 SLSQP 337.0 0.002869140620444522
42 Powell 472.0 0.0006099279648976476
42 nelder-mead 218.0 0.002941046450586494
42 L-BFGS-B 66.0 0.027570873338690554
42 BFGS 1680.0 1.2038440956761933e-07
42 trust-constr 2676.0 1.5612798805036442e-05
43 COBYLA 1000.0 0.07978793620432727
43 COBYQA 1000.0 0.00020598371863355666
43 SLSQP 395.0 0.002744641508948131
43 Powell 565.0 0.00026251356957190366
43 nelder-mead 326.0 0.005665898451296298
43 L-BFGS-B 371.0 0.0021694202909624477
43 BFGS 3715.0 8.540661203993601e-07
43 trust-constr 4347.0 2.5085999407170378e-05
44 COBYLA 1000.0 0.08233794111808837
44 COBYQA 1000.0 0.0006650365315197859
44 SLSQP 535.0 0.0032067490009361457
44 Powell 816.0 0.00026012887967490284
44 nelder-mead 723.0 8.272905977631895e-05
44 L-BFGS-B 126.0 0.027542394225624045
44 BFGS 6159.0 2.2593468907395e-06
44 trust-constr 4968.0 2.6883195826390666e-05
45 COBYLA 149.0 0.02029939321743818
45 COBYQA 42.0 0.00015000751307292886
45 SLSQP 84.0 6.360763241040102e-07
45 Powell 227.0 8.949262200449805e-22
45 nelder-mead 375.0 4.867800022545828e-05
45 L-BFGS-B 42.0 2.400899932123255e-10
45 BFGS 174.0 5.3673974914275e-10
45 trust-constr 78.0 2.454090575098533e-06
46 COBYLA 1000.0 1567939963474.8965
46 COBYQA 106.0 0.0007293992928959043
46 SLSQP 84.0 9.996680902431176e-10
46 Powell 278.0 1.1773407061309293e-20
46 nelder-mead 465.0 8.181505338496236e-05
46 L-BFGS-B 66.0 2.7791623937188346e-10
46 BFGS 210.0 5.368902177156296e-10
46 trust-constr 114.0 2.9953810817563102e-06
47 COBYLA 203.0 0.01914426426703756
47 COBYQA 55.0 5.4087818719220104e-05
47 SLSQP 156.0 8.591197491407533e-10
47 Powell 362.0 3.554425222840407e-20
47 nelder-mead 853.0 0.00010539806608652054
47 L-BFGS-B 63.0 1.7100578981890751e-09
47 BFGS 396.0 8.554467047410975e-10
47 trust-constr 81.0 4.0417068746088015e-06
48 COBYLA 279.0 0.0331710012331183
48 COBYQA 87.0 0.005073729292213983
48 SLSQP 139.0 0.003427873430343299
48 Powell 450.0 8.053464532704803e-20
48 nelder-mead 1117.0 0.00021983838347792528
48 L-BFGS-B 88.0 4.803599008795718e-10
48 BFGS 660.0 1.0683513286135924e-09
48 trust-constr 99.0 4.5226625371459055e-06
49 COBYLA 369.0 0.03428531322541776
49 COBYQA 95.0 0.0036452448361088343
49 SLSQP 304.0 2.269282789883771e-06
49 Powell 540.0 3.020724955433762e-19
49 nelder-mead 1411.0 0.00030925186830669545
49 L-BFGS-B 104.0 6.510760200351575e-10
49 BFGS 741.0 1.2849241618216414e-09
49 trust-constr 208.0 4.879304344714801e-06
50 COBYLA 1000.0 4476072203419.262
50 COBYQA 205.0 0.010356853034351174
50 SLSQP 174.0 16090.611615482758
50 Powell 562.0 2.8587659884100126e-19
50 nelder-mead 2400.0 9.64238414027846
50 L-BFGS-B 156.0 1.885624896893773e-09
50 BFGS 845.0 1.2851588507030082e-09
50 trust-constr 273.0 4.8786393752066874e-06
51 COBYLA 1000.0 0.045532842697210224
51 COBYQA 1000.0 0.0025772034067144288
51 SLSQP 321.0 0.003924596822050221
51 Powell 4594.0 1.0674322967281136e-24
51 nelder-mead 1600.0 0.001974655275796391
51 L-BFGS-B 162.0 0.018002244158092647
51 BFGS 540.0 2.5296977794668327e-10
51 trust-constr 837.0 5.2718565480438994e-08
52 COBYLA 1000.0 894783.8530217626
52 COBYQA 848.0 4.5059640773037355
52 SLSQP 996.0 267.22972398525724
52 Powell 5894.0 12.202607426064228
52 nelder-mead 1600.0 223.19648568816407
52 L-BFGS-B 306.0 4.536639136942162
52 BFGS 669.0 4.5348798664538466
52 trust-constr 432.0 4.544389712796571
============================== ============== =========== ========================
The results for BenchSmoothUnbounded:
[ 6.82%] ··· ====================== ============== ==================== ========================
-- result type
------------------------------------- ---------------------------------------------
test function solver mean_nfev mean_time
====================== ============== ==================== ========================
rosenbrock_slow COBYLA 1000.0 0.10221092700958252
rosenbrock_slow COBYQA 152.8 0.41478211879730226
rosenbrock_slow Powell 887.2 0.09182374477386475
rosenbrock_slow nelder-mead 310.2 0.03188121318817139
rosenbrock_slow L-BFGS-B 148.0 0.017220544815063476
rosenbrock_slow BFGS 177.2 0.02521369457244873
rosenbrock_slow CG n/a n/a
rosenbrock_slow TNC n/a n/a
rosenbrock_slow SLSQP 155.1 0.022461724281311036
rosenbrock_slow Newton-CG n/a n/a
rosenbrock_slow dogleg n/a n/a
rosenbrock_slow trust-ncg n/a n/a
rosenbrock_slow trust-exact n/a n/a
rosenbrock_slow trust-krylov n/a n/a
rosenbrock_slow trust-constr 218.0 0.08552446365356445
rosenbrock_nograd COBYLA 1000.0 0.019648361206054687
rosenbrock_nograd COBYQA 152.8 0.39280545711517334
rosenbrock_nograd Powell 887.2 0.016094064712524413
rosenbrock_nograd nelder-mead 310.2 0.007036185264587403
rosenbrock_nograd L-BFGS-B 148.0 0.005884408950805664
rosenbrock_nograd BFGS 177.2 0.009394955635070801
rosenbrock_nograd CG n/a n/a
rosenbrock_nograd TNC n/a n/a
rosenbrock_nograd SLSQP 155.1 0.007205891609191895
rosenbrock_nograd Newton-CG n/a n/a
rosenbrock_nograd dogleg n/a n/a
rosenbrock_nograd trust-ncg n/a n/a
rosenbrock_nograd trust-exact n/a n/a
rosenbrock_nograd trust-krylov n/a n/a
rosenbrock_nograd trust-constr 218.0 0.07210166454315185
rosenbrock COBYLA 1000.0 0.019016504287719727
rosenbrock COBYQA 152.8 0.3881392240524292
rosenbrock Powell 887.2 0.015932464599609376
rosenbrock nelder-mead 310.2 0.006581926345825195
rosenbrock L-BFGS-B 92.75 0.0037383198738098144
rosenbrock BFGS 110.95 0.006753122806549073
rosenbrock CG 110.0 0.008317852020263672
rosenbrock TNC 84.3 0.0030642032623291017
rosenbrock SLSQP 102.3 0.005086040496826172
rosenbrock Newton-CG 64.5 0.007260346412658691
rosenbrock dogleg 18.4 0.00196533203125
rosenbrock trust-ncg 43.5 0.0044361591339111325
rosenbrock trust-exact 17.5 0.0029834747314453126
rosenbrock trust-krylov 35.7 0.004609107971191406
rosenbrock trust-constr 98.16666666666667 0.04931776523590088
rosenbrock_tight COBYLA 1000.0 0.01909213066101074
rosenbrock_tight COBYQA 180.2 0.4743908166885376
rosenbrock_tight Powell 1070.1 0.019453907012939455
rosenbrock_tight nelder-mead 407.7 0.008971905708312989
rosenbrock_tight L-BFGS-B 99.2 0.0039985418319702145
rosenbrock_tight BFGS 160.9 0.009589815139770507
rosenbrock_tight CG 127.8 0.009342217445373535
rosenbrock_tight TNC 90.4 0.003626441955566406
rosenbrock_tight SLSQP 108.15 0.005452108383178711
rosenbrock_tight Newton-CG 71.6 0.00818331241607666
rosenbrock_tight dogleg 19.2 0.002088189125061035
rosenbrock_tight trust-ncg 44.9 0.004395413398742676
rosenbrock_tight trust-exact 18.4 0.0029824018478393556
rosenbrock_tight trust-krylov 93.7 0.008433985710144042
rosenbrock_tight trust-constr 106.53333333333333 0.052931682268778486
simple_quadratic COBYLA 54.8 0.0005712509155273438
simple_quadratic COBYQA 22.2 0.03898210525512695
simple_quadratic Powell 47.2 0.0005118846893310547
simple_quadratic nelder-mead 152.8 0.0022516965866088865
simple_quadratic L-BFGS-B 7.5 0.0003270387649536133
simple_quadratic BFGS 7.5 0.0003994584083557129
simple_quadratic CG 3.6 0.00027928352355957033
simple_quadratic TNC 7.0 0.00021626949310302735
simple_quadratic SLSQP 6.0 0.0003117680549621582
simple_quadratic Newton-CG 2.0 0.0002607107162475586
simple_quadratic dogleg 3.0 0.00029482841491699217
simple_quadratic trust-ncg 3.0 0.00023899078369140624
simple_quadratic trust-exact 3.0 0.00047221183776855467
simple_quadratic trust-krylov 3.0 0.0003118276596069336
simple_quadratic trust-constr 6.0 0.0033010323842366535
asymmetric_quadratic COBYLA 64.2 0.0006988286972045898
asymmetric_quadratic COBYQA 23.5 0.043817758560180664
asymmetric_quadratic Powell 134.6 0.0012090682983398437
asymmetric_quadratic nelder-mead 162.9 0.0022761344909667967
asymmetric_quadratic L-BFGS-B 7.5 0.00037877559661865235
asymmetric_quadratic BFGS 7.5 0.0003869175910949707
asymmetric_quadratic CG 3.1 0.00026428699493408203
asymmetric_quadratic TNC 12.5 0.0002912759780883789
asymmetric_quadratic SLSQP 6.0 0.00032651424407958984
asymmetric_quadratic Newton-CG 2.0 0.0002610206604003906
asymmetric_quadratic dogleg 2.9 0.00026187896728515627
asymmetric_quadratic trust-ncg 2.9 0.0002459287643432617
asymmetric_quadratic trust-exact 2.9 0.0004597187042236328
asymmetric_quadratic trust-krylov 2.9 0.0003061056137084961
asymmetric_quadratic trust-constr 5.966666666666667 0.0032143274943033854
sin_1d COBYLA 24.7 0.0003074169158935547
sin_1d COBYQA 13.7 0.02740190029144287
sin_1d Powell 28.1 0.00037937164306640626
sin_1d nelder-mead 36.8 0.00053558349609375
sin_1d L-BFGS-B 9.45 0.00045931339263916016
sin_1d BFGS 10.2 0.0007171988487243653
sin_1d CG 5.5 0.00034372806549072267
sin_1d TNC 10.2 0.0002444028854370117
sin_1d SLSQP 6.95 0.00047876834869384763
sin_1d Newton-CG n/a n/a
sin_1d dogleg n/a n/a
sin_1d trust-ncg n/a n/a
sin_1d trust-exact n/a n/a
sin_1d trust-krylov n/a n/a
sin_1d trust-constr 9.6 0.006979584693908691
booth COBYLA 67.1 0.0008264064788818359
booth COBYQA 30.9 0.06615271568298339
booth Powell 63.4 0.0007641792297363281
booth nelder-mead 122.9 0.0021659135818481445
booth L-BFGS-B 11.6 0.0005254864692687989
booth BFGS 15.4 0.001001286506652832
booth CG 6.4 0.00041408538818359374
booth TNC 6.8 0.000226593017578125
booth SLSQP 12.3 0.0006320714950561523
booth Newton-CG n/a n/a
booth dogleg n/a n/a
booth trust-ncg n/a n/a
booth trust-exact n/a n/a
booth trust-krylov n/a n/a
booth trust-constr 14.8 0.008171749114990235
beale COBYLA 461.1 0.004681205749511719
beale COBYQA 206.9 0.45821413993835447
beale Powell 1821.9 0.019892191886901854
beale nelder-mead 141.1 0.002144122123718262
beale L-BFGS-B 60.7 0.0021002888679504395
beale BFGS 311.6 0.018624258041381837
beale CG 32.9 0.0018250226974487304
beale TNC 24.1 0.0005386114120483398
beale SLSQP 25.9 0.0011149048805236816
beale Newton-CG n/a n/a
beale dogleg n/a n/a
beale trust-ncg n/a n/a
beale trust-exact n/a n/a
beale trust-krylov n/a n/a
beale trust-constr 476.35 0.2655073642730713
LJ COBYLA 868.8 0.02984781265258789
LJ COBYQA 696.3 4.859420919418335
LJ Powell 1900.0 0.05867128372192383
LJ nelder-mead 2281.8 0.08925626277923585
LJ L-BFGS-B 278.55 0.011390721797943116
LJ BFGS 722.75 0.03499552011489868
LJ CG 146.9 0.01970562934875488
LJ TNC 115.6 0.011374115943908691
LJ SLSQP 208.5 0.00989387035369873
LJ Newton-CG n/a n/a
LJ dogleg n/a n/a
LJ trust-ncg n/a n/a
LJ trust-exact n/a n/a
LJ trust-krylov n/a n/a
LJ trust-constr 1083.05 0.30552700757980344
====================== ============== ==================== ========================
For the numbered problems, the objective function value from COBYQA is always better, and the histogram of function evaluation counts looks better.
It looks like something went awry with the named problems. COBQA's objective value is always worse; it's almost like it's solving a different problem.
It looks like something went awry with the named problems (BenchSmoothUnbounded)? COBQA's objective value is always worse. Or are the other algorithms really always better on smooth problems?
Hi @mdhaber , if you are referring to the second table posted by Tom @ragonneau , you may note that the column is mean_time rather than min_obj.
Ah. It's tough to compare the performance without knowing the objective function values. I can create an algorithm that runs pretty quickly but doesn't solve the problem : ) But I think the first set is sufficient.
Agree. I do not know the SciPy benchmark, I did not run it, and I do not understand why the data in the two tables are different. Maybe Tom @ragonneau should check.
@ragonneau I suggest you edit your results. They are difficult to read. In addition, there seems no reason to use a layout different from the other one. Thanks.
@ragonneau I suggest you edit your results. They are difficult to read. In addition, there seems no reason to use a table different from the other one. Thanks.
I believe the layout of the table was made by asv. When the last column has more than two subcolumns, it creates a column that is difficult to read, as in my previous comment. If an asv expert is in the thread, I'd be happy to learn how to improve that.
In the meantime, I reran the experiment with only mean_nfev and mean_obj:
[100.00%] ··· ====================== ============== ==================== ========================
-- result type
------------------------------------- ---------------------------------------------
test function solver mean_nfev mean_obj
====================== ============== ==================== ========================
rosenbrock_slow COBYLA 1000.0 1.040669274800954
rosenbrock_slow COBYQA 152.8 3.890437153553753e-08
rosenbrock_slow Powell 887.2 1.577792702290088e-23
rosenbrock_slow nelder-mead 310.2 1.948083288418953e-09
rosenbrock_slow L-BFGS-B 148.0 2.395468489328531e-06
rosenbrock_slow BFGS 177.2 4.675037804518409e-11
rosenbrock_slow CG n/a n/a
rosenbrock_slow TNC n/a n/a
rosenbrock_slow SLSQP 155.1 1.0577238854081483e-05
rosenbrock_slow Newton-CG n/a n/a
rosenbrock_slow dogleg n/a n/a
rosenbrock_slow trust-ncg n/a n/a
rosenbrock_slow trust-exact n/a n/a
rosenbrock_slow trust-krylov n/a n/a
rosenbrock_slow trust-constr 218.0 9.51616929722639e-10
rosenbrock_nograd COBYLA 1000.0 1.040669274800954
rosenbrock_nograd COBYQA 152.8 3.890437153553753e-08
rosenbrock_nograd Powell 887.2 1.577792702290088e-23
rosenbrock_nograd nelder-mead 310.2 1.948083288418953e-09
rosenbrock_nograd L-BFGS-B 148.0 2.395468489328531e-06
rosenbrock_nograd BFGS 177.2 4.675037804518409e-11
rosenbrock_nograd CG n/a n/a
rosenbrock_nograd TNC n/a n/a
rosenbrock_nograd SLSQP 155.1 1.0577238854081483e-05
rosenbrock_nograd Newton-CG n/a n/a
rosenbrock_nograd dogleg n/a n/a
rosenbrock_nograd trust-ncg n/a n/a
rosenbrock_nograd trust-exact n/a n/a
rosenbrock_nograd trust-krylov n/a n/a
rosenbrock_nograd trust-constr 218.0 9.51616929722639e-10
rosenbrock COBYLA 1000.0 1.040669274800954
rosenbrock COBYQA 152.8 3.890437153553753e-08
rosenbrock Powell 887.2 1.577792702290088e-23
rosenbrock nelder-mead 310.2 1.948083288418953e-09
rosenbrock L-BFGS-B 92.75 2.1047384326635483e-06
rosenbrock BFGS 110.95 2.5838983525240928e-11
rosenbrock CG 110.0 2.184879891252899e-09
rosenbrock TNC 84.3 0.41353307449601323
rosenbrock SLSQP 102.3 1.2278847681581272e-05
rosenbrock Newton-CG 64.5 0.00028386319748745415
rosenbrock dogleg 18.4 1.0473748283949003e-12
rosenbrock trust-ncg 43.5 9.096654341078131e-10
rosenbrock trust-exact 17.5 2.3688425209734387e-11
rosenbrock trust-krylov 35.7 1.2080822443269492e-09
rosenbrock trust-constr 98.16666666666667 7.582424644998224e-10
rosenbrock_tight COBYLA 1000.0 1.0395001250506932
rosenbrock_tight COBYQA 180.2 1.0833609443328238e-15
rosenbrock_tight Powell 1070.1 5.649905225233614e-24
rosenbrock_tight nelder-mead 407.7 1.7872342256810196e-17
rosenbrock_tight L-BFGS-B 99.2 2.4807079357316176e-10
rosenbrock_tight BFGS 160.9 1.6841960135232725e-11
rosenbrock_tight CG 127.8 5.959714896391864e-19
rosenbrock_tight TNC 90.4 0.41043093540666387
rosenbrock_tight SLSQP 108.15 5.646132513964207e-10
rosenbrock_tight Newton-CG 71.6 8.645626438121326e-19
rosenbrock_tight dogleg 19.2 7.554726540980743e-22
rosenbrock_tight trust-ncg 44.9 6.432714849941195e-24
rosenbrock_tight trust-exact 18.4 9.805261192705719e-20
rosenbrock_tight trust-krylov 93.7 1.227821735143473e-17
rosenbrock_tight trust-constr 106.53333333333333 1.2124258427696534e-11
simple_quadratic COBYLA 54.8 9.518126940933508e-09
simple_quadratic COBYQA 22.2 8.872232632322827e-31
simple_quadratic Powell 47.2 1.2855981106468597e-31
simple_quadratic nelder-mead 152.8 2.2619242867800513e-09
simple_quadratic L-BFGS-B 7.5 2.4856928645318576e-16
simple_quadratic BFGS 7.5 1.6821868699593401e-15
simple_quadratic CG 3.6 3.154879136346819e-10
simple_quadratic TNC 7.0 2.0304535898501623e-07
simple_quadratic SLSQP 6.0 1.336048763662199e-16
simple_quadratic Newton-CG 2.0 0.0
simple_quadratic dogleg 3.0 2.78107340254829e-32
simple_quadratic trust-ncg 3.0 0.0
simple_quadratic trust-exact 3.0 2.76371887415049e-32
simple_quadratic trust-krylov 3.0 0.0
simple_quadratic trust-constr 6.0 9.386350782614601e-17
asymmetric_quadratic COBYLA 64.2 -0.24999998850473912
asymmetric_quadratic COBYQA 23.5 -0.25
asymmetric_quadratic Powell 134.6 -0.25
asymmetric_quadratic nelder-mead 162.9 -0.24999999810601786
asymmetric_quadratic L-BFGS-B 7.5 -0.24999999999999883
asymmetric_quadratic BFGS 7.5 -0.24999999999999784
asymmetric_quadratic CG 3.1 -0.25
asymmetric_quadratic TNC 12.5 -0.24999858212319187
asymmetric_quadratic SLSQP 6.0 -0.24999999999999972
asymmetric_quadratic Newton-CG 2.0 -0.25
asymmetric_quadratic dogleg 2.9 -0.25
asymmetric_quadratic trust-ncg 2.9 -0.25
asymmetric_quadratic trust-exact 2.9 -0.25
asymmetric_quadratic trust-krylov 2.9 -0.25
asymmetric_quadratic trust-constr 5.966666666666667 -0.24999999999999986
sin_1d COBYLA 24.7 -0.9999999947997484
sin_1d COBYQA 13.7 -0.9999999999766593
sin_1d Powell 28.1 -0.9999999999215394
sin_1d nelder-mead 36.8 -0.9999999998495948
sin_1d L-BFGS-B 9.45 -0.9999999936365207
sin_1d BFGS 10.2 -0.9999999996362824
sin_1d CG 5.5 -0.9999999999451038
sin_1d TNC 10.2 -0.9999999999816668
sin_1d SLSQP 6.95 -0.9999874938865606
sin_1d Newton-CG n/a n/a
sin_1d dogleg n/a n/a
sin_1d trust-ncg n/a n/a
sin_1d trust-exact n/a n/a
sin_1d trust-krylov n/a n/a
sin_1d trust-constr 9.6 -0.9999999995302697
booth COBYLA 67.1 9.727602527729466e-08
booth COBYQA 30.9 1.1072140768523674e-09
booth Powell 63.4 2.1062586169401015e-29
booth nelder-mead 122.9 1.219380568865664
booth L-BFGS-B 11.6 2.1341422667527374e-08
booth BFGS 15.4 1.4028567484760426e-10
booth CG 6.4 8.662008283688798e-26
booth TNC 6.8 4.8330558571088694e-15
booth SLSQP 12.3 8.848768281252701e-08
booth Newton-CG n/a n/a
booth dogleg n/a n/a
booth trust-ncg n/a n/a
booth trust-exact n/a n/a
booth trust-krylov n/a n/a
booth trust-constr 14.8 5.502755360622469e-11
beale COBYLA 461.1 1.0645878833664317
beale COBYQA 206.9 1.2864101661603227
beale Powell 1821.9 0.41411155100942254
beale nelder-mead 141.1 3.970084546666909e-10
beale L-BFGS-B 60.7 1.1948511188692428
beale BFGS 311.6 0.13604038140884195
beale CG 32.9 1.5004900363633769
beale TNC 24.1 0.01830610047951784
beale SLSQP 25.9 1.5104357514520017
beale Newton-CG n/a n/a
beale dogleg n/a n/a
beale trust-ncg n/a n/a
beale trust-exact n/a n/a
beale trust-krylov n/a n/a
beale trust-constr 476.35 1.4994513515047054
LJ COBYLA 868.8 -4.290638261230069
LJ COBYQA 696.3 -5.999999907301164
LJ Powell 1900.0 -5.7224243446285765
LJ nelder-mead 2281.8 -3.8851133797531867
LJ L-BFGS-B 278.55 -3.2094944991218455
LJ BFGS 722.75 -5.999999999973245
LJ CG 146.9 -5.999999999936375
LJ TNC 115.6 -5.487077119375702
LJ SLSQP 208.5 -1.9506977348879857
LJ Newton-CG n/a n/a
LJ dogleg n/a n/a
LJ trust-ncg n/a n/a
LJ trust-exact n/a n/a
LJ trust-krylov n/a n/a
LJ trust-constr 1083.05 -4.104444644769407
====================== ============== ==================== ========================
Based on the benchmarks, I would consider COBYQA a great improvement over COBYLA and for our derivative-free solvers in general. As it is implemented in Python, it should not expose extraordinary maintenance burden as well. On how to proceed: as the current trend seems to be git submodules, let's try that here as well?
Based on the benchmarks, I would consider COBYQA a great improvement over COBYLA and for our derivative-free solvers in general. Also it should be noted that currently COBYLA is our only derivative-free local solver that works with constraints. As it is implemented in Python, COBYQA should not cause extraordinary maintenance burden as well. On how to proceed: as the current trend seems to be git submodules, l would advocate for that here as well.
But let's wait for a few more opinions :).
For the moment, we use setuptools to create the package. Would it be useful if we switch to meson?
For the moment, we use setuptools to create the package. Would it be helpful if we switch to meson?
It's pure Python, so it really doesn't matter.
as the current trend seems to be git submodules, l would advocate for that here as well.
It depends a bit on plans I'd say. E.g., is the standalone package expected to continue to be developed for years into the future? Or once it's stable in SciPy, that will become the main implementation?
It depends a bit on plans I'd say. E.g., is the standalone package expected to continue to be developed for years into the future? Or once it's stable in SciPy, that will become the main implementation?
The goal is to maintain cobyqa as a standalone package and to continue to develop it in the future, possibly to add new features, etc. I believe this implies adding it as a submodule in scipy/_lib? I tried on my side; my scipy/optimize/_cobyqa_py.py looks like this:
import importlib.util
import sys
import numpy as np
from ._optimize import _check_unknown_options
__all__ = []
def _minimize_cobyqa(fun, x0, args=(), bounds=None, constraints=(),
callback=None, disp=False, maxfev=1000, maxiter=1000,
target=-np.inf, feasibility_tol=1e-8, radius_init=1.0,
radius_final=1e-6, **unknown_options):
"""
Minimize a scalar function of one or more variables using the
Constrained Optimization BY Quadratic Approximations (COBYQA) algorithm.
Options
-------
disp : bool
Set to True to print information about the optimization procedure.
maxfev : int
Maximum number of function evaluations.
maxiter : int
Maximum number of iterations.
target : float
Target value for the objective function. The optimization procedure is
terminated when the objective function value of a nearly feasible point
is less than or equal to this target.
feasibility_tol : float
Tolerance for the constraint violation.
radius_init : float
Initial trust-region radius. Typically, this value should be in the
order of one tenth of the greatest expected change to the variables.
radius_final : float
Final trust-region radius. It should indicate the accuracy required in
the final values of the variables.
"""
# Import the cobyqa module.
cobyqa_spec = importlib.util.spec_from_file_location(
'cobyqa', '../_lib/cobyqa/cobyqa/__init__.py')
cobyqa_module = importlib.util.module_from_spec(cobyqa_spec)
sys.modules['cobyqa'] = cobyqa_module
cobyqa_spec.loader.exec_module(cobyqa_module)
# Run the COBYQA method.
_check_unknown_options(unknown_options)
options = {
'disp': bool(disp),
'maxfev': int(maxfev),
'maxiter': int(maxiter),
'target': float(target),
'feasibility_tol': float(feasibility_tol),
'radius_init': float(radius_init),
'radius_final': float(radius_final),
}
return cobyqa_module.minimize(fun, x0, args, bounds, constraints, callback, options)
The problem: I am unsure how to modify meson.build so that the Python files in scipy/_lib/cobyqa/cobyqa are included. There should be a cleaner way.
+1 to what Robert said on using relative imports and avoiding importlib. That's not even much of a compromise I'd say; relative imports within a package are idiomatic.
The problem: I am unsure how to modify meson.build so that the Python files in scipy/_lib/cobyqa/cobyqa are included.
I think it only requires something like this at the bottom of scipy/_lib/meson.build:
py3.install_sources(
[
'cobyqa/__init__.py',
'cobyqa/cobyqa.py',
],
subdir: 'scipy/_lib/cobyqa',
)
If there's some kind of hiccup, then I'm happy to help on a PR or branch, please feel free to ping me.
+1 to what Robert said on using relative imports and avoiding importlib. That's not even much of a compromise I'd say; relative imports within a package are idiomatic.
I already pushed an update to the COBYQA repo, removing all the absolute imports.
I think it only requires something like this at the bottom of scipy/_lib/meson.build:
py3.install_sources(
[
'cobyqa/__init__.py',
'cobyqa/cobyqa.py',
],
subdir: 'scipy/_lib/cobyqa',
)
If there's some kind of hiccup, then I'm happy to help on a PR or branch, please feel free to ping me.
Great! I will work next week on a PR, I will contact you if I face any trouble by then. Thanks 👍
I just created a PR. I would be happy to have some feedback and improve it if necessary.
As a side note, on my challenging noisy function COBYQA showed systematically better convergence to global minimum vs COBYLA and all other optimization functions available in scipy. Looking forward to see it merged in!
As a side note, on my challenging noisy function COBYQA showed systematically better convergence to global minimum vs COBYLA and all other optimization functions available in scipy. Looking forward to see it merged in!
That's great to hear! I'm delighted this work is useful!
| gharchive/issue | 2024-01-19T14:54:04 | 2025-04-01T06:45:44.710652 | {
"authors": [
"BarsMonster",
"dschmitz89",
"ev-br",
"ilayn",
"j-bowhay",
"mdhaber",
"ragonneau",
"rgommers",
"tupui",
"zaikunzhang"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/19918",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2397965503 | BUG: signal: crash in signaltools on free-threaded Python, macOS arm64
In the cron wheel builds on main, there was a failure today in the arm64 macOS 14.0 wheel build. The test suite run exited in signal/tests/test_signaltools.py with:
13615 Trace/BPT trap: 5
Full CI log and raw output (doesn't show anything more informative it looks like):
...
2024-07-09T10:14:17.6854550Z + /bin/sh -c 'bash /Users/runner/work/scipy/scipy/tools/wheels/cibw_test_command.sh /Users/runner/work/scipy/scipy'
2024-07-09T10:14:17.6901730Z ++ python '-cimport sysconfig; print(bool(sysconfig.get_config_var('\''Py_GIL_DISABLED'\'')))'
2024-07-09T10:14:17.7121210Z + FREE_THREADED_BUILD=True
2024-07-09T10:14:17.7121520Z + [[ True == \T\r\u\e ]]
2024-07-09T10:14:17.7122050Z + export PYTHON_GIL=0
2024-07-09T10:14:17.7122330Z + PYTHON_GIL=0
2024-07-09T10:14:17.7122800Z + python -c 'import sys; import scipy; sys.exit(not scipy.test())'
2024-07-09T10:14:25.2555010Z ============================= test session starts ==============================
2024-07-09T10:14:25.2556010Z platform darwin -- Python 3.13.0b2, pytest-8.2.2, pluggy-1.5.0
2024-07-09T10:14:25.2556740Z rootdir: /private/var/folders/h9/l1shxhdd69nct08ylzq0n6q00000gn/T/cibw-run-9m42f5g1/cp313t-macosx_arm64/test_cwd
2024-07-09T10:14:25.2557380Z plugins: hypothesis-6.105.1, cov-5.0.0, xdist-3.6.1
2024-07-09T10:14:25.2557760Z collected 64837 items / 11739 deselected / 53098 selected
2024-07-09T10:14:25.2558000Z
2024-07-09T10:14:25.2607920Z _lib/tests/test__gcutils.py ...... [ 0%]
2024-07-09T10:14:25.2682730Z _lib/tests/test__pep440.py ......... [ 0%]
2024-07-09T10:14:25.2700990Z _lib/tests/test__testutils.py .. [ 0%]
2024-07-09T10:14:25.7786640Z _lib/tests/test__threadsafety.py .. [ 0%]
2024-07-09T10:14:26.5096540Z _lib/tests/test__util.py ........................ [ 0%]
2024-07-09T10:14:26.5173400Z _lib/tests/test_array_api.py ssssssssssssss [ 0%]
2024-07-09T10:14:26.5402140Z _lib/tests/test_bunch.py ............................ [ 0%]
2024-07-09T10:14:26.9057650Z _lib/tests/test_ccallback.py ss.. [ 0%]
2024-07-09T10:14:26.9207020Z _lib/tests/test_deprecation.py . [ 0%]
2024-07-09T10:14:27.0253750Z _lib/tests/test_public_api.py .......................................... [ 0%]
2024-07-09T10:14:27.1161830Z ........................................................... [ 0%]
2024-07-09T10:14:27.1171160Z _lib/tests/test_scipy_version.py . [ 0%]
2024-07-09T10:14:27.1208080Z _lib/tests/test_tmpdirs.py ... [ 0%]
2024-07-09T10:14:27.2026800Z cluster/tests/test_disjoint_set.py ..................... [ 0%]
2024-07-09T10:14:27.2528020Z cluster/tests/test_hierarchy.py ........................................ [ 0%]
2024-07-09T10:14:27.3202600Z .........................................sss........... [ 0%]
2024-07-09T10:14:27.7931330Z cluster/tests/test_vq.py ............................... [ 0%]
2024-07-09T10:14:27.7990700Z constants/tests/test_codata.py ....... [ 0%]
2024-07-09T10:14:27.8054610Z constants/tests/test_constants.py ....... [ 0%]
2024-07-09T10:14:28.4682470Z datasets/tests/test_data.py ..... [ 0%]
2024-07-09T10:14:28.5271930Z fft/_pocketfft/tests/test_basic.py ..................................... [ 0%]
2024-07-09T10:14:28.6184940Z ........................................................................ [ 0%]
2024-07-09T10:14:28.6871320Z ........................................................................ [ 1%]
2024-07-09T10:14:28.7571510Z ........................................................................ [ 1%]
2024-07-09T10:14:28.8555130Z ........................................................................ [ 1%]
2024-07-09T10:14:29.1293840Z ........................................................................ [ 1%]
2024-07-09T10:14:29.5032370Z .............................................................. [ 1%]
2024-07-09T10:14:29.5362690Z fft/_pocketfft/tests/test_real_transforms.py ........................... [ 1%]
2024-07-09T10:14:29.6031280Z ........................................................................ [ 1%]
2024-07-09T10:14:29.6720620Z ........................................................................ [ 1%]
2024-07-09T10:14:29.7397130Z ........................................................................ [ 2%]
2024-07-09T10:14:29.8047270Z ........................................................................ [ 2%]
2024-07-09T10:14:30.1155580Z ........................................................................ [ 2%]
2024-07-09T10:14:30.3840320Z ........................................................................ [ 2%]
2024-07-09T10:14:30.5076490Z ........................................................................ [ 2%]
2024-07-09T10:14:30.5846320Z ........................................................................ [ 2%]
2024-07-09T10:14:30.6565160Z ........................................................................ [ 2%]
2024-07-09T10:14:30.7249790Z ........................................................................ [ 2%]
2024-07-09T10:14:30.8180390Z ........................................................................ [ 3%]
2024-07-09T10:14:30.8948810Z ........................................................................ [ 3%]
2024-07-09T10:14:31.2646790Z ........................................................................ [ 3%]
2024-07-09T10:14:31.5765810Z ........................................................................ [ 3%]
2024-07-09T10:14:31.6457940Z ........................................................................ [ 3%]
2024-07-09T10:14:31.7139080Z ........................................................................ [ 3%]
2024-07-09T10:14:31.7755250Z ........................................................................ [ 3%]
2024-07-09T10:14:31.8335770Z ........................................................................ [ 4%]
2024-07-09T10:14:31.8937290Z ........................................................................ [ 4%]
2024-07-09T10:14:31.9552580Z ........................................................................ [ 4%]
2024-07-09T10:14:32.0153930Z ........................................................................ [ 4%]
2024-07-09T10:14:32.0742110Z ........................................................................ [ 4%]
2024-07-09T10:14:32.1342620Z ........................................................................ [ 4%]
2024-07-09T10:14:32.1948590Z ........................................................................ [ 4%]
2024-07-09T10:14:32.2564340Z ........................................................................ [ 4%]
2024-07-09T10:14:32.3222580Z ........................................................................ [ 5%]
2024-07-09T10:14:32.3705760Z ................................................. [ 5%]
2024-07-09T10:14:32.4125240Z fft/tests/test_backend.py .............................................. [ 5%]
2024-07-09T10:14:32.4125710Z [ 5%]
2024-07-09T10:14:32.5495700Z fft/tests/test_basic.py ................................................ [ 5%]
2024-07-09T10:14:32.7200660Z ........................................................................ [ 5%]
2024-07-09T10:14:33.7593160Z ................... [ 5%]
2024-07-09T10:14:33.8039110Z fft/tests/test_fftlog.py ........................................ [ 5%]
2024-07-09T10:14:33.8371100Z fft/tests/test_helper.py ............................... [ 5%]
2024-07-09T10:14:33.9503940Z fft/tests/test_multithreading.py ....................................... [ 5%]
2024-07-09T10:14:33.9837950Z ................ [ 5%]
2024-07-09T10:14:34.0240480Z fft/tests/test_real_transforms.py ...................................... [ 5%]
2024-07-09T10:14:34.0973000Z ........................................................................ [ 6%]
2024-07-09T10:14:34.1705940Z ........................................................................ [ 6%]
2024-07-09T10:14:34.2434150Z ........................................................................ [ 6%]
2024-07-09T10:14:34.3157850Z ........................................................................ [ 6%]
2024-07-09T10:14:34.3873810Z ........................................................................ [ 6%]
2024-07-09T10:14:34.4592810Z ........................................................................ [ 6%]
2024-07-09T10:14:34.5306500Z ........................................................................ [ 6%]
2024-07-09T10:14:34.6022770Z ........................................................................ [ 6%]
2024-07-09T10:14:34.6741650Z ........................................................................ [ 7%]
2024-07-09T10:14:34.7459530Z ........................................................................ [ 7%]
2024-07-09T10:14:34.8142910Z ........................................................................ [ 7%]
2024-07-09T10:14:34.8838600Z ........................................................................ [ 7%]
2024-07-09T10:14:34.9501190Z ........................................................................ [ 7%]
2024-07-09T10:14:35.0187220Z ........................................................................ [ 7%]
2024-07-09T10:14:35.0876930Z ........................................................................ [ 7%]
2024-07-09T10:14:35.1595300Z ........................................................................ [ 8%]
2024-07-09T10:14:35.2356300Z ........................................................................ [ 8%]
2024-07-09T10:14:35.3065520Z ........................................................................ [ 8%]
2024-07-09T10:14:35.3771920Z ........................................................................ [ 8%]
2024-07-09T10:14:35.4550810Z ........................................................................ [ 8%]
2024-07-09T10:14:35.5306390Z ........................................................................ [ 8%]
2024-07-09T10:14:35.6093410Z ........................................................................ [ 8%]
2024-07-09T10:14:35.6897450Z ........................................................................ [ 8%]
2024-07-09T10:14:35.7699340Z ........................................................................ [ 9%]
2024-07-09T10:14:35.8533310Z ........................................................................ [ 9%]
2024-07-09T10:14:35.9350320Z ........................................................................ [ 9%]
2024-07-09T10:14:36.0177350Z ........................................................................ [ 9%]
2024-07-09T10:14:36.0906920Z ........................................................................ [ 9%]
2024-07-09T10:14:36.1640020Z ........................................................................ [ 9%]
2024-07-09T10:14:36.2412880Z ........................................................................ [ 9%]
2024-07-09T10:14:36.3201230Z ........................................................................ [ 10%]
2024-07-09T10:14:36.3957790Z ........................................................................ [ 10%]
2024-07-09T10:14:36.4712290Z ........................................................................ [ 10%]
2024-07-09T10:14:36.5504580Z ........................................................................ [ 10%]
2024-07-09T10:14:36.6275450Z ........................................................................ [ 10%]
2024-07-09T10:14:36.6996650Z ........................................................................ [ 10%]
2024-07-09T10:14:36.7691050Z ........................................................................ [ 10%]
2024-07-09T10:14:36.8371680Z ........................................................................ [ 11%]
2024-07-09T10:14:36.9076270Z ........................................................................ [ 11%]
2024-07-09T10:14:36.9742240Z ........................................................................ [ 11%]
2024-07-09T10:14:37.0389130Z ........................................................................ [ 11%]
2024-07-09T10:14:37.0908880Z .......................................................... [ 11%]
2024-07-09T10:14:37.2353020Z fftpack/tests/test_basic.py ......x..................................... [ 11%]
2024-07-09T10:14:37.3292180Z ........................................................................ [ 11%]
2024-07-09T10:14:37.4028710Z ........................................................................ [ 11%]
2024-07-09T10:14:37.6076430Z ........................................................................ [ 12%]
2024-07-09T10:14:37.8148540Z .................................... [ 12%]
2024-07-09T10:14:37.8188790Z fftpack/tests/test_helper.py .... [ 12%]
2024-07-09T10:14:37.8864940Z fftpack/tests/test_pseudo_diffs.py ............................. [ 12%]
2024-07-09T10:14:38.1308650Z fftpack/tests/test_real_transforms.py .................................. [ 12%]
2024-07-09T10:14:38.8570470Z ........................................................................ [ 12%]
2024-07-09T10:14:38.9197940Z ........................................................................ [ 12%]
2024-07-09T10:14:38.9647940Z .............................................. [ 12%]
2024-07-09T10:14:39.5883850Z integrate/_ivp/tests/test_ivp.py ....................................... [ 12%]
2024-07-09T10:14:39.7089890Z ............... [ 12%]
2024-07-09T10:14:39.7182320Z integrate/_ivp/tests/test_rk.py .......... [ 12%]
2024-07-09T10:14:41.5625470Z integrate/tests/test__quad_vec.py .......................... [ 12%]
2024-07-09T10:14:41.6248260Z integrate/tests/test_banded_ode_solvers.py . [ 12%]
2024-07-09T10:14:41.7764800Z integrate/tests/test_bvp.py ................ [ 12%]
2024-07-09T10:14:41.8280720Z integrate/tests/test_integrate.py ...................................... [ 12%]
2024-07-09T10:14:41.8478350Z ......... [ 12%]
2024-07-09T10:14:41.8761530Z integrate/tests/test_odeint_jac.py .. [ 12%]
2024-07-09T10:14:42.0584870Z integrate/tests/test_quadpack.py ....................................... [ 12%]
2024-07-09T10:14:43.0279540Z ..sssssssssssssssssssssssssssssssssssssssssss........... [ 13%]
2024-07-09T10:14:43.1334810Z integrate/tests/test_quadrature.py ..................................... [ 13%]
2024-07-09T10:14:43.2340040Z ........................................................................ [ 13%]
2024-07-09T10:14:43.3278030Z ........................................................................ [ 13%]
2024-07-09T10:14:43.4222890Z ........................................................................ [ 13%]
2024-07-09T10:14:43.4286090Z ....... [ 13%]
2024-07-09T10:14:43.5115680Z integrate/tests/test_tanhsinh.py ....................................... [ 13%]
2024-07-09T10:14:43.6967280Z .........ssss........................................................... [ 13%]
2024-07-09T10:14:44.0721280Z .............................ss..ss......ss............................. [ 13%]
2024-07-09T10:14:44.2780360Z .................................ss..................................... [ 14%]
2024-07-09T10:14:44.4808950Z ........................................ [ 14%]
2024-07-09T10:14:44.5441160Z interpolate/tests/test_bsplines.py ..................................... [ 14%]
2024-07-09T10:14:44.6400190Z ........................................................................ [ 14%]
2024-07-09T10:14:45.0000690Z .......................x................................................ [ 14%]
2024-07-09T10:14:45.0590470Z .............................. [ 14%]
2024-07-09T10:14:45.1476160Z interpolate/tests/test_fitpack.py .....................................s [ 14%]
2024-07-09T10:14:45.1579000Z ....s.. [ 14%]
2024-07-09T10:14:45.3876110Z interpolate/tests/test_fitpack2.py ..................................... [ 14%]
2024-07-09T10:14:45.5359100Z ................................ [ 14%]
2024-07-09T10:14:45.5366470Z interpolate/tests/test_gil.py s [ 14%]
2024-07-09T10:14:45.6146460Z interpolate/tests/test_interpnd.py ..................... [ 14%]
2024-07-09T10:14:45.7301320Z interpolate/tests/test_interpolate.py .................................. [ 14%]
2024-07-09T10:14:46.0830570Z ........................................................................ [ 14%]
2024-07-09T10:14:46.4576170Z ........................ [ 14%]
2024-07-09T10:14:46.4888780Z interpolate/tests/test_ndgriddata.py .................... [ 15%]
2024-07-09T10:14:46.4956010Z interpolate/tests/test_pade.py .... [ 15%]
2024-07-09T10:14:46.7352530Z interpolate/tests/test_polyint.py ...................................... [ 15%]
2024-07-09T10:14:46.8566020Z ........................................................................ [ 15%]
2024-07-09T10:14:46.8740440Z ................. [ 15%]
2024-07-09T10:14:46.8957080Z interpolate/tests/test_rbf.py .......... [ 15%]
2024-07-09T10:14:47.0738270Z interpolate/tests/test_rbfinterp.py .................................... [ 15%]
2024-07-09T10:14:47.2650430Z ........................................................................ [ 15%]
2024-07-09T10:14:47.5113650Z ........................................................................ [ 15%]
2024-07-09T10:14:47.5241690Z ..... [ 15%]
2024-07-09T10:14:48.0257510Z interpolate/tests/test_rgi.py ....................s..................... [ 15%]
2024-07-09T10:14:49.7347350Z ..................................ss.......ss..................s.......s [ 15%]
2024-07-09T10:14:50.5596830Z s.......ss............................................s.......ss........ [ 15%]
2024-07-09T10:14:50.6196080Z ..s......................... [ 16%]
2024-07-09T10:14:50.6303250Z io/_harwell_boeing/tests/test_fortran_format.py ........... [ 16%]
2024-07-09T10:14:50.6412170Z io/_harwell_boeing/tests/test_hb.py .. [ 16%]
2024-07-09T10:14:50.8372350Z io/arff/tests/test_arffread.py ........................... [ 16%]
2024-07-09T10:14:50.8392440Z io/matlab/tests/test_byteordercodes.py .. [ 16%]
2024-07-09T10:14:50.9405920Z io/matlab/tests/test_mio.py ............................................ [ 16%]
2024-07-09T10:14:50.9769900Z ............................ [ 16%]
2024-07-09T10:14:50.9829330Z io/matlab/tests/test_mio5_utils.py ...... [ 16%]
2024-07-09T10:14:50.9840530Z io/matlab/tests/test_mio_funcs.py . [ 16%]
2024-07-09T10:14:50.9857970Z io/matlab/tests/test_mio_utils.py .. [ 16%]
2024-07-09T10:14:50.9872670Z io/matlab/tests/test_miobase.py . [ 16%]
2024-07-09T10:14:50.9899830Z io/matlab/tests/test_pathological.py .. [ 16%]
2024-07-09T10:14:51.6076130Z io/matlab/tests/test_streams.py ........... [ 16%]
2024-07-09T10:14:51.6357840Z io/tests/test_fortran.py ........... [ 16%]
2024-07-09T10:14:51.6975860Z io/tests/test_idl.py ................................................... [ 16%]
2024-07-09T10:14:51.7507160Z ................ [ 16%]
2024-07-09T10:14:51.8459390Z io/tests/test_mmio.py .................................................. [ 16%]
2024-07-09T10:14:52.0525170Z ........................................................................ [ 16%]
2024-07-09T10:14:52.1781320Z .............................. [ 16%]
2024-07-09T10:14:52.2297610Z io/tests/test_netcdf.py ......................... [ 16%]
2024-07-09T10:14:52.2471830Z io/tests/test_paths.py ........... [ 16%]
2024-07-09T10:14:52.3103000Z io/tests/test_wavfile.py ................s.............................. [ 16%]
2024-07-09T10:14:52.4051870Z ........................................................................ [ 17%]
2024-07-09T10:14:52.5090840Z ........................................................................ [ 17%]
2024-07-09T10:14:52.6753370Z ........................................................................ [ 17%]
2024-07-09T10:14:52.7456120Z .......................... [ 17%]
2024-07-09T10:14:52.7938530Z linalg/tests/test_basic.py ............................................. [ 17%]
2024-07-09T10:14:52.8787530Z ................................................................s....... [ 17%]
2024-07-09T10:14:52.9472940Z ........................................................................ [ 17%]
2024-07-09T10:14:53.4837720Z ........................................................................ [ 17%]
2024-07-09T10:14:54.3404830Z ........................................................................ [ 17%]
2024-07-09T10:14:54.4335140Z ........................................s............................... [ 18%]
2024-07-09T10:14:54.4576590Z ........................... [ 18%]
2024-07-09T10:14:54.5152150Z linalg/tests/test_blas.py .............................................. [ 18%]
2024-07-09T10:14:54.5625230Z .....x. [ 18%]
2024-07-09T10:14:54.5705570Z linalg/tests/test_cython_blas.py ...... [ 18%]
2024-07-09T10:14:54.5738780Z linalg/tests/test_cython_lapack.py ... [ 18%]
2024-07-09T10:14:54.5977860Z linalg/tests/test_cythonized_array_utils.py ............................ [ 18%]
2024-07-09T10:14:54.6176280Z ...................... [ 18%]
2024-07-09T10:14:54.7476860Z linalg/tests/test_decomp.py ............................................ [ 18%]
2024-07-09T10:14:54.8324210Z ........................................................................ [ 18%]
2024-07-09T10:14:54.9226810Z ........................................................................ [ 18%]
2024-07-09T10:14:58.3365110Z ........................................................................ [ 18%]
2024-07-09T10:14:58.4764580Z ........................................................................ [ 19%]
2024-07-09T10:14:58.5902640Z ........................................................................ [ 19%]
2024-07-09T10:14:59.3769020Z ..............................................x......................... [ 19%]
2024-07-09T10:14:59.4114420Z ................ [ 19%]
2024-07-09T10:14:59.4436090Z linalg/tests/test_decomp_cholesky.py .....s............................. [ 19%]
2024-07-09T10:14:59.4787890Z .................................... [ 19%]
2024-07-09T10:14:59.5673790Z linalg/tests/test_decomp_cossin.py ..................................... [ 19%]
2024-07-09T10:14:59.7591640Z ...................................................... [ 19%]
2024-07-09T10:14:59.8045200Z linalg/tests/test_decomp_ldl.py ............ [ 19%]
2024-07-09T10:14:59.8499530Z linalg/tests/test_decomp_lu.py ......................................... [ 19%]
2024-07-09T10:14:59.8973490Z ................................................... [ 19%]
2024-07-09T10:14:59.9048490Z linalg/tests/test_decomp_polar.py .. [ 19%]
2024-07-09T10:14:59.9945930Z linalg/tests/test_decomp_update.py ..................................... [ 19%]
2024-07-09T10:15:00.1746640Z ........................................................................ [ 20%]
2024-07-09T10:15:00.3120800Z ........................................................................ [ 20%]
2024-07-09T10:15:00.4651120Z ........................................................................ [ 20%]
2024-07-09T10:15:00.6254490Z ........................................................................ [ 20%]
2024-07-09T10:15:00.8040730Z ........................................................................ [ 20%]
2024-07-09T10:15:00.9568510Z ........................................................................ [ 20%]
2024-07-09T10:15:01.1552150Z ........................................................................ [ 20%]
2024-07-09T10:15:01.3566930Z ........................................................................ [ 20%]
2024-07-09T10:15:01.4763820Z .................................................... [ 21%]
2024-07-09T10:15:06.5056140Z linalg/tests/test_extending.py . [ 21%]
2024-07-09T10:15:06.5427880Z linalg/tests/test_fblas.py ............................................. [ 21%]
2024-07-09T10:15:06.6780290Z ........................................................................ [ 21%]
2024-07-09T10:15:06.6882240Z ............ [ 21%]
2024-07-09T10:15:07.5339200Z linalg/tests/test_interpolative.py ..................................... [ 21%]
2024-07-09T10:15:07.5444960Z .......... [ 21%]
2024-07-09T10:15:07.5941610Z linalg/tests/test_lapack.py ............................................ [ 21%]
2024-07-09T10:15:07.6923310Z ....................................................................s... [ 21%]
2024-07-09T10:15:07.7727540Z ........................................................................ [ 21%]
2024-07-09T10:15:07.8527350Z ........................................................................ [ 21%]
2024-07-09T10:15:07.9207110Z ........................................................................ [ 22%]
2024-07-09T10:15:07.9951510Z ........................................................................ [ 22%]
2024-07-09T10:15:08.0691010Z ........................................................................ [ 22%]
2024-07-09T10:15:08.1576350Z ........................................................................ [ 22%]
2024-07-09T10:15:08.2409530Z ........................................................................ [ 22%]
2024-07-09T10:15:08.3137500Z ........................................................................ [ 22%]
2024-07-09T10:15:08.3907690Z ........................................................................ [ 22%]
2024-07-09T10:15:08.4672670Z ........................................................................ [ 22%]
2024-07-09T10:15:08.5464580Z ........................................................................ [ 23%]
2024-07-09T10:15:08.6297740Z ........................................................................ [ 23%]
2024-07-09T10:15:08.7047530Z ........................................................................ [ 23%]
2024-07-09T10:15:08.7771680Z ........................................................................ [ 23%]
2024-07-09T10:15:08.8509100Z ........................................................................ [ 23%]
2024-07-09T10:15:08.9244920Z ........................................................................ [ 23%]
2024-07-09T10:15:09.0135440Z ........................................................................ [ 23%]
2024-07-09T10:15:09.0994090Z ........................................................................ [ 24%]
2024-07-09T10:15:09.1688880Z ........................................................................ [ 24%]
2024-07-09T10:15:09.2433480Z ........................................................................ [ 24%]
2024-07-09T10:15:09.3176420Z ........................................................................ [ 24%]
2024-07-09T10:15:09.3871150Z ........................................................................ [ 24%]
2024-07-09T10:15:09.4503830Z ........................................................................ [ 24%]
2024-07-09T10:15:09.6009720Z ........................................................................ [ 24%]
2024-07-09T10:15:09.6778920Z ........................................................................ [ 25%]
2024-07-09T10:15:09.6901010Z ........... [ 25%]
2024-07-09T10:15:10.1008090Z linalg/tests/test_matfuncs.py ................xxxxx..............x...... [ 25%]
2024-07-09T10:15:10.7697250Z .............X........................ [ 25%]
2024-07-09T10:15:10.7775130Z linalg/tests/test_matmul_toeplitz.py .... [ 25%]
2024-07-09T10:15:10.7912770Z linalg/tests/test_procrustes.py ........... [ 25%]
2024-07-09T10:15:11.0652420Z linalg/tests/test_sketches.py ..... [ 25%]
2024-07-09T10:15:11.0950730Z linalg/tests/test_solve_toeplitz.py ......X......................... [ 25%]
2024-07-09T10:15:11.3559020Z linalg/tests/test_solvers.py ........x....x....x.xx..................... [ 25%]
2024-07-09T10:15:11.3629790Z .... [ 25%]
2024-07-09T10:15:11.4569850Z linalg/tests/test_special_matrices.py .................................. [ 25%]
2024-07-09T10:15:11.5297410Z ........................................................................ [ 25%]
2024-07-09T10:15:11.5357540Z ...... [ 25%]
2024-07-09T10:15:11.6376640Z misc/tests/test_common.py ... [ 25%]
2024-07-09T10:15:11.6426770Z misc/tests/test_config.py .... [ 25%]
2024-07-09T10:15:11.6471870Z misc/tests/test_doccer.py ..... [ 25%]
2024-07-09T10:15:11.6554370Z ndimage/tests/test_c_api.py ... [ 25%]
2024-07-09T10:15:11.7703620Z ndimage/tests/test_datatypes.py .x [ 25%]
2024-07-09T10:15:11.8099170Z ndimage/tests/test_filters.py .......................................... [ 25%]
2024-07-09T10:15:11.8746060Z ........................................................................ [ 25%]
2024-07-09T10:15:11.9408870Z ........................................................................ [ 25%]
2024-07-09T10:15:12.0060500Z ........................................................................ [ 26%]
2024-07-09T10:15:12.0687140Z ........................................................................ [ 26%]
2024-07-09T10:15:12.1318650Z ........................................................................ [ 26%]
2024-07-09T10:15:12.1954470Z ........................................................................ [ 26%]
2024-07-09T10:15:12.2615740Z ........................................................................ [ 26%]
2024-07-09T10:15:12.3250940Z ........................................................................ [ 26%]
2024-07-09T10:15:12.3928960Z ........................................................................ [ 26%]
2024-07-09T10:15:12.4606520Z ........................................................................ [ 27%]
2024-07-09T10:15:12.5452470Z ........................................................................ [ 27%]
2024-07-09T10:15:12.6220390Z ........................................................................ [ 27%]
2024-07-09T10:15:12.7023650Z ........................................................................ [ 27%]
2024-07-09T10:15:12.7842420Z ........................................................................ [ 27%]
2024-07-09T10:15:12.8742090Z ........................................................................ [ 27%]
2024-07-09T10:15:12.9463760Z ........................................................................ [ 27%]
2024-07-09T10:15:13.0147990Z ........................................................................ [ 27%]
2024-07-09T10:15:13.0845290Z ........................................................................ [ 28%]
2024-07-09T10:15:13.1497550Z ........................................................................ [ 28%]
2024-07-09T10:15:13.2107280Z ........................................................................ [ 28%]
2024-07-09T10:15:13.2720800Z ........................................................................ [ 28%]
2024-07-09T10:15:13.3373750Z ........................................................................ [ 28%]
2024-07-09T10:15:13.4042240Z ........................................................................ [ 28%]
2024-07-09T10:15:13.4707980Z ........................................................................ [ 28%]
2024-07-09T10:15:13.5378050Z ........................................................................ [ 29%]
2024-07-09T10:15:13.6017690Z ........................................................................ [ 29%]
2024-07-09T10:15:14.0162760Z ......................................................... [ 29%]
2024-07-09T10:15:14.0562490Z ndimage/tests/test_fourier.py .......................................... [ 29%]
2024-07-09T10:15:14.0848060Z .................................. [ 29%]
2024-07-09T10:15:14.1189730Z ndimage/tests/test_interpolation.py .................................... [ 29%]
2024-07-09T10:15:14.1818590Z ........................................................................ [ 29%]
2024-07-09T10:15:14.2420070Z ........................................................................ [ 29%]
2024-07-09T10:15:14.3013030Z ........................................................................ [ 29%]
2024-07-09T10:15:14.3611000Z ........................................................................ [ 30%]
2024-07-09T10:15:14.4311920Z ........................................................................ [ 30%]
2024-07-09T10:15:14.5121480Z ........................................................................ [ 30%]
2024-07-09T10:15:14.5722300Z ........................................................................ [ 30%]
2024-07-09T10:15:14.6336240Z ........................................................................ [ 30%]
2024-07-09T10:15:14.6953700Z ........................................................................ [ 30%]
2024-07-09T10:15:14.7701420Z ........................................................................ [ 30%]
2024-07-09T10:15:14.8486430Z ........................................................................ [ 31%]
2024-07-09T10:15:14.9185810Z ........................................................................ [ 31%]
2024-07-09T10:15:14.9838040Z ........................................................................ [ 31%]
2024-07-09T10:15:14.9980120Z ................ [ 31%]
2024-07-09T10:15:15.1718370Z ndimage/tests/test_measurements.py ..................................... [ 31%]
2024-07-09T10:15:15.2533650Z ........................................................................ [ 31%]
2024-07-09T10:15:15.2672510Z ................ [ 31%]
2024-07-09T10:15:15.3084360Z ndimage/tests/test_morphology.py ....................................... [ 31%]
2024-07-09T10:15:15.3795670Z ........................................................................ [ 31%]
2024-07-09T10:15:15.4443610Z ........................................................................ [ 31%]
2024-07-09T10:15:15.5042860Z ........................................................................ [ 32%]
2024-07-09T10:15:15.5675510Z ........................................................................ [ 32%]
2024-07-09T10:15:15.6323520Z ........................................................................ [ 32%]
2024-07-09T10:15:15.6942970Z ........................................................................ [ 32%]
2024-07-09T10:15:15.7532250Z ........................................................................ [ 32%]
2024-07-09T10:15:15.8135430Z ........................................................................ [ 32%]
2024-07-09T10:15:15.8747790Z ........................................................................ [ 32%]
2024-07-09T10:15:15.9362540Z ........................................................................ [ 32%]
2024-07-09T10:15:15.9974380Z ........................................................................ [ 33%]
2024-07-09T10:15:16.0146030Z ................. [ 33%]
2024-07-09T10:15:16.0230610Z ndimage/tests/test_ni_support.py .......... [ 33%]
2024-07-09T10:15:16.0489620Z ndimage/tests/test_splines.py .................. [ 33%]
2024-07-09T10:15:16.0801220Z odr/tests/test_odr.py ................s.. [ 33%]
2024-07-09T10:15:16.0919140Z optimize/_trustregion_constr/tests/test_canonical_constraint.py ...... [ 33%]
2024-07-09T10:15:16.1229470Z optimize/_trustregion_constr/tests/test_projections.py .......... [ 33%]
2024-07-09T10:15:16.1436180Z optimize/_trustregion_constr/tests/test_qp_subproblem.py ............... [ 33%]
2024-07-09T10:15:16.1563350Z ........ [ 33%]
2024-07-09T10:15:16.1774570Z optimize/_trustregion_constr/tests/test_report.py .s [ 33%]
2024-07-09T10:15:18.1183910Z optimize/tests/test__basinhopping.py ................................. [ 33%]
2024-07-09T10:15:18.4623840Z optimize/tests/test__differential_evolution.py ......................... [ 33%]
2024-07-09T10:15:22.9791320Z .............................s....... [ 33%]
2024-07-09T10:15:25.3800790Z optimize/tests/test__dual_annealing.py ................................. [ 33%]
2024-07-09T10:15:26.8269010Z ............. [ 33%]
2024-07-09T10:15:26.8392760Z optimize/tests/test__linprog_clean_inputs.py ............. [ 33%]
2024-07-09T10:15:26.9077750Z optimize/tests/test__numdiff.py ................................ [ 33%]
2024-07-09T10:15:26.9533350Z optimize/tests/test__remove_redundancy.py .............................. [ 33%]
2024-07-09T10:15:27.0364490Z ...................................... [ 33%]
2024-07-09T10:15:27.0545480Z optimize/tests/test__root.py ............... [ 33%]
2024-07-09T10:15:28.4213460Z optimize/tests/test__shgo.py ..................ss...................s... [ 33%]
2024-07-09T10:15:29.7282670Z ............................ [ 33%]
2024-07-09T10:15:29.8347330Z optimize/tests/test__spectral.py .... [ 33%]
2024-07-09T10:15:29.8976010Z optimize/tests/test_bracket.py ......................................... [ 34%]
2024-07-09T10:15:30.0832880Z ........................................................................ [ 34%]
2024-07-09T10:15:30.1558930Z ........................................................................ [ 34%]
2024-07-09T10:15:30.2331280Z ........................................................................ [ 34%]
2024-07-09T10:15:30.3187730Z ........................................................................ [ 34%]
2024-07-09T10:15:30.6099050Z ........................................................................ [ 34%]
2024-07-09T10:15:30.7122890Z ................................................................... [ 34%]
2024-07-09T10:15:31.1437480Z optimize/tests/test_chandrupatla.py .................................... [ 34%]
2024-07-09T10:15:31.3160680Z ........................................................................ [ 35%]
2024-07-09T10:15:31.3628760Z ........................... [ 35%]
2024-07-09T10:15:31.3758290Z optimize/tests/test_cobyla.py s..... [ 35%]
2024-07-09T10:15:31.7728800Z optimize/tests/test_cobyqa.py ........ [ 35%]
2024-07-09T10:15:32.7723680Z optimize/tests/test_constraint_conversion.py ........ [ 35%]
2024-07-09T10:15:32.7888780Z optimize/tests/test_constraints.py ................ [ 35%]
2024-07-09T10:15:32.7930720Z optimize/tests/test_cython_optimize.py ..... [ 35%]
2024-07-09T10:15:32.8319130Z optimize/tests/test_differentiable_functions.py ..............x. [ 35%]
2024-07-09T10:15:32.9012880Z optimize/tests/test_differentiate.py ................................... [ 35%]
2024-07-09T10:15:33.2916880Z ........................................................................ [ 35%]
2024-07-09T10:15:33.4273250Z ...............................................XX................... [ 35%]
2024-07-09T10:15:33.5483090Z optimize/tests/test_direct.py ....................................ss.... [ 35%]
2024-07-09T10:15:33.5655220Z ..................... [ 35%]
2024-07-09T10:15:36.2213890Z optimize/tests/test_extending.py . [ 35%]
2024-07-09T10:15:36.2360910Z optimize/tests/test_hessian_update_strategy.py ..... [ 35%]
2024-07-09T10:15:36.2642970Z optimize/tests/test_isotonic_regression.py ............................. [ 35%]
2024-07-09T10:15:36.2786870Z ................ [ 35%]
2024-07-09T10:15:36.3305210Z optimize/tests/test_lbfgsb_hessinv.py .. [ 35%]
2024-07-09T10:15:36.3327930Z optimize/tests/test_lbfgsb_setulb.py .. [ 35%]
2024-07-09T10:15:36.9080380Z optimize/tests/test_least_squares.py ................................... [ 35%]
2024-07-09T10:15:38.1155430Z ........................................................................ [ 35%]
2024-07-09T10:15:38.1291890Z ........... [ 35%]
2024-07-09T10:15:38.1900150Z optimize/tests/test_linear_assignment.py ...................... [ 36%]
2024-07-09T10:15:38.2060650Z optimize/tests/test_linesearch.py ........... [ 36%]
2024-07-09T10:15:38.4155820Z optimize/tests/test_linprog.py ......................................... [ 36%]
2024-07-09T10:15:38.6611000Z .......................................s.ss............................. [ 36%]
2024-07-09T10:15:39.0405520Z ...............................................s.s...................... [ 36%]
2024-07-09T10:15:39.3690070Z ....................................................s...s.ss............ [ 36%]
2024-07-09T10:15:39.4870250Z ................................................................ss...... [ 36%]
2024-07-09T10:15:39.8017070Z .....................................................................ssX [ 36%]
2024-07-09T10:15:40.2077170Z ........................................................................ [ 36%]
2024-07-09T10:15:40.3654850Z ....sssX................................................................ [ 37%]
2024-07-09T10:15:40.4783920Z .............s.ssss..................................................... [ 37%]
2024-07-09T10:15:40.6751150Z ..........................s.sss..............................x.......... [ 37%]
2024-07-09T10:15:40.9290110Z ..................................s..................................... [ 37%]
2024-07-09T10:15:42.1105030Z ........x............................................s.................. [ 37%]
2024-07-09T10:15:42.1674890Z .s........................... [ 37%]
2024-07-09T10:15:42.1775110Z optimize/tests/test_lsq_common.py .......... [ 37%]
2024-07-09T10:15:43.1575130Z optimize/tests/test_lsq_linear.py ..........s........s... [ 37%]
2024-07-09T10:15:43.2816280Z optimize/tests/test_milp.py .x.........s [ 37%]
2024-07-09T10:15:43.5628820Z optimize/tests/test_minimize_constrained.py .....................s...... [ 37%]
2024-07-09T10:15:44.1901520Z ..........s.............................ssssssssssssssssssssssssssssssss [ 37%]
2024-07-09T10:15:45.1899620Z ss.....................s................s............................... [ 38%]
2024-07-09T10:15:46.3663690Z ..s.......x........s.................................s.......x........s. [ 38%]
2024-07-09T10:15:46.7592070Z .....................x........X [ 38%]
2024-07-09T10:15:46.8312760Z optimize/tests/test_minpack.py ......................................... [ 38%]
2024-07-09T10:15:46.9424900Z .................................................... [ 38%]
2024-07-09T10:15:46.9556820Z optimize/tests/test_nnls.py ....... [ 38%]
2024-07-09T10:15:47.1980860Z optimize/tests/test_nonlin.py ....................................... [ 38%]
2024-07-09T10:15:47.6160690Z optimize/tests/test_optimize.py .........................x.............. [ 38%]
2024-07-09T10:15:48.1526430Z ...........x.........................x.........................x........ [ 38%]
2024-07-09T10:15:49.6752560Z .....................................................x.................. [ 38%]
2024-07-09T10:15:50.1991420Z ........................................................................ [ 39%]
2024-07-09T10:15:50.3351140Z ........................................................................ [ 39%]
2024-07-09T10:15:58.4262230Z ......................................................s.ss.ss.ss.ss.ss.s [ 39%]
2024-07-09T10:15:58.5769120Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s [ 39%]
2024-07-09T10:15:58.7756750Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s.................................... [ 39%]
2024-07-09T10:15:58.9463060Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s [ 39%]
2024-07-09T10:15:59.2881390Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s.................. [ 39%]
2024-07-09T10:15:59.3500620Z .... [ 39%]
2024-07-09T10:15:59.3950930Z optimize/tests/test_quadratic_assignment.py .............. [ 39%]
2024-07-09T10:15:59.3976070Z optimize/tests/test_regression.py ... [ 39%]
2024-07-09T10:15:59.4530400Z optimize/tests/test_slsqp.py ..................................... [ 39%]
2024-07-09T10:15:59.5016050Z optimize/tests/test_tnc.py .................... [ 39%]
2024-07-09T10:15:59.5339520Z optimize/tests/test_trustregion.py ........ [ 39%]
2024-07-09T10:15:59.8097390Z optimize/tests/test_trustregion_exact.py .......... [ 40%]
2024-07-09T10:15:59.8149600Z optimize/tests/test_trustregion_krylov.py ..... [ 40%]
2024-07-09T10:15:59.8613950Z optimize/tests/test_zeros.py ........................................... [ 40%]
2024-07-09T10:16:00.0879160Z ........................................................................ [ 40%]
2024-07-09T10:16:00.1109110Z ...................... [ 40%]
2024-07-09T10:16:00.1167350Z signal/tests/test_array_tools.py ...... [ 40%]
2024-07-09T10:16:00.1469750Z signal/tests/test_bsplines.py ......................XXXXXXX.. [ 40%]
2024-07-09T10:16:00.1739540Z signal/tests/test_cont2discrete.py ........................ [ 40%]
2024-07-09T10:16:00.2598200Z signal/tests/test_czt.py ............................................... [ 40%]
2024-07-09T10:16:00.3353940Z ........................................................................ [ 40%]
2024-07-09T10:16:00.3776900Z .......................................... [ 40%]
2024-07-09T10:16:00.4092980Z signal/tests/test_dltisys.py ................................ [ 40%]
2024-07-09T10:16:00.4527410Z signal/tests/test_filter_design.py ..................................... [ 40%]
2024-07-09T10:16:00.7298940Z ........................................................................ [ 40%]
2024-07-09T10:16:01.8037600Z ........................................................................ [ 41%]
2024-07-09T10:16:02.0839580Z .............................................................. [ 41%]
2024-07-09T10:16:02.1196630Z signal/tests/test_fir_filter_design.py ................................. [ 41%]
2024-07-09T10:16:02.2388470Z .... [ 41%]
2024-07-09T10:16:02.3864560Z signal/tests/test_ltisys.py ............................................ [ 41%]
2024-07-09T10:16:02.4179210Z ..................................... [ 41%]
2024-07-09T10:16:02.4248210Z signal/tests/test_max_len_seq.py .. [ 41%]
2024-07-09T10:16:02.4607540Z signal/tests/test_peak_finding.py ...................................... [ 41%]
2024-07-09T10:16:02.5206730Z ............... [ 41%]
2024-07-09T10:16:02.5257490Z signal/tests/test_result_type.py ..... [ 41%]
2024-07-09T10:16:02.5515660Z signal/tests/test_savitzky_golay.py .............. [ 41%]
2024-07-09T10:16:02.6127120Z signal/tests/test_short_time_fft.py .................................... [ 41%]
2024-07-09T10:16:02.8062180Z ........................................................................ [ 41%]
2024-07-09T10:16:03.1699940Z ........................................ [ 41%]
2024-07-09T10:16:03.2585410Z signal/tests/test_signaltools.py ....................................... [ 41%]
2024-07-09T10:16:03.3188100Z ........................................................................ [ 42%]
2024-07-09T10:16:03.5789530Z ........................................................................ [ 42%]
2024-07-09T10:16:03.6483680Z ........................................................................ [ 42%]
2024-07-09T10:16:03.7216200Z ........................................................................ [ 42%]
2024-07-09T10:16:03.7953200Z ........................................................................ [ 42%]
2024-07-09T10:16:03.8685520Z ........................................................................ [ 42%]
2024-07-09T10:16:03.9417370Z ........................................................................ [ 42%]
2024-07-09T10:16:04.0166170Z ........................................................................ [ 43%]
2024-07-09T10:16:04.1058930Z ........................................................................ [ 43%]
2024-07-09T10:16:04.2119420Z ........................................................................ [ 43%]
2024-07-09T10:16:04.3251090Z ........................................................................ [ 43%]
2024-07-09T10:16:04.4468720Z ........................................................................ [ 43%]
2024-07-09T10:16:04.5551980Z ........................................................................ [ 43%]
2024-07-09T10:16:04.6493940Z ........................................................................ [ 43%]
2024-07-09T10:16:04.7529190Z ........................................................................ [ 43%]
2024-07-09T10:16:04.8595980Z ........................................................................ [ 44%]
2024-07-09T10:16:04.9487580Z ........................................................................ [ 44%]
2024-07-09T10:16:04.9614450Z .....ss..../Users/runner/work/scipy/scipy/tools/wheels/cibw_test_command.sh: line 11: 13615 Trace/BPT trap: 5 python -c "import sys; import scipy; sys.exit(not scipy.test())"
2024-07-09T10:16:05.8676330Z ##[endgroup]
2024-07-09T10:16:05.8678070Z [31m✕ [0m121.80s
2024-07-09T10:16:05.8724980Z ##[error]Command ['/bin/sh', '-c', 'bash /Users/runner/work/scipy/scipy/tools/wheels/cibw_test_command.sh /Users/runner/work/scipy/scipy'] failed with code 133.
2024-07-09T10:16:05.8731550Z
2024-07-09T10:16:05.9060960Z ##[error]Process completed with exit code 1.
2024-07-09T10:16:05.9228700Z Post job cleanup.
2024-07-09T10:16:05.9299200Z Post job cleanup.
2024-07-09T10:16:06.2955200Z [command]/opt/homebrew/bin/git version
2024-07-09T10:16:06.3255860Z git version 2.45.2
2024-07-09T10:16:06.3296050Z Copying '/Users/runner/.gitconfig' to '/Users/runner/work/_temp/955ef7cb-151e-4e43-9d1d-139ce053d906/.gitconfig'
2024-07-09T10:16:06.3306900Z Temporarily overriding HOME='/Users/runner/work/_temp/955ef7cb-151e-4e43-9d1d-139ce053d906' before making global git config changes
2024-07-09T10:16:06.3307910Z Adding repository directory to the temporary git global config as a safe directory
2024-07-09T10:16:06.3309400Z [command]/opt/homebrew/bin/git config --global --add safe.directory /Users/runner/work/scipy/scipy
2024-07-09T10:16:06.3536820Z [command]/opt/homebrew/bin/git config --local --name-only --get-regexp core\.sshCommand
2024-07-09T10:16:06.4118110Z [command]/opt/homebrew/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :"
2024-07-09T10:16:06.5565800Z Entering 'doc/source/_static/scipy-mathjax'
2024-07-09T10:16:06.5729790Z Entering 'scipy/_lib/array_api_compat'
2024-07-09T10:16:06.5886670Z Entering 'scipy/_lib/boost_math'
2024-07-09T10:16:06.6096750Z Entering 'scipy/_lib/cobyqa'
2024-07-09T10:16:06.6270050Z Entering 'scipy/_lib/highs'
2024-07-09T10:16:06.6455590Z Entering 'scipy/_lib/pocketfft'
2024-07-09T10:16:06.6631580Z Entering 'scipy/_lib/unuran'
2024-07-09T10:16:06.6809800Z Entering 'scipy/sparse/linalg/_propack/PROPACK'
2024-07-09T10:16:06.6968700Z [command]/opt/homebrew/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader
2024-07-09T10:16:06.7004900Z http.https://github.com/.extraheader
2024-07-09T10:16:06.7012330Z [command]/opt/homebrew/bin/git config --local --unset-all http.https://github.com/.extraheader
2024-07-09T10:16:06.7057840Z [command]/opt/homebrew/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :"
2024-07-09T10:16:06.7621440Z Entering 'doc/source/_static/scipy-mathjax'
2024-07-09T10:16:06.7718540Z http.https://github.com/.extraheader
2024-07-09T10:16:06.7813910Z Entering 'scipy/_lib/array_api_compat'
2024-07-09T10:16:06.7932730Z http.https://github.com/.extraheader
2024-07-09T10:16:06.8060740Z Entering 'scipy/_lib/boost_math'
2024-07-09T10:16:06.8171270Z http.https://github.com/.extraheader
2024-07-09T10:16:06.8279550Z Entering 'scipy/_lib/cobyqa'
2024-07-09T10:16:06.8394170Z http.https://github.com/.extraheader
2024-07-09T10:16:06.8489320Z Entering 'scipy/_lib/highs'
2024-07-09T10:16:06.8592700Z http.https://github.com/.extraheader
2024-07-09T10:16:06.8706510Z Entering 'scipy/_lib/pocketfft'
2024-07-09T10:16:06.8956800Z http.https://github.com/.extraheader
2024-07-09T10:16:06.9190440Z Entering 'scipy/_lib/unuran'
2024-07-09T10:16:06.9285650Z http.https://github.com/.extraheader
2024-07-09T10:16:06.9469110Z Entering 'scipy/sparse/linalg/_propack/PROPACK'
2024-07-09T10:16:06.9562200Z http.https://github.com/.extraheader
2024-07-09T10:16:06.9945710Z Cleaning up orphan processes
2024-07-09T10:16:07.7179740Z Terminate orphan process: pid (4062) (diskimages-help)
Based on the test output (......ss....), and assuming that locally the test order within test_signaltools.py is unchanged from what it was in CI, the crash happened in:
scipy/signal/tests/test_signaltools.py::TestMedFilt::test_medfilt2d_parallel[float32]
which certainly looks likely since it's calling medfilt2d () under ThreadPoolExecutor(max_workers=4):
https://github.com/scipy/scipy/blob/81c53d48a290b604ec5faa34c0a7d48537b487d6/scipy/signal/tests/test_signaltools.py#L1171-L1180
The C code that ends up being called there is:
_sigtools_median2d: https://github.com/scipy/scipy/blob/81c53d48a290b604ec5faa34c0a7d48537b487d6/scipy/signal/_sigtoolsmodule.c#L954
_f_medfilt2d: https://github.com/scipy/scipy/blob/81c53d48a290b604ec5faa34c0a7d48537b487d6/scipy/signal/_medianfilter.c#L130
I'd expect this to be reproducible locally.
I wonder if the following patch still crashes:
$ git diff
diff --git a/scipy/signal/_signaltools.py b/scipy/signal/_signaltools.py
index 13caf1718f..5eb51f3596 100644
--- a/scipy/signal/_signaltools.py
+++ b/scipy/signal/_signaltools.py
@@ -2004,8 +2004,8 @@ def medfilt2d(input, kernel_size=3):
# checking dtype.type, rather than just dtype, is necessary for
# excluding np.longdouble with MS Visual C.
- if image.dtype.type not in (np.ubyte, np.float32, np.float64):
- return medfilt(image, kernel_size)
+# if image.dtype.type not in (np.ubyte, np.float32, np.float64):
+ return medfilt(image, kernel_size)
if kernel_size is None:
kernel_size = [3] * 2
what the patch does, it utlimately calls ndimage.rank_filter instead of multipack-era C implementation. Maybe instead of debugging those, it's easier (and certainly better for the code health overall) to use ndimage filtering across signal filtering functions.
(We did that for medfilt recently, got no complaints so far)
I tried to reproduce this one locally, however, all tests in signal/tests/test_signaltools.py are passing successfully under PYTHON_GIL=0
I was able to trigger the crash. I installed pytest-repeat using pip and then ran the following command:
PYTHON_GIL=0 lldb $(pyenv which python) -- dev.py test -- 'scipy.signal.tests.test_signaltools::TestMedFilt' --count=1000
And after some iterations iterations, hit a crash with the following python traceback on the segfaulting thread:
File "/Users/goldbaum/Documents/scipy/build-install/lib/python3.13/site-packages/scipy/signal/_signaltools.py", line 1951 in medfilt2d
File "/Users/goldbaum/Documents/scipy/build-install/lib/python3.13/site-packages/scipy/signal/tests/test_signaltools.py", line 1168 in apply
File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/concurrent/futures/thread.py", line 58 in run
File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/concurrent/futures/thread.py", line 92 in _worker
File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/threading.py", line 990 in run
File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/threading.py", line 1039 in _bootstrap_inner
File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/threading.py", line 1010 in _bootstrap
and the following C traceback,
(lldb) bt
* thread #3, stop reason = EXC_BAD_ACCESS (code=261, address=0xd50323ff)
* frame #0: 0x00000001842c1698 libsystem_platform.dylib`setjmp + 36
frame #1: 0x00000001006a30f4 _sigtools.cpython-313t-darwin.so`_sigtools_median2d(__NPY_UNUSED_TAGGEDdummy='0x171e12c30', args=('0x53580023728', '0x53580023730')) at _sigtoolsmodule.c:980:9
frame #2: 0x0000000100a878dc libpython3.13t.dylib`cfunction_call + 180
frame #3: 0x0000000100a23368 libpython3.13t.dylib`_PyObject_MakeTpCall + 324
frame #4: 0x0000000100b5efd0 libpython3.13t.dylib`_PyEval_EvalFrameDefault + 9684
frame #5: 0x0000000100a26818 libpython3.13t.dylib`method_vectorcall + 328
frame #6: 0x0000000100c512c0 libpython3.13t.dylib`thread_run + 128
frame #7: 0x0000000100be6f84 libpython3.13t.dylib`pythread_wrapper + 28
frame #8: 0x0000000184292f94 libsystem_pthread.dylib`_pthread_start + 136
Which is in a call to setjmp. I suspect there are hairy thread safety issues around use of longjmp/setjmp.
I think what's happening is described more or less in this StackOverflow:
https://stackoverflow.com/questions/38234546/setjmp-longjmp-between-threads-to-handle-timeout
There is a static jmp_buf in _sigtoolsmodule.c which defines a spot to handle malloc failures. It's only every used in one spot.
IMO, get rid of it and just use a normal goto fail block instead. Unless I'm misunderstanding why this code uses setjmp/longjmp at all.
Unless I'm misunderstanding why this code uses setjmp/longjmp at all.
I think it shouldn't indeed. The check_malloc should also be cleaned up. There's no callback mechanism here that would justify using setjmp/longjmp. Rather, it only avoids a single level of regular error handling it looks like (f_medfilt2d is void rather than int).
gh-21160 fixes this issue.
| gharchive/issue | 2024-07-09T11:38:56 | 2025-04-01T06:45:44.745229 | {
"authors": [
"andfoy",
"ev-br",
"ngoldbaum",
"rgommers"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/21142",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2526940707 | BUG: Windows Arm build fails
Describe your issue.
Can't install scipy on win arm machine for native Python. No pre-built wheels, from source compilation fails.
Reproducing Code Example
pip install scipy
Error message
Downloading scipy-1.14.1.tar.gz (58.6 MB)
--------------------------------------- 58.6/58.6 MB 24.2 MB/s eta 0:00:00
Installing build dependencies: started
Installing build dependencies: finished with status 'done'
Getting requirements to build wheel: started
Getting requirements to build wheel: finished with status 'done'
Installing backend dependencies: started
Installing backend dependencies: finished with status 'done'
Preparing metadata (pyproject.toml): started
Preparing metadata (pyproject.toml): finished with status 'error'
error: subprocess-exited-with-error
Preparing metadata (pyproject.toml) did not run successfully.
exit code: 1
[39 lines of output]
+ meson setup C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh -Dbuildtype=release -Db_ndebug=if-release -Db_vscrt=md --native-file=C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh\meson-python-native-file.ini
The Meson build system
Version: 1.5.1
Source dir: C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c
Build dir: C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh
Build type: native build
Project name: scipy
Project version: 1.14.1
Activating VS 17.8.3
C compiler for the host machine: cl (msvc 19.38.33133 "Microsoft (R) C/C++ Optimizing Compiler Version 19.38.33133 for ARM64")
C linker for the host machine: link link 14.38.33133.0
C++ compiler for the host machine: cl (msvc 19.38.33133 "Microsoft (R) C/C++ Optimizing Compiler Version 19.38.33133 for ARM64")
C++ linker for the host machine: link link 14.38.33133.0
Cython compiler for the host machine: cython (cython 3.0.11)
Host machine cpu family: aarch64
Host machine cpu: aarch64
Program python found: YES (X:\Github\PWABuilder\Windows-DevRel\Tools\ML-tests\.temp\env_scipy\Scripts\python.exe)
Run-time dependency python found: YES 3.11
Program cython found: YES (C:\Users\gkhmyznikov\AppData\Local\Temp\pip-build-env-cw5uc0or\overlay\Scripts\cython.EXE)
Compiler for C supports arguments -Wno-unused-but-set-variable: NO
Compiler for C supports arguments -Wno-unused-function: NO
Compiler for C supports arguments -Wno-conversion: NO
Compiler for C supports arguments -Wno-misleading-indentation: NO
Library m found: NO
..\meson.build:84:0: ERROR: Unknown compiler(s): [['ifort'], ['gfortran'], ['flang'], ['pgfortran'], ['g95']]
The following exception(s) were encountered:
Running `ifort --version` gave "[WinError 2] The system cannot find the file specified"
Running `ifort -V` gave "[WinError 2] The system cannot find the file specified"
Running `gfortran --version` gave "[WinError 2] The system cannot find the file specified"
Running `gfortran -V` gave "[WinError 2] The system cannot find the file specified"
Running `flang --version` gave "[WinError 2] The system cannot find the file specified"
Running `flang -V` gave "[WinError 2] The system cannot find the file specified"
Running `pgfortran --version` gave "[WinError 2] The system cannot find the file specified"
Running `pgfortran -V` gave "[WinError 2] The system cannot find the file specified"
Running `g95 --version` gave "[WinError 2] The system cannot find the file specified"
Running `g95 -V` gave "[WinError 2] The system cannot find the file specified"
A full log can be found at C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh\meson-logs\meson-log.txt
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
error: metadata-generation-failed
System.Management.Automation.RemoteException
Encountered error while generating package metadata.
System.Management.Automation.RemoteException
See above for output.
System.Management.Automation.RemoteException
SciPy/NumPy/Python version and system information
scipy 1.14.1
python 3.11.9 win-arm64
I don't know if the build will succeed but at least to try, you need a fortran compiler on your system. That's what the error is saying.
@ilayn any fortran compiler for win arm is available?
I have no idea unfortunately. Not sure if we even support windows arm. It's not an architecture I am familiar with.
Given that the root cause here is a missing Fortran compiler and the builds from @cgohlke show that a from source build on Windows on Arm works fine at the moment, let's close this issue.
We can open a "provide wheels for Windows on Arm" issue once NumPy has wheels, it's not actionable now.
| gharchive/issue | 2024-09-15T14:31:51 | 2025-04-01T06:45:44.756371 | {
"authors": [
"ilayn",
"khmyznikov",
"rgommers"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/21562",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
214562613 | Add guidance on error/warning/exception handling patterns and best practices
I could not easily find information on what to expect from scipy in terms error/warning/exception handling patterns and best practices, either as a consumer or a contributor.
I looked here: http://scipy.github.io/devdocs/hacking.html
and here: http://scipy.github.io/devdocs/dev/index.html
Not sure what you mean. Maybe an example would help?
So numpy has settings for how to handle floating point errors here. In a "typical" scipy routine, what should cause a warning vs. an exception, and what about returning a separate error value/boolean from a computation vs. simply having a nan for the output to signify something went wrong? How can the python logger be leveraged effectively, if desired? What design pattens allow a scipy consumer flexibility in detecting, handling, and logging warnings and errors?
I have never seen any written down or spelled out policy for this.
based on my experience:
numpy still applies, e.g. Zero Division Warnings in any code depend on the numpy settings
incorrect user input that is not usable raises an Exception, ValueError or similar, This can be either explicit in the scipy function or via numpy if for example shapes don't match.
vectorized functions return np.nan in positions that cannot be computed to a specific value
np.inf is mostly just a regular float and not a value to check for
optional checks can raise if condition applies, e.g. isfinite check in linalg or nan_policy now in stats
specific computational problems like nonconvergence in optimize or integration problems might just issue a specific warning
basic principle:
If the user input is ambiguous or there no useful computations can be performed, then raise.
If there is a problem, but the computation still contains useful information, then warn or return nan and let the users check it. (e.g. nonconvergence or vectorized)
Also a note on
So numpy has settings for how to handle floating point errors
As of 0.19 scipy.special has a similar style of error handling:
https://scipy.github.io/devdocs/special.html#error-handling
Also as this isn't a bug the mailing list would probably be a more appropriate place for this discussion.
Thanks for your feedback here. I'm looking into the mailing list option for ideas around formulating best practices.
The dev mailing list would be the appropriate location to discuss whether this can or should be spelled out. My impression is that it is largely an implicit policy.
(related aside: statsmodels is suffering from "organic growth" in this area without systematic review or policy and insufficient unit tests for flags, warnings and exceptions if problems show up https://github.com/statsmodels/statsmodels/issues/3520
scipy is much better but I would expect that it still has some inconsistencies)
| gharchive/issue | 2017-03-16T00:07:25 | 2025-04-01T06:45:44.764410 | {
"authors": [
"josef-pkt",
"person142",
"thunderfish24"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/7183",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
261725566 | linalg.solve ill-condition warning
For the linalg.solve, since version 19.1, we have the capability of printing the warning the ill conditioning of the matrix. However the result is not quite pretty due to the traceback inclusion. For example:
I think the story got a bit uglier with our latest change that also prints the reported dlamch machine precision since now we actually have to look for the correct number.
I think it's best if we print the truncated versions of those numbers anyhow. But should we switch to a warning printing scheme instead? I don't know yet any way to raise a warning without printing the traceback.
One use case I'd like to retain is that if a user is solving iteratively a lot of equations s/he should still be able to catch a warning if the problem is ill-conditioned (which was the reason to emit a warning in the first place). Or we can emit a warning directly and then do a custom printing. Such that there is a clear distinction between the traceback and the actual warning message.
Can we set the stacklevel so that it's in the calling function? Then the traceback is much more useful.
In general, all uses of warnings.warn should have stacklevel set.
I don't know actually. If it is too deep, let's say (I'm just making it up) something is computing a filter coefficient for conversion from state space to transfer function, to get the frequency response to draw Bode plot, which function should be the source the deepest or the calling function?
If it is too deep,
This is a problem for the caller, whether it is a user or a library. In other words, in your example, it's a problem for the state space -> transfer function code; and we should improve our filter code so that it does the right thing with this warning (avoids it or catches it). So I'd say just go up however many stacklevels (2?) to get to the one that called e.g. scipy.linalg.solve.
I've played around with this a bit more and it turns out that setting the right value to the stacklevel is trickier. I've also asked in SO to see whether someone teaches me more about it. Essentially there are two problems :
from warnings import warn, formatwarning
def _warning_on_one_line(message, category, filename, lineno,
file=None, line=None):
return ' %s:%s: %s:%s' % (filename, lineno, category.__name__, message)
def inner():
formatwarning = _warning_on_one_line
warn('Some warning', RuntimeWarning, stacklevel=3)
return
def middle():
return inner()
def outer():
return middle()
outer()
As you play with stacklevel argument you can basically cause the warning to report different lines in this code. But if you overdo it you start getting IPython's caller stack all the way up to 26 where it finally starts to say sys.1: RuntimeWarning so detecting the correct caller doesn't seem obvious to me. And I agree that I think we have to be a bit more systematic when including a potentially warning emitting code inside a function. Such that we can catch the warning and emit a more meaningful message. So it seems to me that we are stuck with stacklevel=1 for now.
Second problem is how to format the warning message but that is kind how it is done as given in the above code.
Do you have a way to overcome this issue?
My idea is/was:
So I'd say just go up however many stacklevels (2?) to get to the one that called e.g. scipy.linalg.solve.
If we know where the warning is emitted, and we know the codepath to get there from scipy.linalg.solve, we can do this, right?
... in other words, in your example, if scipy.linalg.solve (in our codebase) is inner, you'd use stacklevel=1, if it's middle you'd use stacklevel=2, and if it's outer you'd use stacklevel=3 (maybe +/-1 to all of these, can't remember the level details).
So it seems to me that we are stuck with stacklevel=1 for now.
If you're saying that the warning is currently emitted in scipy.linalg.solve, then yeah I agree stacklevel=1 (assuming that's the one that shows scipy.linalg.solve's caller) is the right thing to do.
The stacktrace indicates the point in the calling code where the
warning is emitted. The stacklevel must correspond to the innermost
public API function. This is what Numpy does, and it is sensible behavior.
For a public API function, we write
def public_api_func():
warnings.warn("some warning", stacklevel=2)
For an internal function for which we know the distance to the API function:
def _helper_func():
warnings.warn("some warning", stacklevel=3)
def public_api_func():
_helper_func()
If we know where the warning is emitted, and we know the codepath to get there from scipy.linalg.solve, we can do this, right?
Yes indeed but depending on platform we might go too far and start addressing irrelevant parts of the current stack, that's what I meant by the 26 argument. For example if I use stacklevel=9
runfile('C:/Users/Ilhan Polat/Documents/pydump/warning_exp.py', wdir='C:/Users/Ilhan Polat/Documents/pydump')
c:\users\ilhan polat\appdata\local\programs\python\python36\lib\site-packages\IPython\core\interactiveshell.py:2808: RuntimeWarning: Some warning
if self.run_code(code, result):
which is related to IPython internals So give me the top of the stack kind of argument won't be accurate to get the actual code path.
The stacktrace indicates the point in the calling code where the warning is emitted. The stacklevel must correspond to the innermost public API function. This is what Numpy does, and it is sensible behavior.
Ah OK now I get your arguments.
Warnings emitted inside Scipy code (e.g. due to calling solve() on a
singular matrix in the implementation of some routine) are probably
implementation defect on our side, although usually probably minor one.
Either we e.g. know the algorithm works correctly despite such
singularity (in which case we should ensure somehow the warning doesn't
get emitted), or the singularity implies a failure of the algorithm (in
which case an error or a more specific warning would be more appropriate
in the perfect world).
I agree but now I'm a bit confused how to detect an ill-conditioned array. Because if we remove the warning, then it goes unnoticed. This is known to matlab users a lot with its silent exception handling and only blowing up when the algorithm cannot move any forward. If we emit the warning only pointing to where linalg.solve is called then we have the current situation where the warning seems to be unrelated to the current caller.
Silencing the warning if it is not related is easy. Modifying the warning message is also doable by catching the warning. However I don't know what should be the way to go with this.
Because if we remove the warning, then it goes unnoticed
IIUC the propsal is to:
Always emit the warning in the scipy.linalg.solve code when ill conditioned matrices are detected.
In places in SciPy where we internally use scipy.linalg.solve, we should "do the right thing". In other words, we should deal with the warning safely so that no warning is ever emitted. This could mean "catch the warning and raise an error" in cases where rank deficiency is bad, or it could mean "catch the warning, and then use linalg.lstsq instead, because rank-deficiency is fine but we'd like the speed of solve if possible".
Point (2) is something we need to decide / do on a case-by-case basis as it comes up, so I don't think we need to worry about it right now.
OK, I'll just try to make the warning message a bit more structured then and leave the decision part for later then.
| gharchive/issue | 2017-09-29T18:27:33 | 2025-04-01T06:45:44.780991 | {
"authors": [
"ilayn",
"larsoner",
"pv"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/7951",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
266012502 | welch method unexpected results for low frequency
When computing the PSD of the synthetic signal below using welch method I obtain an unexpected peak near 0 Hz with no physical meaning. The equivalent MATLAB code produces different results for that region.
Python code example:
import numpy as np
import matplotlib.pyplot as plt
from scipy.signal import welch, hanning
x = np.linspace(0, 10, 100001)
dt = x[1] - x[0]
fs = 1 / dt
a1 = 1
f1 = 500
a2 = 10
f2 = 2000
y = a1 * np.sin(2*np.pi*f1*x) + a2 * np.sin(2*np.pi*f2*x)
datos = y
nblock = 1024
overlap = 128
win = hanning(nblock, True)
f, Pxxf = welch(datos, fs, window=win, noverlap=overlap, nfft=nblock, return_onesided=True)
plt.semilogy(f, Pxxf, '-o')
plt.grid()
plt.show()
MATLAB code example
x = linspace(0, 10, 100001);
dt = x(2) - x(1);
Fs = 1 / dt;
a1 = 1;
f1 = 500;
a2 = 10;
f2 = 2000;
y = a1 * sin(2*pi*f1*x) + a2 * sin(2*pi*f2*x);
nblock=1024;
overlap=128;
windowsel=hann(nblock);
[Pxx,f]=pwelch(y,windowsel,overlap,nblock,Fs,'onesided');
figure()
semilogy(f,Pxx, '-o')
Scipy/Numpy/Python version information:
Scipy 0.19.1
NumPy 1.13.1
Python 3.6.2 final (serial 0)
Matlab:
Python
I can reproduce with simplified scripts that only process a single window and shorter signals:
fs = 10.
n_samp = 32
t = np.arange(n_samp) / fs
y = np.sin(2*np.pi*2*t)
win = windows.hann(n_samp, False)
f, Pxxf = welch(y, fs, window=win, noverlap=0, nfft=n_samp,
return_onesided=True)
plt.semilogy(f, Pxxf, '-o')
plt.xlim([0, 5])
plt.ylim([1e-7, 1])
plt.show()
MATLAB:
clear all; close all; home;
fs = 10;
n_samp = 32;
x = (0:n_samp-1) / fs;
y = sin(2*pi*2*x);
overlap=0;
windowsel=hann(n_samp);
[Pxx,f]=pwelch(y,windowsel,overlap,n_samp,fs,'onesided');
figure();
semilogy(f,Pxx, '-o');
But Octave at least agrees with our output:
I'm not sure what MATLAB is doing to change/correct these values.
I will have a look at the implementation of scipy and try to figure out where this behaviour might be coming from. Do you have any suggestion on that?
I have also found this matlab example that might be relevant: https://es.mathworks.com/matlabcentral/answers/33653-psd-estimation-fft-vs-welch
It compares results obtained using FFT with results using PSD and both show the same behaviour near 0 Hz. It may be a good starting point.
Try using detrend=False. signal.welch defaults to subtracting the mean of the data (detrend='constant') which seems to be giving you this DC artifact.
Thanks @ewmoore I can confirm detrend=False fixes it.
Awesome! Thank you!
It occurred to me that the root cause of this low-frequency energy showing up in the estimated PSD is that detrending takes place before windowing (check _fft_helper()). This means that the DC component of the windowed frame is not zero even though it has been detrended to have zero mean!
To illustrate, I modified the example provided by @larsoner:
fs = 10.
n_samp = 32
t = np.arange(n_samp) / fs
y = np.sin(2*np.pi*2*t)
win = windows.hann(n_samp, False)
windowed_y = y*win # apply window already now
f, Pxxf = welch(windowed_y , fs, window='boxcar', noverlap=0) # do not apply it here
plt.semilogy(f, Pxxf, '-o')
plt.xlim([0, 5])
plt.ylim([1e-7, 1])
plt.show()
This results in:
Which has zero energy at DC, as expected given the default detrending method.
@larsoner, do you think this qualifies as a bug that should be reported?
Sorry for adding a comment years after this issue was marked as closed.
(as a procedural note, feel free to open a new issue in case this doesn't receive a response)
| gharchive/issue | 2017-10-17T06:56:21 | 2025-04-01T06:45:44.791013 | {
"authors": [
"AlexS12",
"akukuq",
"ewmoore",
"larsoner",
"lucascolley"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/issues/8045",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1142111325 | CI: pin asv to avoid slowdowns in 0.5/0.5.1
[skip azp]
[skip actions]
Closes gh-15568
Okay, CI is green again, the end of almost 2 weeks of N emails/day about failures. So in it goes.
Timing for the job is back to ~25 minutes for the asv run, rather than timing out at 40% after an hour.
| gharchive/pull-request | 2022-02-18T00:14:26 | 2025-04-01T06:45:44.792933 | {
"authors": [
"rgommers"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/pull/15616",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
142766020 | ENH: Add gesvd method for svd
First pass at adding gesvd to complement gesdd method for svd.
Closes #4994.
@pv feel free to comment on if you think this approach makes sense. I hope to find time to get it working fully in the next couple of weeks.
When adding something similar to lstsq we called the argument lapack_driver and just directly used the the names of the lapack functions as the option names. It would be good to be consistent.
Sounds good, will change to lapack_driver='gesdd' | 'gesvd'
Okay @pv @ewmoore this one should be good to go now.
@@ master #5994 diff @@
======================================
Files 238 238
Stmts 43803 43809 +6
Branches 8211 8213 +2
Methods 0 0
======================================
+ Hit 34230 34236 +6
Partial 2603 2603
Missed 6970 6970
Review entire Coverage Diff as of 360d271
Powered by Codecov. Updated on successful CI builds.
| gharchive/pull-request | 2016-03-22T20:13:10 | 2025-04-01T06:45:44.797398 | {
"authors": [
"Eric89GXL",
"codecov-io",
"ewmoore"
],
"repo": "scipy/scipy",
"url": "https://github.com/scipy/scipy/pull/5994",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
275455661 | Error response contains request id
Fixes #525
Review Checklist
Tests were added to cover all code changes
Documentation was added / updated
Code and tests follow standards in CONTRIBUTING.md
Codecov Report
Merging #1002 into master will increase coverage by 0.04%.
The diff coverage is 87.5%.
@@ Coverage Diff @@
## master #1002 +/- ##
==========================================
+ Coverage 90.32% 90.36% +0.04%
==========================================
Files 50 50
Lines 6716 6717 +1
==========================================
+ Hits 6066 6070 +4
+ Misses 650 647 -3
| gharchive/pull-request | 2017-11-20T18:34:40 | 2025-04-01T06:45:44.810764 | {
"authors": [
"codecov-io",
"hkethi002"
],
"repo": "scitran/core",
"url": "https://github.com/scitran/core/pull/1002",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2209650577 | Inventory of related efforts
Closes #51.
@nikosft, thoughts? Can we discuss a little more what you want to add here?
@SteveLasker, since you are touching up formatting: I am ok with this as a starting point (by @nikosft and I, thanks to him for his contributions so far). I am sure we can add to it over time incrementally. If you are ready to merge it, so am I. I marked it as ready last evening and hope to add more over time of course, but don't have anything in the immediate future here. 🚢 it when ready.
Is there anything we can do to unblock this pull request?
Is there anything we can do to unblock this pull request?
I am happy with it if you are, we should try to merge it in short order. I probably shouldn't be reviewing and merging a PR I worked on.
I am happy too :)
To my understanding, related efforts are things that have the same goal as SCITT but are not SCITT.
I think it is important to talk about how SCITT relates to other efforts and contextualize it. If not, we presume that information is elsewhere (I think it is not).
Is there something we can to adjust or scope the content in the PR to finalize the review, or should we just close this PR?
@SteveLasker, can we discuss at the tail end of next week's meeting how we intend to handle this?
| gharchive/pull-request | 2024-03-27T01:46:07 | 2025-04-01T06:45:44.814566 | {
"authors": [
"aj-stein-nist",
"nikosft"
],
"repo": "scitt-community/scitt-web",
"url": "https://github.com/scitt-community/scitt-web/pull/52",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
872640566 | Creating a Shared Library
Hello,
When I try to use -DBUILD_SHARED_LIBS=on when using Intel compilers on Windows, I receive the following link errors:
[870/887] Linking Fortran shared library src\cmumps.dll
FAILED: src/cmumps.dll src/cmumps.lib
cmd.exe /C "cmd.exe /C ""C:\Program Files\CMake\bin\cmake.exe" -E __create_def C:\Tools\MUMPS\build\src\CMakeFiles\cmumps.dir.\exports.def C:\Tools\MUMPS\build\src\CMakeFiles\cmumps.dir.\exports.def.objs && cd C:\Tools\MUMPS\build" && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_dll --intdir=src\CMakeFiles\cmumps.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\10.0.18362.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\10.0.18362.0\x64\mt.exe --manifests -- xilink /nologo src\CMakeFiles\cmumps.dir\mumps_c.c.obj src\CMakeFiles\cmumps.dir\cana_aux.F.obj src\CMakeFiles\cmumps.dir\cana_aux_par.F.obj src\CMakeFiles\cmumps.dir\cana_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_ELT_m.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_m.F.obj src\CMakeFiles\cmumps.dir\cfac_front_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_type2_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_sispointers_m.F.obj src\CMakeFiles\cmumps.dir\cfac_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_dynamic.F.obj src\CMakeFiles\cmumps.dir\cfac_omp_m.F.obj src\CMakeFiles\cmumps.dir\cfac_par_m.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_l0omp_m.F.obj src\CMakeFiles\cmumps.dir\clr_core.F.obj src\CMakeFiles\cmumps.dir\clr_stats.F.obj src\CMakeFiles\cmumps.dir\clr_type.F.obj src\CMakeFiles\cmumps.dir\cmumps_comm_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_config_file.F.obj src\CMakeFiles\cmumps.dir\cmumps_load.F.obj src\CMakeFiles\cmumps.dir\cmumps_lr_data_m.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc.F.obj src\CMakeFiles\cmumps.dir\cmumps_sol_es.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore_files.F.obj src\CMakeFiles\cmumps.dir\cmumps_struc_def.F.obj src\CMakeFiles\cmumps.dir\comp_tps_m.F.obj src\CMakeFiles\cmumps.dir\csol_lr.F.obj src\CMakeFiles\cmumps.dir\csol_omp_m.F.obj src\CMakeFiles\cmumps.dir\cstatic_ptr_m.F.obj src\CMakeFiles\cmumps.dir\cini_driver.F.obj src\CMakeFiles\cmumps.dir\cana_driver.F.obj src\CMakeFiles\cmumps.dir\cfac_driver.F.obj src\CMakeFiles\cmumps.dir\csol_driver.F.obj src\CMakeFiles\cmumps.dir\csol_distrhs.F.obj src\CMakeFiles\cmumps.dir\cend_driver.F.obj src\CMakeFiles\cmumps.dir\cana_aux_ELT.F.obj src\CMakeFiles\cmumps.dir\cana_dist_m.F.obj src\CMakeFiles\cmumps.dir\cana_LDLT_preprocess.F.obj src\CMakeFiles\cmumps.dir\cana_reordertree.F.obj src\CMakeFiles\cmumps.dir\carrowheads.F.obj src\CMakeFiles\cmumps.dir\cbcast_int.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_asm.F.obj src\CMakeFiles\cmumps.dir\cfac_b.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_distentry.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_lastrtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_alloc_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_compress_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_free_block_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack.F.obj src\CMakeFiles\cmumps.dir\cfac_process_band.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blfac_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto_LDLT.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto.F.obj src\CMakeFiles\cmumps.dir\cfac_process_bf.F.obj src\CMakeFiles\cmumps.dir\cfac_process_end_facto_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type3.F.obj src\CMakeFiles\cmumps.dir\cfac_process_maprow.F.obj src\CMakeFiles\cmumps.dir\cfac_process_master2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_message.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2son.F.obj src\CMakeFiles\cmumps.dir\cfac_process_rtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_root_parallel.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings.F.obj src\CMakeFiles\cmumps.dir\cfac_determinant.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScaleAbs.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScale_util.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_pool.F.obj src\CMakeFiles\cmumps.dir\cfac_type3_symmetrize.F.obj src\CMakeFiles\cmumps.dir\cini_defaults.F.obj src\CMakeFiles\cmumps.dir\cmumps_driver.F.obj src\CMakeFiles\cmumps.dir\cmumps_f77.F.obj src\CMakeFiles\cmumps.dir\cmumps_gpu.c.obj src\CMakeFiles\cmumps.dir\cmumps_iXamax.F.obj src\CMakeFiles\cmumps.dir\cana_mtrans.F.obj src\CMakeFiles\cmumps.dir\cooc_panel_piv.F.obj src\CMakeFiles\cmumps.dir\crank_revealing.F.obj src\CMakeFiles\cmumps.dir\csol_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd.F.obj src\CMakeFiles\cmumps.dir\csol_c.F.obj src\CMakeFiles\cmumps.dir\csol_fwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_fwd.F.obj src\CMakeFiles\cmumps.dir\csol_matvec.F.obj src\CMakeFiles\cmumps.dir\csol_root_parallel.F.obj src\CMakeFiles\cmumps.dir\ctools.F.obj src\CMakeFiles\cmumps.dir\ctype3_root.F.obj /out:src\cmumps.dll /implib:src\cmumps.lib /pdb:src\cmumps.pdb /dll /version:0.0 /machine:x64 /INCREMENTAL:NO /DEF:src\CMakeFiles\cmumps.dir.\exports.def src\mumps_common.lib PORD\pord.lib "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_scalapack_lp64.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_blacs_intelmpi_lp64.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_intel_lp64.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_sequential.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_core.lib" "C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpifec.lib" "C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpi.lib" user32.lib && cd ."
LINK: command "xilink /nologo src\CMakeFiles\cmumps.dir\mumps_c.c.obj src\CMakeFiles\cmumps.dir\cana_aux.F.obj src\CMakeFiles\cmumps.dir\cana_aux_par.F.obj src\CMakeFiles\cmumps.dir\cana_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_ELT_m.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_m.F.obj src\CMakeFiles\cmumps.dir\cfac_front_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_type2_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_sispointers_m.F.obj src\CMakeFiles\cmumps.dir\cfac_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_dynamic.F.obj src\CMakeFiles\cmumps.dir\cfac_omp_m.F.obj src\CMakeFiles\cmumps.dir\cfac_par_m.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_l0omp_m.F.obj src\CMakeFiles\cmumps.dir\clr_core.F.obj src\CMakeFiles\cmumps.dir\clr_stats.F.obj src\CMakeFiles\cmumps.dir\clr_type.F.obj src\CMakeFiles\cmumps.dir\cmumps_comm_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_config_file.F.obj src\CMakeFiles\cmumps.dir\cmumps_load.F.obj src\CMakeFiles\cmumps.dir\cmumps_lr_data_m.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc.F.obj src\CMakeFiles\cmumps.dir\cmumps_sol_es.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore_files.F.obj src\CMakeFiles\cmumps.dir\cmumps_struc_def.F.obj src\CMakeFiles\cmumps.dir\comp_tps_m.F.obj src\CMakeFiles\cmumps.dir\csol_lr.F.obj src\CMakeFiles\cmumps.dir\csol_omp_m.F.obj src\CMakeFiles\cmumps.dir\cstatic_ptr_m.F.obj src\CMakeFiles\cmumps.dir\cini_driver.F.obj src\CMakeFiles\cmumps.dir\cana_driver.F.obj src\CMakeFiles\cmumps.dir\cfac_driver.F.obj src\CMakeFiles\cmumps.dir\csol_driver.F.obj src\CMakeFiles\cmumps.dir\csol_distrhs.F.obj src\CMakeFiles\cmumps.dir\cend_driver.F.obj src\CMakeFiles\cmumps.dir\cana_aux_ELT.F.obj src\CMakeFiles\cmumps.dir\cana_dist_m.F.obj src\CMakeFiles\cmumps.dir\cana_LDLT_preprocess.F.obj src\CMakeFiles\cmumps.dir\cana_reordertree.F.obj src\CMakeFiles\cmumps.dir\carrowheads.F.obj src\CMakeFiles\cmumps.dir\cbcast_int.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_asm.F.obj src\CMakeFiles\cmumps.dir\cfac_b.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_distentry.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_lastrtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_alloc_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_compress_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_free_block_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack.F.obj src\CMakeFiles\cmumps.dir\cfac_process_band.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blfac_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto_LDLT.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto.F.obj src\CMakeFiles\cmumps.dir\cfac_process_bf.F.obj src\CMakeFiles\cmumps.dir\cfac_process_end_facto_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type3.F.obj src\CMakeFiles\cmumps.dir\cfac_process_maprow.F.obj src\CMakeFiles\cmumps.dir\cfac_process_master2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_message.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2son.F.obj src\CMakeFiles\cmumps.dir\cfac_process_rtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_root_parallel.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings.F.obj src\CMakeFiles\cmumps.dir\cfac_determinant.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScaleAbs.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScale_util.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_pool.F.obj src\CMakeFiles\cmumps.dir\cfac_type3_symmetrize.F.obj src\CMakeFiles\cmumps.dir\cini_defaults.F.obj src\CMakeFiles\cmumps.dir\cmumps_driver.F.obj src\CMakeFiles\cmumps.dir\cmumps_f77.F.obj src\CMakeFiles\cmumps.dir\cmumps_gpu.c.obj src\CMakeFiles\cmumps.dir\cmumps_iXamax.F.obj src\CMakeFiles\cmumps.dir\cana_mtrans.F.obj src\CMakeFiles\cmumps.dir\cooc_panel_piv.F.obj src\CMakeFiles\cmumps.dir\crank_revealing.F.obj src\CMakeFiles\cmumps.dir\csol_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd.F.obj src\CMakeFiles\cmumps.dir\csol_c.F.obj src\CMakeFiles\cmumps.dir\csol_fwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_fwd.F.obj src\CMakeFiles\cmumps.dir\csol_matvec.F.obj src\CMakeFiles\cmumps.dir\csol_root_parallel.F.obj src\CMakeFiles\cmumps.dir\ctools.F.obj src\CMakeFiles\cmumps.dir\ctype3_root.F.obj /out:src\cmumps.dll /implib:src\cmumps.lib /pdb:src\cmumps.pdb /dll /version:0.0 /machine:x64 /INCREMENTAL:NO /DEF:src\CMakeFiles\cmumps.dir.\exports.def src\mumps_common.lib PORD\pord.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_scalapack_lp64.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_blacs_intelmpi_lp64.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_intel_lp64.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_sequential.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_core.lib C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpifec.lib C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpi.lib user32.lib /MANIFEST /MANIFESTFILE:src\cmumps.dll.manifest" failed (exit code 1120) with the following output:
Creating library src\cmumps.lib and object src\cmumps.exp
LINK : warning LNK4098: defaultlib 'LIBCMT' conflicts with use of other libs; use /NODEFAULTLIB:library
cooc_panel_piv.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U
csol_bwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U
csol_fwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U
cfac_front_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U
cfac_front_LU_type1.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U
cfac_front_LU_type2.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U
ctools.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cfac_process_blocfacto.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cooc_panel_piv.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
csol_bwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
csol_fwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cfac_par_m.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cmumps_ooc_buffer.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cfac_process_blocfacto_LDLT.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cfac_front_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cfac_front_LDLT_type1.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cfac_front_LDLT_type2.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cfac_front_type2_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L
cmumps_load.F.obj : error LNK2019: unresolved external symbol MUMPS_FUTURE_NIV2_mp_FUTURE_NIV2 referenced in function CMUMPS_LOAD_mp_CMUMPS_LOAD_UPDATE
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_ERR_STR_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_ERR_STR_OOC
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_DIM_ERR_STR_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_DIM_ERR_STR_OOC
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_MYID_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_MYID_OOC
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_ICNTL1 referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_ICNTL1
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_LOW_LEVEL_STRAT_IO referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_LOW_LEVEL_STRAT_IO
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_VADDR referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_VADDR
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_STEP_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_STEP_OOC
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_INODE_SEQUENCE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_INODE_SEQUENCE
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_NB_FILE_TYPE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_BUF_CLEAN_PENDING
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_NB_FILE_TYPE
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_ADDVIRTLIBRE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_END_OOC_BUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_ADDVIRTLIBRE
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_HBUF_SIZE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_COPY_DATA_TO_BUFFER
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_HBUF_SIZE
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_STRAT_IO_ASYNC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_INIT_DB_BUFFER_PANEL
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_STRAT_IO_ASYNC
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_DIM_BUF_IO referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_INIT_DB_BUFFER_PANEL
cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_KEEP_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_INIT_OOC_BUF
cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_KEEP_OOC
cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_FCT_TYPE referenced in function CMUMPS_OOC_mp_CMUMPS_IS_THERE_FREE_SPACE
cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_WITH_BUF referenced in function CMUMPS_OOC_mp_CMUMPS_NEW_FACTOR
cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_PROCNODE_OOC referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_END_FACTO
cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_SLAVEF_OOC referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_INIT_SOLVE
cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_CB referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_INIT_SOLVE
cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_SOLVE referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_INIT_SOLVE
cfac_process_band.F.obj : error LNK2019: unresolved external symbol MUMPS_FAC_DESCBAND_DATA_M_mp_INODE_WAITED_FOR referenced in function CMUMPS_PROCESS_DESC_BANDE
src\cmumps.dll : fatal error LNK1120: 24 unresolved externals
[873/887] Building Fortran object src\CMakeFiles\zmumps.dir\zmumps_save_restore.F.obj
ninja: build stopped: subcommand failed.
I imagine these symbols are supposed to be in mumps_common.dll/mumps_common.lib. When I run dumpbin /exports mumps_common.lib I see that some of the unresolved symbols above are in mumps_common.lib but others aren't. Any ideas on how to build a shared library in this context?
P.S.: It is possible to build a shared library using MSYS2 and Mingw64 on Windows by specifying -DCMAKE_GNUtoMS=on -DBUILD_SHARED_LIBS=on but I can't use this approach because of an unrelated bug with gfortran and MSMPI.
Thanks.
I added a warning in 751f8252d028d573623ad1c8779212a102bd6864 that this setup (shared libs with MSVC/Intel on Windows) isn't supported. I don't have the time to fix that, but if you have a fix I could try it.
It does work with MSYS2 / MinGW and BUILD_SHARED_LIBS=on
I've refactored some build options to help shared builds on other platforms, which work. However, Windows + Intel oneAPI + shared libs still doesn't work with similar errors to above.
I think this might be an issue for upstream MUMPS library. I simply am providing an easy CMake build instead of the Makefiles that upstream MUMPS provides.
| gharchive/issue | 2021-03-16T14:47:31 | 2025-04-01T06:45:44.880269 | {
"authors": [
"ryagal",
"scivision"
],
"repo": "scivision/mumps",
"url": "https://github.com/scivision/mumps/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
250175557 | Finding selectors on hidden elements...
Hi there ... new to a lot of dev stuff, and agouti in particular. Question on whether I can call Find methods using CSS selectors on elements that are hidden.
Trying to create a selection object by calling the Find method on a file input field that is hidden. (Looking to ultimately call the UploadFile method on the selection object, but failing on the step to create the selection object.)
any guidance would be much-appreciated!
Hi @smithers1221, apologies for the delay!
I recommend using agouti.RunScript to inject JavaScript into the page to manipulate the hidden input field. It's difficult to interact with hidden elements via the WebDriver, as WebDrivers are designed to replicate user actions (which can't happen on hidden elements).
| gharchive/issue | 2017-08-14T22:48:14 | 2025-04-01T06:45:44.882952 | {
"authors": [
"sclevine",
"smithers1221"
],
"repo": "sclevine/agouti",
"url": "https://github.com/sclevine/agouti/issues/115",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2367666386 | Generators: do not hardcode archive exe path
Resolve from archive contents instead.
Done as part of #70
| gharchive/issue | 2024-06-22T07:30:27 | 2025-04-01T06:45:44.897478 | {
"authors": [
"scop"
],
"repo": "scop/wrun",
"url": "https://github.com/scop/wrun/issues/60",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1454973988 | Validation of fields is nearly non-existent
This works and it shouldn't (look at the protocol)
apiVersion: score.dev/v1b1
metadata:
name: frontend
containers:
frontend:
image: nginx:alpine
service:
ports:
www:
port: 80
protocol: Godzilla-growls-angry
targetPort: 80
Translates with score-helm to:
containers:
frontend:
image:
name: nginx:alpine
service:
ports:
- name: www
port: 80
protocol: Godzilla-growls-angry
targetPort: 80
type: ClusterIP
This also "works"
apiVersion: score.dev/v1b1
metadata:
name: frontend
container-with-massive-angry-typo:
frontend:
image: nginx:alpine
service:
ports:
www:
port: 80
protocol: Godzilla-growls-angry
targetPort: 80
Outputs this:
containers: {}
service:
ports:
- name: www
port: 80
protocol: Godzilla-growls-angry
targetPort: 80
type: ClusterIP
And it doesn't error at all, there is no validation as it exists with 0 status
In the first section, you specified an invalid protocol for the port. In the second section, you used an incorrect field name for the container, which caused it to be ignored.
Try this for the first:
apiVersion: score.dev/v1b1 metadata: name: frontend containers: frontend: image: nginx:alpine service: ports: www: port: 80 protocol: TCP # Update the protocol to a valid value, such as TCP targetPort: 80
And this for the helm-part:
apiVersion: score.dev/v1b1 metadata: name: frontend containers: # Update the field name to the correct name, which is "containers" frontend: image: nginx:alpine service: ports: www: port: 80 protocol: TCP targetPort: 80
In the first section, you specified an invalid protocol for the port. In the second section, you used an incorrect field name for the container, which caused it to be ignored.
I think that's exactly @fernando-villalba's point: these errors should be automatically detected by Score implementations to help with troubleshooting.
Duplicate of https://github.com/score-spec/spec/issues/6
| gharchive/issue | 2022-11-18T11:48:08 | 2025-04-01T06:45:44.904330 | {
"authors": [
"adityassharma-ss",
"dserodio",
"fernando-villalba",
"sujaya-sys"
],
"repo": "score-spec/spec",
"url": "https://github.com/score-spec/spec/issues/21",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
35743872 | Grid images with In Home Streaming
There have been reports that grid images don't appear when viewing a stream able game. Instead it displays the file path.
If its not working for non-Ice shortcuts then I don't think there is anything I can do to fix the problem. As such, I am closing the task.
I hope Valve fixes this issue though. I can understand why they would want the client computer to be able to put in their own images, but I dont see any reason why they wouldn't get the custom images from the host computer by default if it has them.
Scott, since you're mucking with pysteam anyways, can you see if there's a way to get the client to automatically grab images for all the remote games it picks up?
Steam lets you manually assign custom grid view images to streamed-only games from a client side, so there should be a way to access that in the code. I personally can't find it right now since it appears to not be in shortcuts.vdf or the games list there, but you may have more insight having worked with this library longer.
Steam lets you manually assign custom grid view images to streamed-only games from a client side, so there should be a way to access that in the code. I personally can't find it right now since it appears to not be in shortcuts.vdf or the games list there, but you may have more insight having worked with this library longer.
Ah, interesting. I have a theory on how this is working, and if I'm right then it should be totally doable to set grid images from a client machine. I won't elaborate too much more though because I haven't looked into it at all and I'm really just guessing.
The bigger problem is - how do I get Ice on the client to know about the Ice that is running on the server? Ice wasn't really meant to handle that use case. I could provide a way to keep Ice running in the background and provide an API for a client to read the list of games (and even the images from the host!) but connecting the client Ice to the server Ice will be hard.
Once I get a streaming box of my own set up I'll try to work on some basic support for Steam In Home Streaming.
Please elaborate! With your familiarity, even a guess is better than my best attempt, and might point me in the right direction. I'm happy to dive into the code myself.
Ice could definitely just run on both machines, and check whatever non-shortcuts.vdf file holds all those references, then assume they are all Ice games. I wish Tags were visible remotely, that would fix any confusion for that.
| gharchive/issue | 2014-06-15T06:55:20 | 2025-04-01T06:45:44.915042 | {
"authors": [
"BlinksTale",
"scottrice"
],
"repo": "scottrice/Ice",
"url": "https://github.com/scottrice/Ice/issues/202",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2022032323 | Variable not in scope: pathParam In examples/basic.hs file
Hello!
I am new to Haskell and Scotty. I am processing examples in basic.hs and I encountered an error (look at the pic below).
this is my cabal file, I think I am using the latest version.
Name: scotty-examples
Version: 0.1
Synopsis: scotty examples
Homepage: https://github.com/scotty-web/scotty
Bug-reports: https://github.com/scotty-web/scotty/issues
License: BSD3
License-file: LICENSE
Author: Andrew Farmer <xichekolas@gmail.com>
Maintainer: Andrew Farmer <xichekolas@gmail.com>
Copyright: (c) 2012-Present, Andrew Farmer and the Scotty contributors
Category: Web
Stability: experimental
Build-type: Simple
Cabal-version: >= 1.10
Description: Example programs using @scotty@
tested-with: GHC == 8.10.7
, GHC == 9.0.2
, GHC == 9.2.8
, GHC == 9.4.6
, GHC == 9.6.2
executable CarShop
main-is: main.hs
default-language: Haskell2010
hs-source-dirs: src
build-depends: base >= 4.6 && < 5,
http-types,
mtl,
random,
scotty == 0.20.1,
text,
wai-extra
GHC-options: -Wall -threaded
I find it will work when using queryParam. I don't know why, but maybe it is better to fix the problem in the examples.
Finally, if you consider it as a problem, maybe I can try to make an easy-fix pull request for you.
Hi @fongfiafia , could you share the imports block in your main.hs ?
您好,来件已收到,我会尽快给您回复。
Hi @fongfiafia , could you share the imports block in your main.hs ?
sure~
{-# LANGUAGE OverloadedStrings #-}
module Main (main) where
import Web.Scotty
main :: IO ()
main = do
scotty 8081 $
get "/user" $ do
v <- pathParam "uid"
html $ mconcat ["<h1>", v, "</h1>"]
| gharchive/issue | 2023-12-02T13:02:01 | 2025-04-01T06:45:44.926090 | {
"authors": [
"fongfiafia",
"ocramz"
],
"repo": "scotty-web/scotty",
"url": "https://github.com/scotty-web/scotty/issues/355",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2257430350 | feat: add support for unqualified path expressions
Part of #29.
So, the tests are failing because we don't have Span::join is nightly-only, and we use the start span as a fallback 😿
| gharchive/pull-request | 2024-04-22T20:43:09 | 2025-04-01T06:45:44.943325 | {
"authors": [
"scrabsha"
],
"repo": "scrabsha/expandable",
"url": "https://github.com/scrabsha/expandable/pull/40",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
136588782 | Splash timeout on all requests.
I have set up docker to work with splash. Docker starts up fine but even the simple pages like google.com time out. Follows the splash terminal output for splash startup.
docker run -p 8050:8050 scrapinghub/splash
2016-02-26 03:39:07+0000 [-] Log opened.
2016-02-26 03:39:07.903064 [-] Splash version: 2.0.1
2016-02-26 03:39:07.903849 [-] Qt 5.5.1, PyQt 5.5.1, WebKit 538.1, sip 4.17, Twisted 15.5.0, Lua 5.2
2016-02-26 03:39:07.903997 [-] Python 3.4.3 (default, Oct 14 2015, 20:28:29) [GCC 4.8.4]
2016-02-26 03:39:07.904119 [-] Open files limit: 524288
2016-02-26 03:39:07.904248 [-] Open files limit increased from 524288 to 1048576
2016-02-26 03:39:08.112195 [-] Xvfb is started: ['Xvfb', ':1047', '-screen', '0', '1024x768x24']
2016-02-26 03:39:08.216409 [-] proxy profiles support is enabled, proxy profiles path: /etc/splash/proxy-profiles
2016-02-26 03:39:08.364839 [-] verbosity=1
2016-02-26 03:39:08.365024 [-] slots=50
2016-02-26 03:39:08.365640 [-] Web UI: enabled, Lua: enabled (sandbox: enabled)
2016-02-26 03:39:08.368755 [-] Site starting on 8050
2016-02-26 03:39:08.369102 [-] Starting factory <twisted.web.server.Site object at 0x7fb0ee1c4860>
Can you please help me with this.
Docker starts up fine but even the simple pages like google.com time out
how are you requesting those simple pages? Logs you posted dont contain any trace of requests being made to Splash, they just show Splash is started. If you posted your logs in full it means that requests you make dont reach Splash. Can you ping localhost:8050 and get demo page? When requests are timing out what error are you getting? Are they timing out with 504 status code or you dont get any status code?
hey Pawl,
Thanks.
I started docker with this command:
docker run -p 5023:5023 -p 8050:8050 -p 8051:8051 scrapinghub/splash
telnet "ping" output is:
telnet localhost 5023
Trying 127.0.0.1...
Connected to localhost.
Escape character is '^]'.
Connection closed by foreign host.
nmap command trace:
nmap -p 8050 localhost
Starting Nmap 7.00 ( https://nmap.org ) at 2016-02-26 21:50 PKT
Nmap scan report for localhost (127.0.0.1)
Host is up (0.00013s latency).
PORT STATE SERVICE
8050/tcp open unknown
Nmap done: 1 IP address (1 host up) scanned in 0.04 seconds
Now I write in address line Firefox (44.0.2): localhost:8050
I write in text-field provided next to "Render" Button. "http://google.com"(its already there btw) and press the button.
I am attaching compete logs of what the splash process outputs. If you look at the logs there is a problem that /etc/machine-id is not found. Well it is a problem in Ubuntu. I "patched" it, Can this be the problem?
Anyways after processing it gives Error 504
HTTP Error 504 (Gateway Time-out)
Type: GlobalTimeoutError
Timeout exceeded rendering page
{
"info": {
"timeout": 60
},
"description": "Timeout exceeded rendering page",
"error": 504,
"type": "GlobalTimeoutError"
}
Thanks again. :)
splash.txt
thanks for the detailed report. Is this happening for all splash endpoints or just execute? (when you press render on startpage it makes request to localhost:8050/execute).
can you try following url?
http://localhost:8050/render.json?url=http://httpbin.org
this makes request to render.json. is it responding with 504 too?
Yes it responds with 504 too
{"description": "Timeout exceeded rendering page", "error": 504, "type": "GlobalTimeoutError", "info": {"timeout": 30}}
attached are the logs
splash-json.txt
Thanks a lot.
Ok now this works fine. The only thing I have changed is my internet connection. Now I have a faster internet and it does not timeout.
The machine-id error is still there so it was not the case.
docker run -p 8050:8050 scrapinghub/splash --max-timeout 3600
Hello Ahmed.
I am retrieving the same mistake all the time
Some scripts is working fine because it runs in less than 30 seconds
Other script is always failing because it is slower. For this, each time that I execute it retrieves "GlobalTimeoutError"
The server is deploy inside "docker container" with the following parameters:
docker run -p 8050: 8050 --proxy-profiles-path / etc / splash / proxy-profiles --max-timeout 240 --filters-path / etc / splash / filters --js-profiles-path / etc / splash / Js-profiles --lua-sandbox-allowed-modules "populare; utils; popularetarj" -lua-package-path /etc/splash/lua_modules/?.lua --slots 500 -v 2
I feel that the configuration of "-max-timeout" is not working but I do not understand why. Whatever help is welcome.
Thank you in advance
Details:
{
"timestamp": 1491407822,
"user-agent": null,
"client_ip": "192.168.69.136",
"method": "POST",
"_id": 140363517724656,
"path": "/execute",
"load": [
0.09,
0.04,
0.05
],
"status_code": 504,
"error": {
"error": 504,
"info": {
"timeout": 240
},
"type": "GlobalTimeoutError",
"description": "Timeout exceeded rendering page"
},
"active": 0,
"maxrss": 1953936,
"qsize": 0,
"args": {
"wait": "0.5",
"uid": 140363517724656,
"expand": "1",
"lua_source": "local mymodule = require('popularetarj') function main(splash) splash.images_enabled = false splash.response_body_enabled = true args = {} args['user']='XXXX' args['pass']='XXXX' args['contract']='XXXX' args['fromDate']='05/03/2017' args['toDate']='05/04/2017' return mymodule.execute(splash, args) end",
"images": "1",
"timeout": "240.0",
"url": "http://google.com"
},
"fds": 28,
"rendertime": 30.02491569519043
}
@mgcuesta --max-timeout is only an upper bound of the timeout you can set per request. I didn't find a global config to set timeout for all requests though. If you want longer timeout you'll have to set them in each SplashRequest.
@FreeTymeKiyan How to set for each SplashRequest?
| gharchive/issue | 2016-02-26T03:51:27 | 2025-04-01T06:45:44.962429 | {
"authors": [
"AhmedRafikDjerah",
"FreeTymeKiyan",
"gotounix",
"mgcuesta",
"pawelmhm",
"shafaq"
],
"repo": "scrapinghub/splash",
"url": "https://github.com/scrapinghub/splash/issues/394",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
896917627 | use switch for platforms
using switch over the plain string is more idiomatic I guess
oh my, yep definitely nicer and more go looking than my hacky python :)! Thanks for this @hellt !
| gharchive/pull-request | 2021-05-20T14:38:03 | 2025-04-01T06:45:44.963849 | {
"authors": [
"carlmontanari",
"hellt"
],
"repo": "scrapli/scrapligo",
"url": "https://github.com/scrapli/scrapligo/pull/2",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1111639289 | 🛑 Animate Me is down
In b4687f2, Animate Me (https://animate-me.atanas.info) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Animate Me is back up in 1b024e8.
| gharchive/issue | 2022-01-22T17:51:16 | 2025-04-01T06:45:44.981220 | {
"authors": [
"scriptex"
],
"repo": "scriptex/uptime",
"url": "https://github.com/scriptex/uptime/issues/1133",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1108577037 | 🛑 Material Tetris is down
In dbb5eb0, Material Tetris (https://tetris.atanas.info) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Material Tetris is back up in 3dbcc8d.
| gharchive/issue | 2022-01-19T21:26:21 | 2025-04-01T06:45:44.983840 | {
"authors": [
"scriptex"
],
"repo": "scriptex/uptime",
"url": "https://github.com/scriptex/uptime/issues/640",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2337503924 | add row_usages to ChunkProof struct
Purpose or design rationale of this PR
add row_usages to ChunkProof struct
PR title
Your PR title must follow conventional commits (as we are doing squash merge for each PR), so it must start with one of the following types:
[ ] build: Changes that affect the build system or external dependencies (example scopes: yarn, eslint, typescript)
[ ] ci: Changes to our CI configuration files and scripts (example scopes: vercel, github, cypress)
[ ] docs: Documentation-only changes
[ ] feat: A new feature
[ ] fix: A bug fix
[ ] perf: A code change that improves performance
[ ] refactor: A code change that doesn't fix a bug, or add a feature, or improves performance
[ ] style: Changes that do not affect the meaning of the code (white-space, formatting, missing semi-colons, etc)
[ ] test: Adding missing tests or correcting existing tests
Deployment tag versioning
Has tag in common/version.go been updated or have you added bump-version label to this PR?
[ ] No, this PR doesn't involve a new deployment, git tag, docker image tag
[ ] Yes
Breaking change label
Does this PR have the breaking-change label?
[ ] No, this PR is not a breaking change
[ ] Yes
Codecov Report
All modified and coverable lines are covered by tests :white_check_mark:
Project coverage is 54.31%. Comparing base (c862cb6) to head (86e573b).
Additional details and impacted files
@@ Coverage Diff @@
## chore/libzkp-v0.11.0 #1366 +/- ##
=====================================================
Coverage 54.31% 54.31%
=====================================================
Files 156 156
Lines 11922 11922
=====================================================
Hits 6475 6475
Misses 4952 4952
Partials 495 495
Flag
Coverage Δ
coordinator
19.93% <ø> (ø)
database
42.85% <ø> (ø)
prover
26.55% <ø> (ø)
Flags with carried forward coverage won't be shown. Click here to find out more.
:umbrella: View full report in Codecov by Sentry.
:loudspeaker: Have feedback on the report? Share it here.
btw, we can merge this PR to develop branch and test the backward compatibility in sepolia. @amoylan2
| gharchive/pull-request | 2024-06-06T07:08:56 | 2025-04-01T06:45:45.015169 | {
"authors": [
"amoylan2",
"codecov-commenter",
"colinlyguo"
],
"repo": "scroll-tech/scroll",
"url": "https://github.com/scroll-tech/scroll/pull/1366",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
417660381 | Does it have a way to custom button icon?
I want to change the icons.
Anyone know how to do this?
Thank you!
https://github.com/scrumpy/tiptap/issues/193#issuecomment-461126841
| gharchive/issue | 2019-03-06T07:28:33 | 2025-04-01T06:45:45.017123 | {
"authors": [
"philippkuehn",
"tdkihrr"
],
"repo": "scrumpy/tiptap",
"url": "https://github.com/scrumpy/tiptap/issues/222",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
470177544 | substraTEE-client must accept a SHA256 hash of the WASM to be executed
Currently (M4), a WASM file has to be specified (or the default file is taken) and the SHA256 hash is calculated over the content.
The client is finally separated from the worker and has no access to his internals/files. Therefore, the user must be able to enter the SHA256 hash of the WASM code he/she wants to have executed in the TEE.
as of #61, WASM is no longer supported until we do #15
| gharchive/issue | 2019-07-19T07:20:44 | 2025-04-01T06:45:45.018369 | {
"authors": [
"brenzi",
"electronix"
],
"repo": "scs/substraTEE-worker",
"url": "https://github.com/scs/substraTEE-worker/issues/56",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1969948791 | fix: send OK message for ephemeral events
Hi, according to last changes in NIP01, the OK message MUST be sent in response to every EVENT message: https://github.com/nostr-protocol/nips/blob/4b9f13d983245e4dd166f102308afc28b8bb1603/01.md?plain=1#L153
Thanks for noticing and fixing this! Merged.
| gharchive/pull-request | 2023-10-31T08:55:11 | 2025-04-01T06:45:45.020245 | {
"authors": [
"scsibug",
"yukibtc"
],
"repo": "scsibug/nostr-rs-relay",
"url": "https://github.com/scsibug/nostr-rs-relay/pull/168",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
933807126 | Add softsegs in derivatives of spine-generic-processed
Description
The purpose of this issue is to discuss where we will include the softsegs in the derivatives of spine-generic-processed and how we will name them.
Suggestion
We could inlcude new soft segmentations under /derivatives/labels_softseg/ and have the same structure than under /derivatives/labels/
spine-generic-processed
│
├──derivatives
│ │
│ └── labels
│ │ ├── sub-amu01
│ │ │ ├── anat
│ │ │ │ ├── sub-amu01_acq-T1w_MTS_seg-manual.json
│ │ │ │ ├── sub-amu01_acq-T1w_MTS_seg-manual.nii.gz
│ │ │ │ ├── sub-amu01_acq-MTon_MTS_seg-manual.json
│ │ │ │ ├── sub-amu01_acq-MTon_MTS_seg-manual.nii.gz
│ │ │ │ ├── sub-amu01_T2star_seg-manual.json
│ │ │ │ ├── sub-amu01_T2star_seg-manual.nii.gz
│ │ │ │ ├── sub-amu01_T1w_seg-manual.json
│ │ │ │ ├── sub-amu01_T1wr_seg-manual.nii.gz
│ │ │ │ ├── sub-amu01_T2w_seg-manual.json
│ │ │ │ └── sub-amu01_T2w_seg-manual.nii.gz
│ │ │ └── dwi
│ │ │ ├── sub-amu01_rec-average_dwi_seg-manual.json
│ │ │ └── sub-amu01_rec-average_dwi_seg-manual.nii.gz
│ │ ...
│ └── labels_softseg
...
We also have to decide how to name the softsegs, the suffix -manual isn't appropriate here since softsegs are created by averaging manual segmentations.
Thank you for the heads up @sandrinebedard !
Tagging @kousu @alexfoias @copperwiring and @charleygros so they are aware
I have generated the soft segmentations, I would be ready to add them to spine-generic-processed dataset.
The plan for now would be to add them under derivatives/labels_softseg
We could directly omit the suffix -manual since they are reconstructed, we don't need a suffix necessarily for the derivatives from what I understand
sub-amu01_T2star_softseg.nii.gz
or we could also use the entity desc to differentiate with normal softseg since they are artificially generated here
Found this on probalistic segmentations : https://bids-specification.readthedocs.io/en/stable/05-derivatives/03-imaging.html#probabilistic-segmentations
If you have any insights/suggestions feel free to tell me!
also tagging @uzaymacar so that you are aware 😊!
to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz?
to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz?
although a big problem with that is that it is misleading: the soft segmentation does not come from softseg but from the average of multiple manual segs. So the suffix "manual" still makes sense, no?
to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz?
although a big problem with that is that it is misleading: the soft segmentation does not come from softseg but from the average of multiple manual segs. So the suffix "manual" still makes sense, no?
I would assume that the soft segmentations from softseg would be under /derivatives/labels/ and not /derivatives/labels_softseg, but yeah it may be still be confusing, we could absolutely add the suffix manual.
It would also depend on what we call the soft segmentations from softseg. For now, for SC segmentation, we add -manual since it was manually QC-ed or corrected, if we do the same for softsegs, the same confusion would occur...
to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz?
although a big problem with that is that it is misleading: the soft segmentation does not come from softseg but from the average of multiple manual segs. So the suffix "manual" still makes sense, no?
I would assume that the soft segmentations from softseg would be under /derivatives/labels/ and not /derivatives/labels_softseg, but yeah it may be still be confusing, we could absolutely add the suffix manual.
It would also depend on what we call the soft segmentations from softseg. For now, for SC segmentation, we add -manual since it was manually QC-ed or corrected, if we do the same for softsegs, the same confusion would occur...
given that the project is very specific to soft training, maybe we should in fact put those derivatives under derivatives/labels_softseg. This is not a bad idea. In that case we would not need the suffix "-manual" because we could add a README.md inside derivatives/labels_softseg/ that clearly explains how these labels were generated.
| gharchive/issue | 2021-06-30T15:04:17 | 2025-04-01T06:45:45.031354 | {
"authors": [
"jcohenadad",
"sandrinebedard"
],
"repo": "sct-pipeline/contrast-agnostic-softseg-spinalcord",
"url": "https://github.com/sct-pipeline/contrast-agnostic-softseg-spinalcord/issues/10",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
568337555 | request: reliable means to interrogate sparsity
It would be very helpful if the API had a standard way to query the anndata X sparsity, and sparse format, independent of the backed mode.
In other words, a safe / reliable / performant equivalent equivalent of the following on memory-backed adata:
scipy.sparse.issparse(adata.X) -- bool indicates if is sparse
adata.X.get_format() -- return scipy sparse format, eg, 'csr'
Is there a way to do this now? Or a hack to work around the lack of API today?
One possible hack-around relies on the weak test for the format_str internal to spmatrix (these are overly wordy for clarity):
from scipy import sparse
def issparse(adata):
if adata.isbacked:
return hasattr(adata.X, 'format_str')
else:
return sparse.issparse(adata.X)
def get_sparse_format(adata):
if issparse(adata):
if adata.isbacked:
return getattr(adata.X, 'format_str', 'und')
else:
return adata.X.getformat()
return False
If you just want to test sparsity, and not if X is some other kind of array-like thing, you’d try scipy.sparse.issparse(adata.X) or isinstance(anndata.X, scipy.sparse.spmatrix). For the sparse format, you’d use scipy.sparse.isspmatrix_csr or isinstance(..., ...csr_matrix), right?
This should actually work for BackedSparseMatrix, as all its subclasses actually inherit from the spmatrix subclasses:
https://github.com/theislab/anndata/blob/4440b90ff3dff213b4c512478e21426cf210368d/anndata/_core/sparse_dataset.py#L117
https://github.com/theislab/anndata/blob/4440b90ff3dff213b4c512478e21426cf210368d/anndata/_core/sparse_dataset.py#L143
If you just had plain scipy.sparse classes and wanted to test sparsity, you’d use scipy.sparse.issparse(adata.X) or isinstance(anndata.X, scipy.sparse.spmatrix), right?
Yes, that would be perfect.
For the sparse format, you’d use scipy.sparse.isspmatrix_csr or isinstance(..., ...csr_matrix), right?
I would probably use spmatrix.getformat(), but that is equivalent to scipy.sparse.isspmatrix_csr(), and both should work if adata.X is a subclass of the appropriate spmatrix.
That said, in 0.7.1, it doesn't seem to work. The type of adata.X is a SparseMatrix, not a BackedSparseMatrix.
>>> adata = anndata.read_h5ad('pbmc3k-CSC-gz.h5ad', backed='r')
>>> adata
AnnData object with n_obs × n_vars = 2638 × 1838 backed at 'pbmc3k-CSC-gz.h5ad'
obs: 'n_genes', 'percent_mito', 'n_counts', 'louvain'
var: 'n_cells'
uns: 'draw_graph', 'louvain', 'louvain_colors', 'neighbors', 'pca', 'rank_genes_groups'
obsm: 'X_pca', 'X_tsne', 'X_umap', 'X_draw_graph_fr'
varm: 'PCs'
>>> adata.X
<HDF5 sparse dataset: format 'csc', shape (2638, 1838), type '<f4'>
>>> type(adata.X)
<class 'anndata._core.sparse_dataset.SparseDataset'>
>>> scipy.sparse.isspmatrix(adata.X)
False
>>> scipy.sparse.isspmatrix_csc(adata.X)
False
>>> isinstance(adata.X, scipy.sparse.spmatrix)
False
>>> isinstance(adata.X, scipy.sparse.csc_matrix)
False
The SparseMatrix class has some intriguing parameters (eg, format_str), but I assume those are private?
That’s very weird, X is supposed to be a matrix class, not a HDF5 dataset.
LMK if you need a more complete test case.
@bkmartinjr You should be able to get some idea from type(adata.X).
Do you have a use case where it's important to know whether it's dense or sparse, but it's not important to know whether it's in memory or on disk?
It might be nice to have this in the future, but I'd really like anndata to be more agnostic about the type of the array. That is, someone defines an array type and it works fine without us having to approve it. These trait detection methods should also be implemented in upstream packages (I've suggested this here: https://github.com/scipy/scipy/issues/11565).
For right now, I think you could get away with:
if isinstance(X, (scipy.sparse.spmatrix, SparseDataset)):
# It's sparse
elif isinstance(X, (np.ndarray, h5py.Dataset)):
# It's dense
@flying-sheep if the AnnData is backed and not a view, X is a h5py.Dataset or our SparseDataset.
SparseDataset should get exported, and we should make its to_backed() method private.
Why isn’t X a matrix-like class that produces matrices when sliced?
You should be able to get some idea from type(adata.X)
True, but doing that is what got this entire conversation started. AFAIK, there are no public types usable in that manner. As a hack, it is definitely an alternative. I had picked format_str instead of the type, as it seemed to be more broadly backward compatible.
It might be nice to have this in the future, but I'd really like anndata to be more agnostic
Fully agree with this, as long as you define the minimum interface that the array must support. If that abstract interface is sufficient, all would work fine.
I consider sparsity to be necessary in that interface, as it is very hard to write high performance code, for large matrices, without knowing something about how you should partition and stride through it. It is more important, IMHO, than isbacked.
For right now, I think you could get away with:
This is similar to the hack I was using before, which broke when the internal class structure changed.
TL;DR - I like your "agnostic" approach, but what it needs to work IMHO is:
a "guaranteed" interface for all array-ish things that can be stashed in X
if there are modes of behavior for X (like sparse/dense), interfaces that let users introspect these modes
Coming in because I've ran into the exact same issue. Specifically I'm wondering if this will be fixed
adata = anndata.read("file.h5ad", "r")
type(adata.X)
# anndata._core.sparse_dataset.SparseDataset
adata = anndata.read("file.h5ad")
type(adata.X)
# scipy.sparse.csr.csr_matrix
Why is the type of X SparseDataset when read in backed mode?
With #219 (backed zarr), the incantation for this would get harder. what’s the plan here @ivirshup?
Would still like to not implement our own solution, especially one we export, since I don't think we can reliably cover all cases. I would ideally like scipy to make issparse single dispatch, but idk if that's going to happen.
Realistically I think the thing to do will be something like:
SPARSE_TYPES = (scipy.sparse.sparray, scipy.sparse.spmatrix, ...)
isinstance(x, SPARSE_TYPES)
| gharchive/issue | 2020-02-20T14:25:59 | 2025-04-01T06:45:45.061646 | {
"authors": [
"bkmartinjr",
"flying-sheep",
"ivirshup",
"pablo-gar"
],
"repo": "scverse/anndata",
"url": "https://github.com/scverse/anndata/issues/326",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
837525473 | Opening file with "w" while it's open in backed mode elsewhere still deletes file contents
[x] I have checked that this issue has not already been reported.
[x] I have confirmed this bug exists on the latest version of scanpy. - Version '1.7.1'
[ ] (optional) I have confirmed this bug exists on the master branch of scanpy.
When I open adata in backed='r' mode in one script and then I try to modify the same adata/file from another script the adata file gets corrupted (its size becomes 0). I would expect writing to return an error before corrupting the file.
Versions
-----
anndata 0.7.4
scanpy 1.7.1
sinfo 0.3.1
-----
PIL 7.2.0
anndata 0.7.4
backcall 0.2.0
cairo 1.19.1
cffi 1.14.0
cloudpickle 1.3.0
colorama 0.4.4
cycler 0.10.0
cython_runtime NA
dask 2.21.0
dateutil 2.8.0
decorator 4.4.2
future_fstrings NA
get_version 2.1
google NA
h5py 2.10.0
igraph 0.8.2
ipykernel 5.3.3
ipython_genutils 0.2.0
jedi 0.17.2
joblib 0.16.0
kiwisolver 1.2.0
legacy_api_wrap 0.0.0
leidenalg 0.8.1
llvmlite 0.35.0
louvain 0.6.1
matplotlib 3.3.0
mpl_toolkits NA
natsort 7.0.1
numba 0.52.0
numexpr 2.7.1
numpy 1.19.4
packaging 20.9
pandas 1.0.5
parso 0.7.0
pexpect 4.8.0
pickleshare 0.7.5
pkg_resources NA
prompt_toolkit 3.0.5
psutil 5.7.2
ptyprocess 0.6.0
pygments 2.6.1
pyparsing 2.4.7
pytz 2020.1
scanpy 1.7.1
scipy 1.4.1
setuptools_scm NA
simplejson 3.17.2
sinfo 0.3.1
six 1.15.0
sklearn 0.23.1
storemagic NA
tables 3.6.1
texttable 1.6.2
tlz 0.10.0
toolz 0.10.0
tornado 6.0.4
traitlets 4.3.3
typing_extensions NA
wcwidth 0.2.5
yaml 5.3.1
zmq 19.0.1
zope NA
-----
IPython 7.16.1
jupyter_client 6.1.6
jupyter_core 4.6.3
notebook 6.0.3
-----
Python 3.8.5 | packaged by conda-forge | (default, Jul 22 2020, 17:31:50) [GCC 7.5.0]
Linux-3.10.0-1160.11.1.el7.x86_64-x86_64-with-glibc2.10
64 logical CPU cores, x86_64
-----
Session information updated at 2021-03-21 21:14
</details>
@Hrovatin, I'd check out the issue at h5py for more info. As it pertains to AnnData though, what exactly were you trying to do?
Did you want the file to be overwritten by the second process, or were you looking for it to update it? E.g. would you expect one of those processes to error?
I had file open in one process in basked mode. I then decided that I need to make some stuff in another notebook and forgot about having the file open in the first notebook. When I tried to save the file from the second notebook it corrupted it. I would expect it at least to err out before corrupting it, so that I could close the file in the first notebook before trying to save it in the second notebook.
I think not letting you modify a file while it is opened as basked somewhere else is the safest option.
Any update on this? yet another of my adatas was deleted this way.
@ivirshup Maybe a temporary solution would be info on how to close a backed file - now I usually just copy info from adata and delete the object immediately, but if there was a "close" function that would be nice to know.
@ivirshup any updates on this? Alternatively, is there a good way to check if anndata file is open somewhere before saving?
@Hrovatin, no updates yet. But reading this with fresh eyes, I might have something.
What do you mean by corrupted? Is the data not just overwritten? Or does it fail to overwrite somehow?
Looking some more and found an old issue I opened: https://github.com/h5py/h5py/issues/1864
It looks like this is basically an upstream bug (link to jira tracker, a little difficult to use), and it's unclear if/ when it will be fixed. It looks like there isn't a clean solution from python.
One thing that could work is if the backed file was opened with h5py.File(..., locking=False). Then the file should just be overwritten.
I'm not sure if this is a good default, since it changes behavior from hdf5. Ideally we would error before any data gets truncated. However, we could allow passing that argument through when reading a file in backed mode.
Whatever you can do to prevent deletion of files would be very useful for me, especially now when we no longer have snapshots on the server. Because sometimes I forget to shut down a notebook where I have something backed opened and then data is gone.
If you want to just prevent deletion in general, we could let you pass "w-" or "x" to write_h5ad. But this is specifically if you don't want to overwrite data regardless of whether any other process has it open.
No, I would like to overwrite files, but prevent that they get corrupted if another process has it open. - I often update adatas so I want to change (rewrite) existing object on disk.
Seems like this is a valid use case, and we’re just not clear how to enable it.
Or is something still unclear about this?
| gharchive/issue | 2021-03-21T20:18:56 | 2025-04-01T06:45:45.071437 | {
"authors": [
"Hrovatin",
"flying-sheep",
"ivirshup"
],
"repo": "scverse/anndata",
"url": "https://github.com/scverse/anndata/issues/522",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1262764397 | suddenly unable to read .h5ad file
there were no problem reading .h5ad file but all of the sudden, this error came up and unable to read .h5ad files,,
someone have any idea??
Thank you!!
It was probably written by a newer anndata version. Try updating anndata, maybe that helps.
We have a PR that improves error messages in this regard. @ivirshup could you please merge or adapt and then merge #734?
Thanks for your help! I have updated anndata and h5py and re-start my jupyter noteobok and it works!!
| gharchive/issue | 2022-06-07T06:02:05 | 2025-04-01T06:45:45.074384 | {
"authors": [
"flying-sheep",
"parkjooyoung99"
],
"repo": "scverse/anndata",
"url": "https://github.com/scverse/anndata/issues/780",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2732938152 | Error in import squidpy: cannot import name 'SparseCSCView' from 'anndata._core.views'
Description
Hi, I'm trying to use squidpy but when I try to import it returns an error. It seems to search for a function in anndata which I couldn't find. I might just have misunderstood the error, and if that's the case, I apologize. Could you please help me understand or resolve this issue?
Thanks in advance
...
Minimal reproducible example
#the package was downloaded from conda with other packages like sainsc. I'm running the code in this environment on Spyder 6. Anndata version=0.11.1
import squidpy as sq
...
Traceback
ImportError Traceback (most recent call last)
Cell In[18], line 1
----> 1 import squidpy as sq
File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/__init__.py:5
1 from __future__ import annotations
3 from importlib import metadata
----> 5 from squidpy import datasets, gr, im, pl, read, tl
7 try:
8 md = metadata.metadata(__name__)
File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/gr/__init__.py:5
1 """The graph module."""
3 from __future__ import annotations
----> 5 from squidpy.gr._build import spatial_neighbors
6 from squidpy.gr._ligrec import ligrec
7 from squidpy.gr._nhood import centrality_scores, interaction_matrix, nhood_enrichment
File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/gr/_build.py:31
29 from squidpy._docs import d, inject_docs
30 from squidpy._utils import NDArrayA
---> 31 from squidpy.gr._utils import (
32 _assert_categorical_obs,
33 _assert_positive,
34 _assert_spatial_basis,
35 _save_data,
36 )
38 __all__ = ["spatial_neighbors"]
41 @d.dedent
42 @inject_docs(t=Transform, c=CoordType)
43 def spatial_neighbors(
(...)
56 copy: bool = False,
57 ) -> tuple[csr_matrix, csr_matrix] | None:
File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/gr/_utils.py:12
10 import pandas as pd
11 from anndata import AnnData
---> 12 from anndata._core.views import ArrayView, SparseCSCView, SparseCSRView
13 from anndata.utils import make_index_unique
14 from pandas import CategoricalDtype
ImportError: cannot import name 'SparseCSCView' from 'anndata._core.views' (/home/onco-lab/anaconda3/envs/sainsc/lib/python3.12/site-packages/anndata/_core/views.py)
...
Version
Name Version Build Channel
squidpy 1.5.0 pyhd8ed1ab_0 conda-forge
Anndata version=0.11.1
...
https://github.com/scverse/squidpy/issues/915
I also had the same issue and could fix it by installing version 1.6.2.
| gharchive/issue | 2024-12-11T13:08:39 | 2025-04-01T06:45:45.080941 | {
"authors": [
"HeesooSong",
"diego-iac"
],
"repo": "scverse/squidpy",
"url": "https://github.com/scverse/squidpy/issues/928",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2638467631 | scylla_node: ignore scylla-tools files if scylla-tools is missing
This is a follow-up to https://github.com/scylladb/scylla-ccm/pull/619. Patch scylla_node.py so every access to java tools dir is conditional on the dir existing in the first place. With this PR, tests are passing with the java tools dropped form the unified package (https://github.com/scylladb/scylladb/pull/20740).
Refs: https://github.com/scylladb/scylladb/issues/14856
@denesb are you sure you tested ontop of #619 ?
since it was merged 4 hours ago, and you have a conflict with it.
@denesb are you sure you tested ontop of #619 ?
since it was merged 4 hours ago, and you have a conflict with it.
Ah, seems to be the same thing. I guess my PR is not needed after all. Closing.
| gharchive/pull-request | 2024-11-06T15:39:20 | 2025-04-01T06:45:45.093034 | {
"authors": [
"denesb",
"fruch"
],
"repo": "scylladb/scylla-ccm",
"url": "https://github.com/scylladb/scylla-ccm/pull/623",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1425590323 | Update files.rst
Fixing typo
Muchas gracias por la corrección! Saludos
| gharchive/pull-request | 2022-10-27T12:53:55 | 2025-04-01T06:45:45.281855 | {
"authors": [
"quobit",
"sdelquin"
],
"repo": "sdelquin/aprendepython",
"url": "https://github.com/sdelquin/aprendepython/pull/32",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
182256277 | commit fails after file rename
colemanbroaddus@myers-mac-10-display-1 ~/P/cell_tracker> gl status
On branch master, repo-directory //
Tracked files with modifications:
➜ these will be automatically considered for commit
➜ use gl untrack f if you don't want to track changes to file f
➜ if file f was committed before, use gl checkout f to discard local changes
master.py
truetree.py (new file)
viewer.py (deleted)
Untracked files:
➜ these won't be considered for commit
➜ use gl track f if you want to track changes to file f
There are no untracked files to list
colemanbroaddus@myers-mac-10-display-1 ~/P/cell_tracker> gl commit -m "Merged changes from nodeless."
✘ Some internal error occurred
➜ If you want to help, see http://gitless.com for info on how to report bugs and include the following information:
0.8.3
Traceback (most recent call last):
File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.cli.gl", line 69, in main
File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.cli.gl_commit", line 58, in main
File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.core", line 1179, in create_commit
File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.core", line 1152, in get_tree_and_update_index
File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.core", line 1145, in update
File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/pygit2.index", line 168, in remove
File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/pygit2.errors", line 53, in check_error
IOError: Index does not contain viewer.py at stage 0
This happened directly after a merge and resolve, although they did not include the file causing the error: (viewer.py renamed to truetree.py).
| gharchive/issue | 2016-10-11T12:54:59 | 2025-04-01T06:45:45.294422 | {
"authors": [
"colemanbroad"
],
"repo": "sdg-mit/gitless",
"url": "https://github.com/sdg-mit/gitless/issues/63",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
725377654 | Changing name and website
In Competitive programming list,the name Code Fights is was changed to Code Signal in 2018.So I changed the website and name of CodeFights to CodeSignal.
Thanks
| gharchive/pull-request | 2020-10-20T09:13:49 | 2025-04-01T06:45:45.305557 | {
"authors": [
"nyilynnhtwe",
"sdmg15"
],
"repo": "sdmg15/Best-websites-a-programmer-should-visit",
"url": "https://github.com/sdmg15/Best-websites-a-programmer-should-visit/pull/618",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
474405694 | add images
add images
Please add some more information about the problem or feature request
| gharchive/issue | 2019-07-30T06:51:14 | 2025-04-01T06:45:45.314133 | {
"authors": [
"Saileek-bidchat",
"max-carroll"
],
"repo": "sdoomz/react-google-picker",
"url": "https://github.com/sdoomz/react-google-picker/issues/20",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
346998052 | Make it possible to ignore usage of invalid SSL certificates
Since #15 was closed the world has gotten better, with the merge of https://github.com/sfackler/rust-native-tls/pull/56
Since the native_tl crate now support this, it should be easy to implement it in reqwest.
This probably requires bumping the native-tls crate to 0.2.0 from 0.1.5 though.
Could we help to implement this feature?
Certainly! To get this into 0.8.x, it'd probably be easiest to just inline the code that reqwest uses from hyper-tls and tokio-tls (since tokio-tls is stagnating/abandoned, and the code duplicate code wouldn't be much), and update it to work with native-tls 0.2.
I started the #325 PR. Any policy to implement test for this feature?
This feature is part of the v0.9 release.
| gharchive/issue | 2018-08-02T12:51:51 | 2025-04-01T06:45:45.398616 | {
"authors": [
"adamtulinius",
"seanmonstar",
"yageek"
],
"repo": "seanmonstar/reqwest",
"url": "https://github.com/seanmonstar/reqwest/issues/324",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
477474073 | Algorithm runs exceptionally slow
With addition of 2 children per crossover, alg runs insanely slow. Looking into issue now
Selection was chunking parents every 1 step, doubling the number of parents necessary. Fixed with last commit.
| gharchive/issue | 2019-08-06T16:17:50 | 2025-04-01T06:45:45.399592 | {
"authors": [
"seanmor5"
],
"repo": "seanmor5/genex",
"url": "https://github.com/seanmor5/genex/issues/17",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
536783457 | implementation DatasourceProxy without AlibabaDruid
Why you need it?
DatasourceProxy design reason is adapt to more difference JDBC Datasource implementation(Hikaricp, c3p0...)
How it could be?
io.seata.rm.datasource.sql.SQLVisitorFactory using AlibabaDruid SQLStatement, We can use JSqlParser or custom SqlUtils.
refer: https://github.com/seata/seata/pull/1703
after 1.0
Thanks!
@funcfoo Seat just uses druid internally as sqlparser, and the database connection pool used by the business can be arbitrary.
Yes, I know that.
@slievrly Yes, I know that. but I can't agree with this design. We just use a little library, Why using a Heavyweight library.
| gharchive/issue | 2019-12-12T06:29:47 | 2025-04-01T06:45:45.405356 | {
"authors": [
"funcfoo",
"slievrly",
"zjinlei"
],
"repo": "seata/seata",
"url": "https://github.com/seata/seata/issues/2025",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1082074566 | optimize:simplify the interface of console
[ ] I have registered the PR changes.
Ⅰ. Describe what this PR did
简化控制台接口数量,提高拓展性:
1.控制台接口合并,使用对象作为入参;
2.防止查询条件变更时,要增加接口提供拓展;
老接口:
PageResult<GlobalLockVO> queryByTable(String tableName);
PageResult<GlobalLockVO> queryByXid(String xid);
新接口:
PageResult<GlobalLockVO> query(GlobalLockParam param);
public class GlobalLockParam implements Serializable {
private static final long serialVersionUID = 615412528070131284L;
private String xid;
private String tableName;
}
Ⅱ. Does this pull request fix one issue?
Ⅲ. Why don't you add test cases (unit test/integration test)?
Ⅳ. Describe how to verify it
Ⅴ. Special notes for reviews
Codecov Report
Merging #4212 (d3b7568) into develop (caa8443) will decrease coverage by 0.04%.
The diff coverage is 0.00%.
:exclamation: Current head d3b7568 differs from pull request most recent head 9d7a43b. Consider uploading reports for the commit 9d7a43b to get more accurate results
@@ Coverage Diff @@
## develop #4212 +/- ##
=============================================
- Coverage 49.10% 49.06% -0.05%
+ Complexity 3774 3771 -3
=============================================
Files 719 722 +3
Lines 23948 23959 +11
Branches 2926 2926
=============================================
- Hits 11760 11755 -5
- Misses 10987 11000 +13
- Partials 1201 1204 +3
Impacted Files
Coverage Δ
...o/seata/core/console/param/BranchSessionParam.java
0.00% <0.00%> (ø)
...a/io/seata/core/console/param/GlobalLockParam.java
0.00% <0.00%> (ø)
...o/seata/core/console/param/GlobalSessionParam.java
0.00% <0.00%> (ø)
...java/io/seata/core/console/vo/BranchSessionVO.java
0.00% <ø> (ø)
...in/java/io/seata/core/console/vo/GlobalLockVO.java
0.00% <ø> (ø)
...java/io/seata/core/console/vo/GlobalSessionVO.java
0.00% <ø> (ø)
...er/console/controller/BranchSessionController.java
50.00% <ø> (ø)
...erver/console/controller/GlobalLockController.java
50.00% <0.00%> (ø)
...er/console/controller/GlobalSessionController.java
33.33% <0.00%> (+8.33%)
:arrow_up:
...er/console/impl/db/BranchSessionDBServiceImpl.java
0.00% <ø> (ø)
... and 11 more
改动点:
1.vo添加tostring;
2.部分类添加序列化;
3.接口合并;
4.result类移动位置;
5.分页结果pageResult部分无用方法删除;
| gharchive/pull-request | 2021-12-16T11:21:39 | 2025-04-01T06:45:45.422566 | {
"authors": [
"codecov-commenter",
"lightClouds917"
],
"repo": "seata/seata",
"url": "https://github.com/seata/seata/pull/4212",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2159152510 | Update ChartRendererConfigOptions interface colors
Update ChartRendererConfigOptions interface colors accordingly https://docs.seats.io/docs/renderer/colors
Thanks for creating this pull request, we missed the type update for this one.
| gharchive/pull-request | 2024-02-28T14:50:39 | 2025-04-01T06:45:45.437581 | {
"authors": [
"RafaelAlkmimDias",
"mortendevold"
],
"repo": "seatsio/seatsio-types",
"url": "https://github.com/seatsio/seatsio-types/pull/8",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1126522786 | Don't include screens without questions in total block count
To determine if an application is ok to submit or requires more answers from the applicant, we count the number of active blocks and compare that with the number of active and completed blocks. This caused a bug where blocks that have no questions would count towards the total of active but incomplete blocks, making it impossible to complete the application.
This change alters the logic for tallying completed blocks by not including blocks that have no questions in the tally.
closing in favor of an alternative approach: https://github.com/seattle-uat/civiform/pull/1894
| gharchive/pull-request | 2022-02-07T21:33:38 | 2025-04-01T06:45:45.439412 | {
"authors": [
"bion"
],
"repo": "seattle-uat/civiform",
"url": "https://github.com/seattle-uat/civiform/pull/1893",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
806870904 | The ability to bring up a full dev stack which will automatically tear itself down.
Description
There is one TODO - the actual ARN for the role that needs to bring up the stack needs to change, or else the perms for that role need to be changed. Waiting on John for that. So I figured I'd send this for review and then edit that before submitting.
Let's see if that testing github action works! (Will add reviewers once I see whether that needs changes)
Checklist
[ ] Created tests which fail without the change (if possible)
[ ] Extended the README / documentation, if necessary
Cool - I think that's right.
| gharchive/pull-request | 2021-02-12T01:10:40 | 2025-04-01T06:45:45.441382 | {
"authors": [
"ndmckinley"
],
"repo": "seattle-uat/universal-application-tool",
"url": "https://github.com/seattle-uat/universal-application-tool/pull/144",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2502228010 | updated modal related files to make sure the double scroll bar will n…
fix: modal double scroll bar issue fixed
Continued in #1572
| gharchive/pull-request | 2024-09-03T08:20:13 | 2025-04-01T06:45:45.452156 | {
"authors": [
"GauriKudtarkar",
"splashdust"
],
"repo": "seb-oss/green",
"url": "https://github.com/seb-oss/green/pull/1553",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2095106296 | Make templates able to access instVars of the object sent as rendering context
Make the syntax of the template to be cleaner.
For example, given the following model:
model := STTDummyJobModel new
firstName: 'Questioner';
answer: 42;
yourself.
Instead of
<div><st= self firstName, ' ', self answer asString></div>
It should work if you don't use the accessors too, like in:
<div><st= firstName, ' ', answer asString></div>
PR here https://github.com/sebastianconcept/STTemplate/pull/10
Merged
| gharchive/issue | 2024-01-23T02:12:44 | 2025-04-01T06:45:45.468761 | {
"authors": [
"sebastianconcept"
],
"repo": "sebastianconcept/STTemplate",
"url": "https://github.com/sebastianconcept/STTemplate/issues/9",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
874353209 | Error loading module 'tabline'
Hey seblj
I install this plugin throw vim-plug
After sourcing the lines in my init.vim
require('tabline').setup{
no_name = '[No Name]', -- Name for buffers with no name
modified_icon = '', -- Icon for showing modified buffer
close_icon = '', -- Icon for closing tab with mouse
separator = "▌", -- Separator icon on the left side
space = 3, -- Prefix and suffix space
color_all_icons = false, -- Color devicons in active and inactive tabs
always_show_tabs = false, -- Always show tabline
right_separator = false, -- Show right separator on the last tab
-- highlights = { -- Default colors for highlights.
-- fill = {
-- guifg = M.inactive_background,
-- guibg = M.inactive_background
-- },
-- filename_active = {
-- guifg = M.active_text,
-- guibg = M.active_background,
-- gui = 'bold,italic'
-- },
-- filename_inactive = {
-- guifg = M.inactive_text,
-- guibg = M.inactive_background
-- },
-- padding_active = {
-- guifg = M.active_background,
-- guibg = M.active_background
-- },
-- padding_inactive = {
-- guif = M.inactive_background,
-- guibg = M.inactive_background
-- },
-- separator_active = {
-- guifg = M.active_separator,
-- guibg = M.active_background
-- },
-- separator_inactive = {
-- guifg = M.inactive_text,
-- guibg = M.inactive_background
-- },
-- modified_active = {
-- guifg = M.active_text,
-- guibg = M.active_background
-- },
-- modified_inactive = {
-- guifg = M.inactive_text,
-- guibg = M.inactive_background
-- },
-- close_active = {
-- guifg = M.active_text,
-- guibg = M.active_background
-- },
-- close_inactive = {
-- guifg = M.inactive_text,
-- guibg = M.inactive_background
-- }
-- }
}
I think the problem is with how you require things. Do you have a file called tabline.lua where you setup the plugin? If so, lua gets confused on which file to require since your config file, and my plugin is called the same thing. You should try to add another folder inside the lua folder (for example named config). Then you can do require('config.tabline') from your init.vim.
Your path to your config for the plugin will then be: /home/va/.config/nvim/lua/config/tabline.lua
Ok that fixed the ISSUE 👍🏻
| gharchive/issue | 2021-05-03T08:26:31 | 2025-04-01T06:45:45.483762 | {
"authors": [
"seblj",
"terminal-guy"
],
"repo": "seblj/nvim-tabline",
"url": "https://github.com/seblj/nvim-tabline/issues/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
394861824 | Unsupported "Accept" header
I have run the app in env test. I have attached env file too. I get the below error
env.txt
You forgot to specify Accept in your request headers. It's supposed to equal "application/json"
You'll also need to pass Content-Type with the same value
how do I use Accep in header
| gharchive/issue | 2018-12-30T10:31:51 | 2025-04-01T06:45:45.517994 | {
"authors": [
"hlogeon",
"thomsonimjd"
],
"repo": "secret-tech/backend-ico-dashboard",
"url": "https://github.com/secret-tech/backend-ico-dashboard/issues/193",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2243298845 | 【文档】真是部署中,数据是怎么传输的
Hi
我在看文档使用SPU进行逻辑回归。
我理解,该文档对应的是单机演示环境的。那我有如下几个问题,
假设两方Alice(192.168.1.2 有x_1) 和Bob(192.168.1.3 有x_2 和y)
该文档的代码,是在Alice 执行还是在Bob执行
假设是在Alice执行,那Bob的 x_2 和y 如何传递给Alice(或Alice的代码应该做怎样的调整?)即要在Alice的机器上执行什么操作,在Bob的机器上执行什么操作?
Reproduction code to reproduce the issue.
1
请查看这个文档
| gharchive/issue | 2024-04-15T10:29:22 | 2025-04-01T06:45:45.520636 | {
"authors": [
"anakinxc",
"gxcuit"
],
"repo": "secretflow/secretflow",
"url": "https://github.com/secretflow/secretflow/issues/1251",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1732274507 | Update awesome-pets.md
Added my github id as a contributor to the zkp branch.
Thank you for your approval !
Hi, would you mind to paste your merged PR here? Just to make sure everything is on track. Thanks
@zlyber @Jamie-Cui https://github.com/secretflow/secretflow/pull/597
Thanks @xfap , I'm merge this PR
| gharchive/pull-request | 2023-05-30T13:24:26 | 2025-04-01T06:45:45.522667 | {
"authors": [
"Jamie-Cui",
"xfap",
"zlyber"
],
"repo": "secretflow/secretflow",
"url": "https://github.com/secretflow/secretflow/pull/602",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
832400094 | Angular 11 - JWT Authentication Example & Tutorial With PHP
Article checklist
NOTE: (Please ensure that you have only one open issue + linked pull request at a time. This will ensure that we complete the article in a timely manner from inception to publishing.)
Preliminary Checks - Formatting and Structure
[x] Does your article follow any of the suggested structure formats? - see example formats
[x] Is your article properly formatted in Markdown? - see Github Markdown guide
[x] Have you used the correct folder and file structure? - see Contribution Guidelines
[x] Is your article over 750 words? - Tool: Wordcounter.net
[x] Have you used our preferred heading structure? - H3 (###) and up
[x] Does your article provide enough value and detail about your topic? - Articles should be clear, accurate, and fully explained.
[x] Can your article be understood by beginners? - Assume the audience is smart but has no prior exposure to the common terminology in your article.
[x] Have you included a hero image that is 600x400 pixels and have the copyright to use it? - Tools: Pixlr Image Editor for resizing and Unsplash for Creative Commons images
[x] In your PR - add a comment that states "This fixes #(enter your Idea Content Form #)" to link and close your Idea Content (issue) when the PR is merged.
Grammar & plagiarism checks
[x] Have you spellchecked and grammar checked your article? - Tools: Write&Improve - ProWritingAid - Grammarly
[x] Have you placed your article through a 3rd party plagiarism checker? We suggest using Quetext, this tool is free and has a daily limit. We typically accept articles with 10% or less.
[x] Have you checked your article for readability? - Tool: Hemmingway
[x] Have you added sources for quotes and images that aren't yours?
Technical checks
[x] Are your code snippets properly formatted for syntax highlighting - see Syntax guide
Contribution guidelines
For first-time contributors and for more details, see Contributing Guidelines
This fixes #1691
Hello @jumamiller. The title of this article conflicts with that of another already published article. See https://www.section.io/engineering-education/getting-started-with-jwt-using-angular8-and-nodejs/. The article talks about jwt in angular 11 using a node.js server.
Looking at your article, seems like there's some PHP used. Just a suggestion, maybe you could update the title to include PHP, to differentiate it from the other article.
@paulodhiambo, done!
@jumamiller an amazing article on Angular 11 authentication. @hectorkambow this article is ready for final review.
@jumamiller 👋 Good afternoon - Upon running your article through our 3rd party plagiarism checker it seemed to raise a few flags and the % was higher than we typically accept.
Please see attached PDF - and revisit the article to ensure we are contributing wholly unique and original content.
Be sure to see our resources page to see more info on plagiarism and what is considered as such.
🤔
Angular 11 - JWT Authentication Example & Tutorial With PHP #1699.pdf
@hectorkambow, I've updated most of the boilerplate codes, you can recheck!
cc @paulodhiambo
| gharchive/pull-request | 2021-03-16T04:32:44 | 2025-04-01T06:45:45.537000 | {
"authors": [
"geoffrey45",
"hectorkambow",
"jumamiller",
"paulodhiambo"
],
"repo": "section-engineering-education/engineering-education",
"url": "https://github.com/section-engineering-education/engineering-education/pull/1699",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1070455479 | Creating Responsive Line and Bar charts with React and Semantic-UI
Article checklist
Pre-submission checks
Please ensure that you have only one open topic suggestion (issue) + in-review article (linked pull request) at a time. This will ensure that we complete the review process in a timely manner from inception to publishing
If you have not already, please go over our Resources Page for more tips and tricks on how to improve your overall technical writing so reviews are swifter, increase the chance of a payout and provide more value to our readers.
For first-time contributors and for more details on our submission guidelines, see our Contributing Guidelines.
Formatting and structure checks
[x] Have you used the correct folder and file structure? - see Contribution Guidelines
[x] Is your article properly formatted in Markdown? - see Github Markdown guide
[x] Have you used our preferred heading structure? - H3 (###) and up
[x] Does your article follow any of the suggested structure formats? - see example formats
[x] Is your article over 750 words? - Tool: Wordcounter.net
[x] Is the article you are submitting an in-depth and unique article? Does it go beyond what is in the official docs and what is covered in other blog sites. See these articles as examples.
[x] Does your article provide enough value and detail about your topic? - Articles should be clear, accurate, and fully explained.
[x] Can your article be understood by beginners? - Assume the audience is smart but has no prior exposure to the common terminology in your article.
[x] Have you included a hero image that is 600x400 pixels and have the copyright to use it? - Tools: Pixlr Image Editor for resizing and Unsplash for Creative Commons images
Grammar & plagiarism checks
[x] Have you spellchecked and grammar checked your article? - Tools: Write&Improve - ProWritingAid - Grammarly
[x] Please place your article through a 3rd party plagiarism checker. We suggest using Quetext, this tool is free and has a daily limit. We typically accept articles with 10% or less.
[x] Have you checked your article for readability? - Tool: Hemmingway
[x] Have you added sources for quotes and images that aren't yours?
Technical checks
[x] Are your code snippets properly formatted for syntax highlighting - see Syntax guide
[x] Have you checked your code runs correctly and you've highlighted all necessary dependencies for installation?
[x] Are the software programs and packages you're highlighting in your article up to date, using current versions and not deprecated?
Topic suggestion this closes
Remove the backticks and add the issue number below to link and close your Topic Suggestion (issue) when your article has been published (PR has been merged). See this video for more details.
This closes #4724
Finally, delete the article checklist notes in blockquotes and submit your PR. We look forward to reviewing your article.
hello
@lalith1403, @jerimkaura
I have made the required changes to the previous PR #4944. please proceed with the final review thank you
cc.
@hectorkambow
@jerimkaura, @lalith1403 I'm yet to receive any response from you. #4944 was one commit away from being merged.
Thank you @srishilesh
@jerimkaura is this article ready for a final review?
Take in mind that (generally) any duplicate article should NOT require any additional revisions - & should be fully ready to publish.
If any more (major) changes, revisions, or edits are required we will be closing the PR indefinitely.
cc @FredBenson
Thank you @srishilesh
@jerimkaura is this article ready for a final review?
Take in mind that (generally) any duplicate article should NOT require any additional revisions - & should be fully ready to publish.
If any more (major) changes, revisions, or edits are required we will be closing the PR indefinitely.
cc @FredBenson
Yes the article is ready for final review.
@FredBenson Please reach out via Slack DM 👍
| gharchive/pull-request | 2021-12-03T10:43:48 | 2025-04-01T06:45:45.554156 | {
"authors": [
"FredBenson",
"hectorkambow",
"jerimkaura"
],
"repo": "section-engineering-education/engineering-education",
"url": "https://github.com/section-engineering-education/engineering-education/pull/5366",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1394806607 | Update readme with About + links to the main technologies used.
Make it more searchable.
Done
| gharchive/issue | 2022-10-03T14:09:54 | 2025-04-01T06:45:45.607004 | {
"authors": [
"UberMeatShield"
],
"repo": "secureworks/azure_auth",
"url": "https://github.com/secureworks/azure_auth/issues/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
211930910 | How to make it work with Fish?
I've tried eval (dircolors ~/.dir_colors), replacing backsticks with parentheses but it didn't help. Fish outputs error.
I've also tried to manually set and export LS_COLORS environment variable, but also without results.
The string is incredibly long.
Please help with fish shell, the standard command for exporting variable is set -x, and there's no equal (=) character when setting the value (set var value is enough) unlike var=value;export $var of zsh and bash.
I could finally solve it.
set -x LS_COLORS "no=00;38;5;244:rs=0:di=00;38;5;33:ln=00;38;5;37:mh=00:pi=48;5;230;38;5;136;01:so=48;5;230;38;5;136;01:do=48;5;230;38;5;136;01:bd=48;5;230;38;5;244;01:cd=48;5;230;38;5;244;01:or=48;5;235;38;5;160:su=48;5;160;38;5;230:sg=48;5;136;38;5;230:ca=30;41:tw=48;5;64;38;5;230:ow=48;5;235;38;5;33:st=48;5;33;38;5;230:ex=00;38;5;64:*.tar=00;38;5;61:*.tgz=00;38;5;61:*.arj=00;38;5;61:*.taz=00;38;5;61:*.lzh=00;38;5;61:*.lzma=00;38;5;61:*.tlz=00;38;5;61:*.txz=00;38;5;61:*.zip=00;38;5;61:*.z=00;38;5;61:*.Z=00;38;5;61:*.dz=00;38;5;61:*.gz=00;38;5;61:*.lz=00;38;5;61:*.xz=00;38;5;61:*.bz2=00;38;5;61:*.bz=00;38;5;61:*.tbz=00;38;5;61:*.tbz2=00;38;5;61:*.tz=00;38;5;61:*.deb=00;38;5;61:*.rpm=00;38;5;61:*.jar=00;38;5;61:*.rar=00;38;5;61:*.ace=00;38;5;61:*.zoo=00;38;5;61:*.cpio=00;38;5;61:*.7z=00;38;5;61:*.rz=00;38;5;61:*.apk=00;38;5;61:*.gem=00;38;5;61:*.jpg=00;38;5;136:*.JPG=00;38;5;136:*.jpeg=00;38;5;136:*.gif=00;38;5;136:*.bmp=00;38;5;136:*.pbm=00;38;5;136:*.pgm=00;38;5;136:*.ppm=00;38;5;136:*.tga=00;38;5;136:*.xbm=00;38;5;136:*.xpm=00;38;5;136:*.tif=00;38;5;136:*.tiff=00;38;5;136:*.png=00;38;5;136:*.PNG=00;38;5;136:*.svg=00;38;5;136:*.svgz=00;38;5;136:*.mng=00;38;5;136:*.pcx=00;38;5;136:*.dl=00;38;5;136:*.xcf=00;38;5;136:*.xwd=00;38;5;136:*.yuv=00;38;5;136:*.cgm=00;38;5;136:*.emf=00;38;5;136:*.eps=00;38;5;136:*.CR2=00;38;5;136:*.ico=00;38;5;136:*.tex=00;38;5;245:*.rdf=00;38;5;245:*.owl=00;38;5;245:*.n3=00;38;5;245:*.ttl=00;38;5;245:*.nt=00;38;5;245:*.torrent=00;38;5;245:*.xml=00;38;5;245:*Makefile=00;38;5;245:*Rakefile=00;38;5;245:*Dockerfile=00;38;5;245:*build.xml=00;38;5;245:*rc=00;38;5;245:*1=00;38;5;245:*.nfo=00;38;5;245:*README=00;38;5;245:*README.txt=00;38;5;245:*readme.txt=00;38;5;245:*.md=00;38;5;245:*README.markdown=00;38;5;245:*.ini=00;38;5;245:*.yml=00;38;5;245:*.cfg=00;38;5;245:*.conf=00;38;5;245:*.h=00;38;5;245:*.hpp=00;38;5;245:*.c=00;38;5;245:*.cpp=00;38;5;245:*.cxx=00;38;5;245:*.cc=00;38;5;245:*.objc=00;38;5;245:*.sqlite=00;38;5;245:*.go=00;38;5;245:*.sql=00;38;5;245:*.csv=00;38;5;245:*.log=00;38;5;240:*.bak=00;38;5;240:*.aux=00;38;5;240:*.lof=00;38;5;240:*.lol=00;38;5;240:*.lot=00;38;5;240:*.out=00;38;5;240:*.toc=00;38;5;240:*.bbl=00;38;5;240:*.blg=00;38;5;240:*~=00;38;5;240:*#=00;38;5;240:*.part=00;38;5;240:*.incomplete=00;38;5;240:*.swp=00;38;5;240:*.tmp=00;38;5;240:*.temp=00;38;5;240:*.o=00;38;5;240:*.pyc=00;38;5;240:*.class=00;38;5;240:*.cache=00;38;5;240:*.aac=00;38;5;166:*.au=00;38;5;166:*.flac=00;38;5;166:*.mid=00;38;5;166:*.midi=00;38;5;166:*.mka=00;38;5;166:*.mp3=00;38;5;166:*.mpc=00;38;5;166:*.ogg=00;38;5;166:*.opus=00;38;5;166:*.ra=00;38;5;166:*.wav=00;38;5;166:*.m4a=00;38;5;166:*.axa=00;38;5;166:*.oga=00;38;5;166:*.spx=00;38;5;166:*.xspf=00;38;5;166:*.mov=00;38;5;166:*.MOV=00;38;5;166:*.mpg=00;38;5;166:*.mpeg=00;38;5;166:*.m2v=00;38;5;166:*.mkv=00;38;5;166:*.ogm=00;38;5;166:*.mp4=00;38;5;166:*.m4v=00;38;5;166:*.mp4v=00;38;5;166:*.vob=00;38;5;166:*.qt=00;38;5;166:*.nuv=00;38;5;166:*.wmv=00;38;5;166:*.asf=00;38;5;166:*.rm=00;38;5;166:*.rmvb=00;38;5;166:*.flc=00;38;5;166:*.avi=00;38;5;166:*.fli=00;38;5;166:*.flv=00;38;5;166:*.gl=00;38;5;166:*.m2ts=00;38;5;166:*.divx=00;38;5;166:*.webm=00;38;5;166:*.axv=00;38;5;166:*.anx=00;38;5;166:*.ogv=00;38;5;166:*.ogx=00;38;5;166:"
This long string should be put inside ~/.config/config.fish.
Just to make sure:
dircolors command, based on dircolors.256dark, generate value for LS_COLORS environment variable which contains color codes for ls command.
Thanks for the great colorscheme, the issue is solved.
This works as of the current version of FISH:
eval (dircolors -c LS_COLORS)
In fish 3.1.2, I neede to use this:
eval (dircolors -c ~/.dircolors)
| gharchive/issue | 2017-03-05T05:11:13 | 2025-04-01T06:45:45.641178 | {
"authors": [
"Chinggis6",
"blinry",
"duganchen"
],
"repo": "seebi/dircolors-solarized",
"url": "https://github.com/seebi/dircolors-solarized/issues/76",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2209457030 | Move table name and description creation into project-id-metadata-create
Right now the table names and descriptions are created in the project-id-view.html as a dialog box. I think it makes more sense to have that creation in the project id metadata create. It would simplify things, like not needing to pass around the table metadata object and it would make it so the create URL is more consistent with our naming scheme. Right now, the URL is project id metadata id create which feels a bit inconsistent, since in order to have an id, we need to create it first, but in this case we have an id as we're creating it...
This is on hold until the team has agreed on the design, see issue #397
We decided to start the web app from scratch after the major design/architecture decisions and changes we made. No longer relevant.
| gharchive/issue | 2024-03-26T22:29:54 | 2025-04-01T06:45:45.646284 | {
"authors": [
"lwjohnst86",
"signekb"
],
"repo": "seedcase-project/seedcase-sprout",
"url": "https://github.com/seedcase-project/seedcase-sprout/issues/341",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
138823853 | PhoneNumberOfflineGeocoder ?
Hello,
i'm trying to get the geo info for a number but cant seen to find the PhoneMetadataCollection function/values ??
i even searched the source but not luck :fearful:
Thanks
Unfortunately, PhoneNumberOfflineGeocoder has not been ported to javascript so don't expect to see this functionality on this wrapper package any time soon. You might want to submit an issue on the official repository.
:disappointed: ... will just write my own geo location function i guess
Thanks
You could submit a PR with the port to their repository :)
@Ahmed-90 did you find/write anything?
I can't even seem to find an NPM package that maps USA numbers to the state.
I scraped wikipedia for North America ... I saved the code here:
https://gist.github.com/avimar/bcc7a8159d214c51b5496be5f5e14362
I asked for it here, apparently it's a leet request... https://github.com/googlei18n/libphonenumber/issues/1337
@avimar yeah.. well kindof.. anyway you can find the geo data in the resources https://github.com/googlei18n/libphonenumber/tree/master/resources/geocoding/en
sorted by the geo code also the nanpa has public listings of the codes -> https://www.nationalnanpa.com/reports/area_code_relief_planning.html
@avimar i just used my own lookup without the lib so .. simply converted the data to json object (keyBy areacode) and just did a lookup when i can't identify the number
not the cleanest way :P but worked fine
Is there any update on this??
Can i use geo-location and timezone mapper facility with Javascript/Node?? If so please reply...
@sahilkanani1812, see original comment https://github.com/seegno/google-libphonenumber/issues/46#issuecomment-192973464. It's still valid.
| gharchive/issue | 2016-03-06T19:03:42 | 2025-04-01T06:45:45.652970 | {
"authors": [
"Ahmed-90",
"avimar",
"ruimarinho",
"sahilkanani1812"
],
"repo": "seegno/google-libphonenumber",
"url": "https://github.com/seegno/google-libphonenumber/issues/46",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
334315951 | feat(disableAsideGap): Adding new props for Aside
** Please provide as much detail as possible, but feel free to remove irrelevant sections **
In the split mode it do not need the space in between. At the same time doing come clean up in showmore component as well
BREAKING CHANGE:
adding a new props
RFC URL:
none
MIGRATION GUIDE:
After:
<AsidedLayout
disableGapInBetween
renderAside={() => {...}}
size="30%"
>
EXAMPLE USAGE:
<AsidedLayout
disableGapInBetween
renderAside={() => {...}}
size="30%"
>
<Card>
<Section>
<Text shouting>
Main Content
</Text>
<Text>
This card is provided as children.
</Text>
</Section>
</Card>
<Card>
<Section>
<Text shouting>
Another Card
</Text>
<Text>
Here's another card for good measure.
</Text>
</Section>
</Card>
</AsidedLayout>
Try not to merge with only one approving review
| gharchive/pull-request | 2018-06-21T02:34:55 | 2025-04-01T06:45:45.657687 | {
"authors": [
"ChanJit",
"PepeFranco"
],
"repo": "seekinternational/seek-asia-style-guide",
"url": "https://github.com/seekinternational/seek-asia-style-guide/pull/139",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2300093385 | 线条插槽不能使用-- 并且使用了ovUseLineSlot属性
relation-graph@2.1.42
vue@2.6.10
使用 线条插槽 组件里复制的dome
<template #line="{ line, link, lineIndex }">
线条不生效 报错 Cannot read properties of undefined (reading 'getArrow')
说明 relationGraph 为undefined
relation-graph@2.1.x中在自定义节点、线条插槽中通过依赖注入获取graphInstance的方式发生了变化:
以前是:通过依赖的对象graph获取:graph.instance
现在是:通过依赖的方法graphInstance来获取:graphInstance()
你可以将你的代码以下内容按照下图修改一下就可以了:
相关的示例:
https://www.relation-graph.com/#/demo/vue2?id=adv-line-slot2
| gharchive/issue | 2024-05-16T11:02:45 | 2025-04-01T06:45:45.661071 | {
"authors": [
"ZhouJing0",
"seeksdream"
],
"repo": "seeksdream/relation-graph",
"url": "https://github.com/seeksdream/relation-graph/issues/309",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2704448650 | Dependency lib can upgrade: System.Text.Json
Here is the detail.
https://github.com/segmentio/Serialization.NET/issues/16
hi @arlyxiao, thanks for reporting this. we will get it updated. in the meantime, you should be able to override the dependency version by directly referencing it. see here
| gharchive/issue | 2024-11-29T09:08:19 | 2025-04-01T06:45:45.666881 | {
"authors": [
"arlyxiao",
"wenxi-zeng"
],
"repo": "segmentio/Analytics-CSharp",
"url": "https://github.com/segmentio/Analytics-CSharp/issues/118",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
541753549 | Using credential_process when multiple IAM roles available
I'm trying to get aws-okta configured as the credential_process for a particular profile. The Okta application configured at the SAML URL has multiple AWS IAM roles assigned, requiring the particular role to be chosen interactively. This interactive prompt causes aws-okta to hang indefinitely. I attempted to use the role_arn option to specify the desired role, but this seems to cause the CLI to bypass the credential_process option entirely, throwing an error about partial credentials.
Is it possible to configure profiles to assume specific roles directly without an interactive prompt?
been struggling with that as well, this is due to AWS behaviour, when role_arn is specified AWS sdk will either look for source_profile or direct credentials for this profile in credentials file, it does not support credential_process
on main account i use credential_process and am not specifing role_arn and for roles i create profiles with source_profile pointing to main profile. This will AssumeRole in chain and work fine except the assumed roles will hit AWS hard limit for 1hr sts token when assumerolechain, which will be ok as okta will extend when needed.
BUT if some shit dont understand credential_process in assumedchain roles like terrafrom s3 backend bucket then you need aws-okta env to inject session token to env and that one expires within 1hr no option for change and have to remove it manually
I found a relevant issue in the botocore library: https://github.com/boto/botocore/issues/1329. It looks like there's still some work to be done in the AWS SDKs to make this work smoothly.
| gharchive/issue | 2019-12-23T13:23:48 | 2025-04-01T06:45:45.686588 | {
"authors": [
"bgshacklett",
"gacopl"
],
"repo": "segmentio/aws-okta",
"url": "https://github.com/segmentio/aws-okta/issues/266",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
129546671 | adds validation for context.app.name
This will return an error when context.app.name isn't passed through a server-side .screen() call. GA requires that an : Application Name be passed through .screen() calls.
@sperand-io :eyes:
@f2prateek updated tests
| gharchive/pull-request | 2016-01-28T19:06:53 | 2025-04-01T06:45:45.701685 | {
"authors": [
"ladanazita"
],
"repo": "segmentio/integration-google-analytics",
"url": "https://github.com/segmentio/integration-google-analytics/pull/28",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
160463778 | Allow for async processing
This builds on #5 and adds options.async to requests, which internally uses null for the request id. This just means that the JSON-RPC server should not block the client while it waits for an answer, it will simply complete the request asynchronously.
This could be useful where we don't necessarily care about waiting for the response, such as during sign up. Subsequent gets will just silently fail until the background work is done. This is just an idea, we'll have to see if it works in practice.
/cc @stephenmathieson @stevenmiller888
Good idea!
actually can we do this on the per-request basis? we generally use the same rpc client to make requests we do care about AND requests we don't
@stephenmathieson: yeah, this is per-request (you add { async: true } after your params)
@stephenmathieson: yeah, this is per-request (you add { async: true } after your params, this is not part of the constructor)
haha gotcha whoops
test added!
lgtm
| gharchive/pull-request | 2016-06-15T16:17:27 | 2025-04-01T06:45:45.704719 | {
"authors": [
"dominicbarnes",
"stephenmathieson",
"stevenmiller888"
],
"repo": "segmentio/jsonrpc2.js",
"url": "https://github.com/segmentio/jsonrpc2.js/pull/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
119697909 | Add json options docs
Highlight which options metalsmith.json recognises. Given these are detailed in the API section, I've just referred on to it.
Fixes #106
:+1:
I guess I have the ability to merge this myself now. So... poof! :smile:
| gharchive/pull-request | 2015-12-01T11:31:08 | 2025-04-01T06:45:45.706042 | {
"authors": [
"Ajedi32",
"iwootten"
],
"repo": "segmentio/metalsmith",
"url": "https://github.com/segmentio/metalsmith/pull/189",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
542121406 | support pidfd
related to #295
Job object maintain pidfds instead of pids for race-free process management
use epoll for process waiting
need linux kernel 5.4 or later
no longer need it, since now call waitpid(-1)
| gharchive/issue | 2019-12-24T12:57:11 | 2025-04-01T06:45:45.724387 | {
"authors": [
"sekiguchi-nagisa"
],
"repo": "sekiguchi-nagisa/ydsh",
"url": "https://github.com/sekiguchi-nagisa/ydsh/issues/392",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
80949710 | how do i assign data attribute of select option on ajax request.
Hello there,
i have html select and i am binding data dynamically, now i have scenario i want to set three value of each select option so i want to assign data attribute of select option using ajaxrequest like:
HELLO WORLD
$(".select2").select2({
ajax: {
url: '@Url.Action("GetProjectList","Test")',
dataType: 'json',
delay: 250,
data: function (params) {
return {
search: params.term, // search term
page: params.page
};
},
processResults: function (data, page) {
// response from server
// how can i create the option data attribute
return {
results: data
};
},
cache: true
},
escapeMarkup: function (markup) { return markup; }, // let our custom formatter work
minimumInputLength: 1,
//templateResult: formatRepo, // omitted for brevity, see the source of this page
//templateSelection: formatRepoSelection // omitted for brevity, see the source of this page
});
This appears to be more of a usage question than a bug report or common feature request. As we continue to keep the Select2 issue tracker geared towards these two areas, bug reports and common feature requests, we are redirecting usage questions to other communities such as the mailing list, IRC, or Stack Overflow.
| gharchive/issue | 2015-05-26T12:53:42 | 2025-04-01T06:45:45.730607 | {
"authors": [
"Mubusher87",
"kevin-brown"
],
"repo": "select2/select2",
"url": "https://github.com/select2/select2/issues/3408",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
103528380 | Select2 Onchange event triggered while applying data manually(how to suppress this)
Previously I am using select2 3.3.2, in that while applying data through this syntax,
jQuery('#elementId').select2('data', {"id":1,"text": "some value"}, true);
It is not triggering that elements onchange event.
I upgraded to 3.5.1, now its triggering the onchange event.
If the change is intentional, is there any way to suppress that onchange event by passing a parameter to select2?
Thanks in advance.
Just for reference, 3.3.2 was released on May 9, 2013 and is pretty far out of the support cycle. 3.5.1 also isn't the latest release in the 3.5.x line, I'd recommend 3.5.2.
The select2('data') method supports a triggerChange argument as the third parameter, and you're passing in true so it does trigger the change event.
Try passing in false.
| gharchive/issue | 2015-08-27T14:53:42 | 2025-04-01T06:45:45.733594 | {
"authors": [
"jaggu3017",
"kevin-brown"
],
"repo": "select2/select2",
"url": "https://github.com/select2/select2/issues/3714",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
173218182 | cannot display the placeholder with the numbers
I use this sccript:
<select class = "form-control select2" placeholder data = "1" style = "width: 100%;">
I have a problem displaying a placeholder number.
give me a solution. thank you
Do you mean when it initial by the following snippet , and it's not working ?
$("#test").select2({
placeholder:134
});
if it's the situation you mention ,
you can initial it by add double quote like the following
$("#test").select2({
placeholder:"134"
});
its work. Thanks.
| gharchive/issue | 2016-08-25T14:26:32 | 2025-04-01T06:45:45.735807 | {
"authors": [
"fannysispriadi",
"holmes2136"
],
"repo": "select2/select2",
"url": "https://github.com/select2/select2/issues/4551",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
206498173 | Tags case sensitivity
Well the problem is again with tags and case sensitivity. Select2 allows inserting tags with different cases, where (at least in my opinion) it should not allow it. If a list has a "purple" tag the user should not be able to add "PURPLE", "PuRPLe" and so on. Granted I now have to manually choose to input the duplicate tag by pressing UP on the keyboard, which makes it a little bit harder but not impossible, which is what I'd like to achieve.
This is the old bug report from 2015 july: https://github.com/select2/select2/issues/3566
In any case it's still possible to input duplicate tags:
Prerequisites
[ ] I have searched for similar issues in both open and closed tickets and cannot find a duplicate
[x] The issue still exists against the latest master branch of Select2
[x] This is not a usage question (Those should be directed to the community)
[x] I have attempted to find the simplest possible steps to reproduce the issue
[ ] I have included a failing test as a pull request (Optional)
Steps to reproduce the issue
Go to https://select2.github.io/examples.html#tags
Input PURPLE
Press UP and ENTER
Input PuRPLe
Press UP and ENTER
Etc.
Expected behavior and actual behavior
When I follow those steps, I see that the duplicate tags are being added to the list.
I was expecting Select2 to limit me from inputing duplicate tags.
Environment
Browsers
[x] Google Chrome
[x] Mozilla Firefox
[x] Internet Explorer
Operating System
[x] Windows
[ ] Mac OS X
[ ] Linux
[ ] Mobile
Libraries
jQuery version: The one on the examples page
Select2 version: The one on the examples page
Isolating the problem
[x] This bug happens on the examples page
[x] The bug happens consistently across all tested browsers
[x] This bug happens when using Select2 without other pluigns
[ ] I can reproduce this bug in a jsbin
It looks like this issue was actually fixed in https://github.com/select2/select2/commit/cb9a90457867ffb14c7b1550bb67e872e0a5c2dd, but it just hasn't made it into a release yet. I'm trying to get a new release out ASAP (and tighten the overall release cycle), so I would say that this issue is resolved.
| gharchive/issue | 2017-02-09T13:06:44 | 2025-04-01T06:45:45.745597 | {
"authors": [
"alexweissman",
"loshose"
],
"repo": "select2/select2",
"url": "https://github.com/select2/select2/issues/4784",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
900027640 | Word wrap / Text Wrap within a Hidden Input Multiple Dropdown
I am currently trying to word-wrap / text-wrap within the options (or the select) portion of the drop-down menu that has the hidden input with multiple selections with a size of 4, this way the text can be fully read and not be cut off. And instead of being cut off, the text goes down to the second or third line, aka a wrap Trying to work on the CSS and the Select2 jquery but unable to solve this issue.
Below is my HTML.
<html>
<form>
<select id="select2" class="select2 narrow wrap" multiple size="4">
<option value="01">01 Really long text description that normarly</option>
<option value="02">02 Really long text description that normarly</option>
<option value="03">03 Really long text description</option>
<option value="04">04 Really long text description that normarly</option>
<option value="05">05 Really long text description</option>
<option value="06">06 Really long text description that normarly</option>
</select>
</form>
</html>
Here is my CSS:
.select2.narrow {
width: 200px;
}
.wrap.select2-selection--single {
height: 100%;
}
.select2-container .wrap.select2-selection--multiple .select2-selection__rendered {
word-wrap: break-word !important;
text-overflow: inherit !important;
white-space: normal !important;
}
.wrap {
word-wrap: break-word;
text-overflow: inherit;
white-space: normal;
}
And here is my javascript:
var $select2 = $('.select2').select2({
containerCssClass: "wrap"
})
var $select2 = $('#select2')#select2({
minimumResultsForSearch: -4
});
Any help to see if this can be done is appreciated.
This appears to be more of a usage question than a bug report or common feature request. As we continue to keep the Select2 issue tracker geared towards these two areas, bug reports and common feature requests, we are redirecting usage questions to other communities such as the mailing list, IRC, or Stack Overflow.
| gharchive/issue | 2021-05-24T21:26:38 | 2025-04-01T06:45:45.749624 | {
"authors": [
"camanicnack",
"kevin-brown"
],
"repo": "select2/select2",
"url": "https://github.com/select2/select2/issues/6038",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1164887002 | On open second time modal - dropdown list duplicates
Hi all, i run into problem, on second time when i open my modal popup, select2 dropdown gets duplicated..
1st, 3rd, 4th ..etc.. works normally... Don't blame for Russian flag.. :D
Tried using 4.0.7 / 4.0.12 / 4.0.13 / 4.1.0rc
Using with transition, keepalive, router-view, tried without its, directly anyway got this problem... ;(
Browser: Brave Version 1.36.105 Chromium: 99.0.4844.45 (Official Build) unknown (64-bit) / Opera
OS: Linux Manjaro XFCE
"jquery": "^3.6.0",
"vue": "^3.0.0",
temporary fix
.select2:nth-child(even) {
display: none;
}
This appears to be more of a usage question than a bug report or common feature request. As we continue to keep the Select2 issue tracker geared towards these two areas, bug reports and common feature requests, we are redirecting usage questions to other communities such as the mailing list, IRC, or Stack Overflow.
If it's appears more of a usage question, than you know the answer. If you know the answer and you do not tell me, this appears to be be a you dont give f*** about that question... Sorry <3 Thanks!
Neither of those inferences is logically valid.
i know i just try to understand, why i cant find that problem "or question or whatever this kind of entity is" at www...
This has been classified as a usage question instead of a bug because it appears to be caused by the specific usage that you describe. Please take this usage question to one of the recommended forums if you're looking to get your specific configuration working. Thanks!
You didn't follow the required issue template for bugs and feature requests, so that's also a huge sign that this is actually a usage question.
| gharchive/issue | 2022-03-10T08:03:23 | 2025-04-01T06:45:45.756794 | {
"authors": [
"iBrabus",
"jonahgreenthal",
"kevin-brown"
],
"repo": "select2/select2",
"url": "https://github.com/select2/select2/issues/6148",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
292498728 | Actualize readme
fix typo "alrady" => elaborate "around_filter/around_action" instead
b/c it actually has been battle tested against 5.0
Fixed it in master, thanks for reporting!
| gharchive/pull-request | 2018-01-29T17:44:46 | 2025-04-01T06:45:45.807626 | {
"authors": [
"AlexKVal",
"semaperepelitsa"
],
"repo": "semaperepelitsa/subdomain_locale",
"url": "https://github.com/semaperepelitsa/subdomain_locale/pull/10",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2282552283 | Playbook error
I have a playbook with this list of tasks:
`
- name: Fetching available collections
command: ansible-galaxy collection list
register: result
- name: Available collections
debug:
var: result.stdout_lines
- name: Fetching available roles
command: ansible-galaxy list
register: result
- name: Available roles
debug:
var: result.stdout_lines
`
and I get this error:
TASK [Fetching available collections] ******************************************
fatal: [localhost]: FAILED! => {
"changed":true,
"cmd":[
"ansible-galaxy",
"collection",
"list"
],
"delta":"0:00:02.423948",
"end":"2024-04-30 13:26:21.383339",
"msg":"non-zero return code",
"rc":5,
"start":"2024-04-30 13:26:18.959391",
"stderr":"\u001b[0;31mERROR! - None of the provided paths were usable. Please specify a valid path with --collections-path\u001b[0m",
"stderr_lines":[
"\u001b[0;31mERROR! - None of the provided paths were usable. Please specify a valid path with --collections-path\u001b[0m"
],
"stdout":"usage: ansible-galaxy [-h] [--version] [-v] TYPE ...\n\nPerform various Role and Collection related operations.\n\npositional arguments:\n TYPE\n collection Manage an Ansible Galaxy collection.\n role Manage an Ansible Galaxy role.\n\noptions:\n --version show program's version number, config file location,\n configured module search path, module location, executable\n location and exit\n -h, --help show this help message and exit\n -v, --verbose Causes Ansible to print more debug messages. Adding multiple\n -v will increase the verbosity, the builtin plugins currently\n evaluate up to -vvvvvv. A reasonable level to start is -vvv,\n connection debugging might require -vvvv.",
"stdout_lines":[
"usage: ansible-galaxy [-h] [--version] [-v] TYPE ...",
"",
"Perform various Role and Collection related operations.",
"",
"positional arguments:",
" TYPE",
" collection Manage an Ansible Galaxy collection.",
" role Manage an Ansible Galaxy role.",
"",
"options:",
" --version show program's version number, config file location,",
" configured module search path, module location, executable",
" location and exit",
" -h, --help show this help message and exit",
" -v, --verbose Causes Ansible to print more debug messages. Adding multiple",
" -v will increase the verbosity, the builtin plugins currently",
" evaluate up to -vvvvvv. A reasonable level to start is -vvv,",
" connection debugging might require -vvvv."
]
}
I have a ansible.cfg that is specifying the COLLECTIONS_PATH and DEFAULT_ROLES_PATH to custom paths.
This tasks are working fine on my local machine when I run the playbook but in in Ansible Semaphore UI.
I guess this could be related to #1984 Not possible to set working directory for a template task
In this version (v2.10.2-cc91501-1718257299) it seems to be working:
Now the problem is that the following is still not wrking:
name: Fetching available roles
command: ansible-galaxy list
register: result
I have read som documentation to specify the paths for changing paths for collections and roles and Semaphore UI does not seem to respect them if I add them as CLI arguments for a task.
[
"--connection=local",
"-e",
"ANSIBLE_CONFIG=folder/ansible/ansible.cfg"
]
or
[
"--connection=local",
"-e",
"ANSIBLE_COLLECTIONS_PATH=folder/ansible/.collections"
]
| gharchive/issue | 2024-05-07T07:53:34 | 2025-04-01T06:45:45.814596 | {
"authors": [
"mikaelparkefelt"
],
"repo": "semaphoreui/semaphore",
"url": "https://github.com/semaphoreui/semaphore/issues/1998",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
363521472 | SC-2853 LA config improvements
By default, US region is set and it can be changed to EU on container start to update receivers endpoints. This PR solves the issue when custom LOG_URL is set and invalidates region settings. Also fixes logging message with receivers that are in use.
Docs need to be changed before merging this PR.
Docs need to be changed before merging this PR.
Is there a PR for that @komljen ?
Not yet, they will be added to the same PR.
| gharchive/pull-request | 2018-09-25T11:00:16 | 2025-04-01T06:45:45.816513 | {
"authors": [
"komljen",
"otisg"
],
"repo": "sematext/logagent-js",
"url": "https://github.com/sematext/logagent-js/pull/139",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
701728716 | GraphQL Get group filter missing
The GraphQL query with group filter for entity merging is missing. Queries like this are impossible with the client now:
{
Get {
Things {
Publication(
group:{
type: merge,
force:0.05
}
) {
name
}
}
}
}
Should be fixed/implemented in v1.0.0-rc4. Keeping this issues open until I have feedback that it works as desired.
Confirmed working by @laura-ham. Closing.
| gharchive/issue | 2020-09-15T08:34:25 | 2025-04-01T06:45:45.818180 | {
"authors": [
"etiennedi",
"laura-ham"
],
"repo": "semi-technologies/weaviate-javascript-client",
"url": "https://github.com/semi-technologies/weaviate-javascript-client/issues/7",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2084756789 | Test stream blocks
I've added a single test for the stream_blocks function. Since this function is async, I've had to use tokio_test::block_on to test it, added it as a dev dependency.
One thing I noticed is that stream_blocks just inputs everything into the writer but there isn't a way to read blocks from this writer later. extract_blocks could be adapted to do this for the stream of blocks, however. Should we be concerned about this or this unnecessary?
I've added a single test for the stream_blocks function. Since this function is async, I've had to use tokio_test::block_on to test it, added it as a dev dependency. One thing I noticed is that stream_blocks just inputs everything into the writer but there isn't a way to read blocks from this writer later. extract_blocks could be adapted to do this for the stream of blocks, however. Should we be concerned about this or this unnecessary?
Regarding this, its a nice to have but not urgent, since most components are tested individually
| gharchive/pull-request | 2024-01-16T19:58:39 | 2025-04-01T06:45:45.820484 | {
"authors": [
"pedrohba1",
"srosati"
],
"repo": "semiotic-ai/flat-files-decoder",
"url": "https://github.com/semiotic-ai/flat-files-decoder/pull/6",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
304192043 | Failed to execute 'attachShader' on 'WebGLRenderingContext': parameter 1 is not of type 'WebGLProgram'
After rendering the background for a while, the example throws this error 1/second and stops rendering:
Uncaught TypeError: Failed to execute 'attachShader' on 'WebGLRenderingContext': parameter 1 is not of type 'WebGLProgram'.
at WebGLSpriteRenderer.createProgram (VM253 three.js:4648)
at WebGLSpriteRenderer.init (VM253 three.js:4451)
at WebGLSpriteRenderer.render (VM253 three.js:4488)
at WebGLRenderer.render (VM253 three.js:13947)
at THREE.RenderPass.render (VM513 RenderPass.js:51)
at THREE.EffectComposer.render (VM512 EffectComposer.js:88)
at render (webgl_materials_cubemap_dynamic.html:981)
at animate (webgl_materials_cubemap_dynamic.html:866)
https://htmlpreview.github.io/?https://github.com/semleti/three.ts/blob/master/examples/webgl_materials_cubemap_dynamic.html
same for sprites:
https://htmlpreview.github.io/?https://github.com/semleti/three.ts/blob/master/examples/webgl_sprites.html
same for misc ubiquity test2:
https://htmlpreview.github.io/?https://github.com/semleti/three.ts/blob/master/examples/misc_ubiquity_test2.html
Solved.
Added a 'this' instead of using the local let in WebGLSpiteRenderer.createProgram .
| gharchive/issue | 2018-03-11T20:42:48 | 2025-04-01T06:45:45.823866 | {
"authors": [
"semleti"
],
"repo": "semleti/three.ts",
"url": "https://github.com/semleti/three.ts/issues/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1955020470 | Error handling
Closes #5
Hey @semolex, any thoughts here?
Sorry for delay. Let's merge it
| gharchive/pull-request | 2023-10-20T21:14:25 | 2025-04-01T06:45:45.824933 | {
"authors": [
"krasnoukhov",
"semolex"
],
"repo": "semolex/novaposhta-python-client",
"url": "https://github.com/semolex/novaposhta-python-client/pull/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2010883715 | Model on online demo
Hi.
It's not clear to me what model is being used in the online demo, but it appears to be LLaMA 2, correct me if I am wrong.
If this is so, may I suggest considering switching to something like mistral-openorca, zephyr, dolphin2.2.1-mistral, openhermes2.5-mistral, openchat or neural-chat?
Unlike LLaMA (2), which has a proprietary license by Meta, these models are all under an open source license and are based on Mistral, which is under the Apache 2.0 license and some of them may even outperform LLaMA 2.
Hi! The demo is using OpenChat 3.5 by Alignment Lab, which is Mistral based.
Thank you, @kasumi-1, my bad. I will close the issue.
Is this documented anywhere?
The reason I thought it was LLaMA 2 is that it appears as the suggested default model if you open the Ollama settings:
I don't think it is documented, just from launch announcement. I'll add to the faq on heyamica.com - thank you
Thank you a lot!
Please, consider updating Ollama default settings for consistency.
| gharchive/issue | 2023-11-26T04:10:05 | 2025-04-01T06:45:45.828628 | {
"authors": [
"Aspie96",
"kasumi-1"
],
"repo": "semperai/amica",
"url": "https://github.com/semperai/amica/issues/3",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
374747926 | Update readme.md
This change is optional but most important to get updated npm
Description of the change:
Reason for the change:
Link to original source:
Closes #
Hi, I don't have merge permission, it's just 1 line in this commit, so can you please try resolving that ?
Hello @vk-sourcefuse,Thanks again for the PR! It's HACKTOBERFEST! We want to show our appreciation by sending you some special Hacktoberfest swag. If you have not already, could you please fill out this form so we can send it to you? Thanks! Team SendGrid DX
@vk-sourcefuse - I accidentially overwrote your PR when I was fixing the conflict. Thanks for adding this - I made sure you got the hacktoberfest credit with us and with digital ocean!
| gharchive/pull-request | 2018-10-28T12:20:28 | 2025-04-01T06:45:45.833147 | {
"authors": [
"ksigler7",
"thinkingserious",
"vk-sourcefuse"
],
"repo": "sendgrid/docs",
"url": "https://github.com/sendgrid/docs/pull/4553",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
455948736 | Adding link to Pipedream pipeline
Description of the change: This Pipedream pipeline ingests Sendgrid webhook data and lets you run SQL on event-level engagement data so you can ask interesting questions. It's all free, thought it would be useful to other devs.
Reason for the change: Just thought it would be a cool, useful resource for other devs.
Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it.
Hello @dylburger,Thanks again for the PR! We want to show our appreciation by sending you some swag. Could you please fill out this form so we can send it to you? Thanks! Team SendGrid DX
| gharchive/pull-request | 2019-06-13T20:55:40 | 2025-04-01T06:45:45.837188 | {
"authors": [
"SendGridDX",
"dylburger",
"thinkingserious"
],
"repo": "sendgrid/docs",
"url": "https://github.com/sendgrid/docs/pull/5303",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
279418139 | < is not interpreted as less than sign
Hi,
I am trying to add text that contains a less than sign and has special styling, so I am using html converter to apply this and when I convert the less than character to < it doesn't get rendered in the generated file although the greater than sign is rendered the same way by replcing it to > in the HTML text.
Is there a way to apply this
Thanks
@ahmedelmahalawey so after doing some digging the reason we lose HTML entities is that we pass the provided HTML string though Nokogiri::HTML.fragment here: https://github.com/senny/sablon/blob/2d2399e7343618f82a63f2335ddd144023e85dc9/lib/sablon/html/converter.rb#L18 and then through Nokogiri::XML.fragment here: https://github.com/senny/sablon/blob/2d2399e7343618f82a63f2335ddd144023e85dc9/lib/sablon/content.rb#L75
The HTML entities defined initially get converted by Nokogiri during the HTML parsing step and then get lost in the translation to XML. This stack overflow answer might be a possible solution but I'll need to investigate/test more.
I ran into this issue as well but since I was rendering the HTML from Rails partials I figured it was a Rails produced problem and never looked into it.
I think I have a fix in #80, would you mind testing it in your work flow @ahmedelmahalawey ?
Yeah it worked perfectly, thanks.
| gharchive/issue | 2017-12-05T15:31:52 | 2025-04-01T06:45:45.860035 | {
"authors": [
"ahmedelmahalawey",
"stadelmanma"
],
"repo": "senny/sablon",
"url": "https://github.com/senny/sablon/issues/79",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1253175013 | 🛑 Blockly is down
In 4526eb2, Blockly (https://blockly.sensebox.de/ardublockly/?board=sensebox-mcu) was down:
HTTP code: 0
Response time: 0 ms
Resolved: Blockly is back up in 7fe21af.
| gharchive/issue | 2022-05-30T23:05:21 | 2025-04-01T06:45:45.864644 | {
"authors": [
"mpfeil"
],
"repo": "sensebox/upptime",
"url": "https://github.com/sensebox/upptime/issues/95",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.