id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
98973380
About how to security the view with no blocking. What about this: controller val dataFuture: Future[Data] = ??? val subjectWithPermissionsFromCacheOrDBFuture: Future[CaseSubject] = ??? for { data <- data subject <- subjectWithPermissionFromCacheOrDBFuture } yield { Ok(view(data)(SubjectWrap(subject))) } view.scala.html @(data: Data)(wrap: SubjectWrap) @wrap.require("superAdmin") { @data.name } I mean get the subject informations with future in the controller first and render the view with no db connection. Oh, when I review this issue, I think I need to say more specific. model.scala case class Subject(name: String, permissions: List[Permission]) { def require(permission: String)(content: => T) = if(permissions.exists(_.getName == permission)) { Option(content) } else { None } } controller.scala val dataFuture = Future[Data] val subjectFuture = Future[Subject] for { data <- dataFuture subject <- subjectFuture } yield { Ok(views.html.index(data)(subject)) } view.html ... @subject.require("admin"){ @data.name } ... That's an interesting idea. However, if you fetch the Subject (or DynamicResourceHandler, or whatever) in a non-blocking controller call, you can make it immediately accessible via the DeadboltHandler and set the blocking duration to -1.
gharchive/issue
2015-08-04T13:43:27
2025-04-01T06:45:44.047058
{ "authors": [ "djx314", "schaloner" ], "repo": "schaloner/deadbolt-2-scala", "url": "https://github.com/schaloner/deadbolt-2-scala/issues/14", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
102640377
Remove ExplicitUnicodeString and add NonUnicodeString As I explained here, code generatiion for SQL Server should default to Unicode string literals, so I removed ExplicitUnicodeString and added NonUnicodeString instead. This change affected SQL Server CE too, which I think defaults to Unicode, so 'caș' and N'caș' are equivalent there, but there shouldn't be any harm in using N'. Hana was also using ExplicitUnicodeString, so I went ahead and changed it too. @FabioNascimento, is it ok if the Hana generator dafaults to Unicode string literals? I removed the ExplicitUnicodeString class. This means people who were using it will have to change their code. I can add it back and deprecate it if anyone will miss it, but I think migration code is cleaner when you use normal .NET strings. Fixes #636. Hello @vgrigoriu , Sorry for the late answer, but I just see now. I test your changes in HANA db and everything work fine... Fabio Wow, that's a blast from the past 😃 Who is taking care of this repo now? @fubar-coder ? Any chance of merging this in if I solve the conflicts? This seems to be a breaking change for the end user due to the removal of ExplicitUnicodeString. Can you please keep ExplicitUnicodeString as obsolete (with explanation)? Can it cause problems when we switch to unicode by default? All right, I'll put ExplicitUnicodeString back in as obsolete. I explained [here|https://github.com/schambers/fluentmigrator/issues/636#issuecomment-129391177] what I think are the consequences of switching to Unicode. Also, please don't forget to handle ExplicitUnicodeString instances 😉 . @schambers, I guess, that this change is OK for a minor library update when the only user visible change is a deprecated class? Here you are. Thank you
gharchive/pull-request
2015-08-23T17:45:53
2025-04-01T06:45:44.053427
{ "authors": [ "FabioNascimento", "fubar-coder", "vgrigoriu" ], "repo": "schambers/fluentmigrator", "url": "https://github.com/schambers/fluentmigrator/pull/642", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
1404978108
забанен незнаю за что почему меня забанили? моды стоят только с вашего сервера больше не где не играл!!! This is the source code for the chat mod. Bans are determined by the servers that use it, not the mod itself. а как тогда играть если банит?
gharchive/issue
2022-10-11T17:17:04
2025-04-01T06:45:44.055402
{ "authors": [ "dimooon08101997", "schana" ], "repo": "schana/dayz-mod-global-chat", "url": "https://github.com/schana/dayz-mod-global-chat/issues/10", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
384618412
best practice to make use of pageviewcontroller in layout I'd like to implement a tabbar controller positioned at top, so I think of using UIPageViewController, but I did not find the document about this? Is it possible to use the controller? any example? Thanks in advance thanks for clarification, closing this issue
gharchive/issue
2018-11-27T04:33:51
2025-04-01T06:45:44.122863
{ "authors": [ "wpcfan" ], "repo": "schibsted/layout", "url": "https://github.com/schibsted/layout/issues/163", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
132000641
Use websockets for updating client from server (instead of polling) [need help] Basically I think need to keep track of the connections, as in conn, err := wsupgrader.Upgrade(w, r, nil) in websockets.go... However, this is only really needed if people are going to use this like a chatroom - i.e. realtime viewing of the same document.
gharchive/issue
2016-02-07T19:20:54
2025-04-01T06:45:44.312744
{ "authors": [ "schollz" ], "repo": "schollz/AwwKoala", "url": "https://github.com/schollz/AwwKoala/issues/5", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1637966907
Maintenance of installation module Module to maintain Website: https://school-brainhack.github.io/modules/installation/ Code: https://github.com/school-brainhack/school-brainhack.github.io/tree/main/content/en/modules/installation Work to be done Please ensure: [ ] links, resources and exercises reflect the state-of-the-art [ ] videos are of good quality and reflect the materials. If any update / re-write of the module is needed, feel free to suggest alternative material in a comment! If you think major changes are needed, we suggest you open an issue for new modules. Hi, this is Yu-Shiang Su from Taiwan hub. It seems that the installation instructions for Windows users are outdated. Microsoft has updated their instructions to install WSL, and the steps are much easier now. Importantly, Docker Toolbox was deprecated, and students should install Docker Desktop which now have better support for WSL. Dear @yushiangsu thank you for letting us know, that's amazing you caught up this. I wonder would you like to help us with updating the instructions based on new improvements please? All you need to do is to fork the repo and make the changes and PR, then we will review the changes accordingly. Please have a look at the instructions for the module maintenance here and do not hesitate to reach us with any questions you might have. Yes, I glad to help. I will prepare a PR later this week. Closing this since @yushiangsu completed a PR and I have merged Re-opening since we decided to keep these open until modules have been 'piloted' before school starts. It seems that new macOS versions (after Monterey 12.3) dropped support for python 2.x. Details here. This then results in an error when checking installation using the following command from the Installation module: https://github.com/school-brainhack/school-brainhack.github.io/blob/57f93c5f21bd350a39c7cf3e895029f4033dfa81/content/en/modules/installation/index.md?plain=1#L85 Terminal output valosek@macbook-pro:~$ bash <( curl -s https://raw.githubusercontent.com/brainhackorg/school/master/content/en/modules/installation/nds_check_install.sh ) Missing software program: python. Check installation instructions The proper command here, in this case, should be hash python3 instead of hash python: valosek@macbook-pro:~$ hash python hash: no such command: python valosek@macbook-pro:~$ hash python3 Thanks for that @valosekj, please do let us know if you spot any other bugs!
gharchive/issue
2023-03-23T17:25:41
2025-04-01T06:45:44.339490
{ "authors": [ "clarkenj", "complexbrains", "valosekj", "yushiangsu" ], "repo": "school-brainhack/school-brainhack.github.io", "url": "https://github.com/school-brainhack/school-brainhack.github.io/issues/175", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
646942953
the bit unpacking of the raw Bayer data is wrong for V1 and V2 camera Line 179 and following of core.py does the bit unpacking for V1 and V2 camera: data = data.astype(np.uint16) << 2 for byte in range(4): data[:, byte::5] |= ((data[:, 4::5] >> ((byte+1) * 2)) & 0b11) data = np.delete(data, np.s_[4::5], 1) This does an unpacking for the scheme: AAAAAAAA BBBBBBBB CCCCCCCC DDDDDDDD DDCCBBAA But the documentation in https://picamera.readthedocs.io/en/release-1.12/recipes2.html#raw-bayer-data-captures states this: AAAAAAAA BBBBBBBB CCCCCCCC DDDDDDDD AABBCCDD and the code: data = data.astype(np.uint16) << 2 for byte in range(4): data[:, byte::5] |= ((data[:, 4::5] >> ((4 - byte) * 2)) & 0b11) data = np.delete(data, np.s_[4::5], 1) Good catch, not sure how I let that one slip by. Will implement in next release.
gharchive/issue
2020-06-28T15:10:30
2025-04-01T06:45:44.342640
{ "authors": [ "schoolpost", "scriptorron" ], "repo": "schoolpost/PyDNG", "url": "https://github.com/schoolpost/PyDNG/issues/14", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2518058503
Question about examples In some of the examples you just run (eg. here) compute.run(...) and in some you do all the steps manually (eg. here) try compute.task { task in try task { dispatch in try dispatch(....) } } Is there any reason to prefer the second version before the first? They're both equivalent. run is just more convenient at the expensive of less flexibility.. compute.run(A) compute.run(B) compute.run(C) compute.task { task in task { dispatch in dispatch(A) dispatch(B) dispatch(C) } } The first example will wait for the first shader to complete before running the second shader… and so on. The second example submits them all to Metal in the same pass and allows metal to optimise if needed. It only waits for completion once. I am thinking of deprecating run possibly. Thanks. I see... so if I want to pass a texture for example through multiple shaders without hitting the CPU in between I would use the second one. I think the variant of run that is used in the example in the readme is actually deprecated.
gharchive/issue
2024-09-10T22:26:09
2025-04-01T06:45:44.355103
{ "authors": [ "potmo", "schwa" ], "repo": "schwa/Compute", "url": "https://github.com/schwa/Compute/issues/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
222827773
Clean brew install error with punic init - AttributeError: 'set' object has no attribute 'keys' Call stack: Traceback (most recent call last): File "/usr/local/bin/punic", line 11, in <module> load_entry_point('punic==0.2.9', 'console_scripts', 'punic')() File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/punic_cli.py", line 397, in main punic_cli() File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 697, in main rv = self.invoke(ctx) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 1066, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/punic_cli.py", line 341, in init config_init(**kwargs) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/config_init.py", line 74, in config_init xcode_version = _prompt("Xcode Version", _xcode_versions()) File "/usr/local/Cellar/punic/0.2.9/libexec/lib/python2.7/site-packages/punic/config_init.py", line 41, in _xcode_versions return sorted([six.text_type(version) for version in Xcode._all_xcodes.keys()]) AttributeError: 'set' object has no attribute 'keys' I had punic installed a while back, not sure if that makes a difference, so I uninstalled punic first, ie. pip uninstall punic. Did a fresh install after updating/upgrading brew. Running punic init yields the following callstack. Is it a problem with my Python version (not versed in Python)? It just seems to be a problem with the object type of_all_xcodes. Thanks! How did you install? Are you running from a non-release branch? I'm doing a straight brew install punic, which installed 0.2.9, so I don't believe so. I'm using Xcode 8.3.1 only (ie. no beta/old versions alongside). Ah ok - that's bad. Will look into this. OK fixed and the home-brew tap is updated as well. Can you verify with a brew update and brew upgrade punic? Seems to have done the trick. Thanks!
gharchive/issue
2017-04-19T18:25:24
2025-04-01T06:45:44.359559
{ "authors": [ "mmmilo", "schwa" ], "repo": "schwa/punic", "url": "https://github.com/schwa/punic/issues/67", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
309528578
Add ability to cancel execution (I.e. stop button) ... aca33d494a8775737a99f8dba393391f80a0a492
gharchive/issue
2018-03-28T20:56:15
2025-04-01T06:45:44.360531
{ "authors": [ "Lucaszw" ], "repo": "sci-bots/microdrop-3", "url": "https://github.com/sci-bots/microdrop-3/issues/57", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1722976753
[Experiement] virtual simple index. Having to publish all wheels in the same repository is complicated from the size of the storage and token regenration. Any changes to the index URL needs update to all the CI. So the idea is to have a virtual index that just point to the original one(s) that we can manipulate the way we want: Not sure this is the best place, but that may affect spec 4 and 5 it does not support the top /simple/ that list all the packages, nor the data-python-requires. It appears also the Pypi now expose wheel metadata w/o downloading them, but i'm unsure how.
gharchive/pull-request
2023-05-24T00:30:18
2025-04-01T06:45:44.386548
{ "authors": [ "Carreau" ], "repo": "scientific-python/specs", "url": "https://github.com/scientific-python/specs/pull/194", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1979776241
ERROR: Could not build wheels for ninja The problem: I'm trying to install ninja in a Docker container on an arm32v7/debian image which doesn't have a corresponding wheel on PyPi and I am getting an error during the build process. Currently I'm using python 3.9 and pip 23.3.1 I tried different versions of either of them and got the same result Dockerfile: FROM arm32v7/debian:bullseye ENV DEBIAN_FRONTEND=noninteractive RUN \ apt-get update --yes && \ apt-get install --no-install-recommends --yes \ build-essential cmake bison flex re2c binutils \ fakeroot file ca-certificates \ subversion git \ binutils-arm-linux-gnueabihf \ pkg-config libtool automake autoconf \ python python3-pip zlib1g-dev libxrandr-dev \ libdrm-dev libxdamage-dev libxcb-glx0-dev libxcb-shm0-dev \ libx11-xcb-dev libxcb-dri2-0-dev libxcb-dri3-dev \ libxcb-present-dev libxshmfence-dev libxxf86vm-dev RUN python3 -m pip install --upgrade pip RUN python3 -m pip install --upgrade setuptools RUN python3 -m pip install --upgrade wheel RUN python3 -m pip install --upgrade scikit-build RUN python3 -m pip install ninja --verbose Console output: [+] Building 439.2s (11/11) FINISHED docker:desktop-linux => [internal] load build definition from Dockerfile 0.0s => => transferring dockerfile: 867B 0.0s => [internal] load .dockerignore 0.0s => => transferring context: 2B 0.0s => [internal] load metadata for docker.io/arm32v7/debian:bullseye 1.6s => [auth] arm32v7/debian:pull token for registry-1.docker.io 0.0s => [1/7] FROM docker.io/arm32v7/debian:bullseye@sha256:bede2623dae269454c5b6dd4af15a10810a5f4ef75963d4eb6531628f98bd633 0.0s => CACHED [2/7] RUN apt-get update --yes && apt-get install --no-install-recommends --yes build-essential cmake bison flex re2c binutils fakeroot file ca-certificates subversion git binutils-arm-linux-gnueabihf pkg 0.0s => CACHED [3/7] RUN python3 -m pip install --upgrade pip 0.0s => CACHED [4/7] RUN python3 -m pip install --upgrade setuptools 0.0s => CACHED [5/7] RUN python3 -m pip install --upgrade wheel 0.0s => CACHED [6/7] RUN python3 -m pip install --upgrade scikit-build 0.0s => ERROR [7/7] RUN python3 -m pip install ninja --verbose 437.6s ------ > [7/7] RUN python3 -m pip install ninja --verbose: 1.604 Using pip 23.3.1 from /usr/local/lib/python3.9/dist-packages/pip (python 3.9) 2.285 Collecting ninja 2.584 Downloading ninja-1.11.1.1.tar.gz (132 kB) 2.897 ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 132.4/132.4 kB 458.9 kB/s eta 0:00:00 2.977 Installing build dependencies: started 2.978 Running command pip subprocess to install build dependencies 5.822 Collecting setuptools>=42 5.831 Using cached setuptools-68.2.2-py3-none-any.whl.metadata (6.3 kB) 5.975 Collecting scikit-build 5.983 Using cached scikit_build-0.17.6-py3-none-any.whl.metadata (14 kB) 6.333 Collecting setuptools-scm[toml] 6.543 Downloading setuptools_scm-8.0.4-py3-none-any.whl.metadata (6.4 kB) 6.973 Collecting packaging>=20 (from setuptools-scm[toml]) 6.981 Using cached packaging-23.2-py3-none-any.whl.metadata (3.2 kB) 7.119 Collecting typing-extensions (from setuptools-scm[toml]) 7.185 Downloading typing_extensions-4.8.0-py3-none-any.whl.metadata (3.0 kB) 7.318 Collecting tomli>=1 (from setuptools-scm[toml]) 7.324 Using cached tomli-2.0.1-py3-none-any.whl (12 kB) 7.583 Collecting distro (from scikit-build) 7.590 Using cached distro-1.8.0-py3-none-any.whl (20 kB) 7.806 Collecting wheel>=0.32.0 (from scikit-build) 7.813 Using cached wheel-0.41.3-py3-none-any.whl.metadata (2.2 kB) 7.880 Using cached setuptools-68.2.2-py3-none-any.whl (807 kB) 7.893 Using cached scikit_build-0.17.6-py3-none-any.whl (84 kB) 7.900 Using cached packaging-23.2-py3-none-any.whl (53 kB) 7.908 Using cached wheel-0.41.3-py3-none-any.whl (65 kB) 7.958 Downloading setuptools_scm-8.0.4-py3-none-any.whl (42 kB) 8.026 ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 42.1/42.1 kB 686.9 kB/s eta 0:00:00 8.084 Downloading typing_extensions-4.8.0-py3-none-any.whl (31 kB) 8.813 Installing collected packages: wheel, typing-extensions, tomli, setuptools, packaging, distro, setuptools-scm, scikit-build 11.24 Successfully installed distro-1.8.0 packaging-23.2 scikit-build-0.17.6 setuptools-68.2.2 setuptools-scm-8.0.4 tomli-2.0.1 typing-extensions-4.8.0 wheel-0.41.3 11.24 WARNING: Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv 11.36 Installing build dependencies: finished with status 'done' 11.37 Getting requirements to build wheel: started 11.37 Running command Getting requirements to build wheel 12.27 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:308: UserWarning: git archive did not support describe output 12.27 warnings.warn("git archive did not support describe output") 12.27 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:327: UserWarning: unprocessed git archival found (no export subst applied) 12.28 warnings.warn("unprocessed git archival found (no export subst applied)") 12.41 running egg_info 12.41 writing src/ninja.egg-info/PKG-INFO 12.42 writing dependency_links to src/ninja.egg-info/dependency_links.txt 12.42 writing entry points to src/ninja.egg-info/entry_points.txt 12.43 writing requirements to src/ninja.egg-info/requires.txt 12.43 writing top-level names to src/ninja.egg-info/top_level.txt 12.49 ERROR setuptools_scm._file_finders.git listing git files failed - pretending there aren't any 12.49 reading manifest file 'src/ninja.egg-info/SOURCES.txt' 12.50 reading manifest template 'MANIFEST.in' 12.50 warning: no previously-included files matching '*' found under directory '_skbuild' 12.50 warning: no previously-included files matching '*' found under directory 'Ninja-src' 12.50 warning: no files found matching 'versioneer.py' 12.50 adding license file 'LICENSE_Apache_20' 12.50 adding license file 'AUTHORS.rst' 12.51 writing manifest file 'src/ninja.egg-info/SOURCES.txt' 12.56 Getting requirements to build wheel: finished with status 'done' 12.57 Preparing metadata (pyproject.toml): started 12.57 Running command Preparing metadata (pyproject.toml) 13.48 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:308: UserWarning: git archive did not support describe output 13.48 warnings.warn("git archive did not support describe output") 13.48 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:327: UserWarning: unprocessed git archival found (no export subst applied) 13.48 warnings.warn("unprocessed git archival found (no export subst applied)") 13.59 running dist_info 13.60 creating /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info 13.60 writing /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/PKG-INFO 13.61 writing dependency_links to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/dependency_links.txt 13.61 writing entry points to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/entry_points.txt 13.62 writing requirements to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/requires.txt 13.62 writing top-level names to /tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/top_level.txt 13.62 writing manifest file '/tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/SOURCES.txt' 13.67 ERROR setuptools_scm._file_finders.git listing git files failed - pretending there aren't any 13.68 reading manifest file '/tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/SOURCES.txt' 13.68 reading manifest template 'MANIFEST.in' 13.68 warning: no previously-included files matching '*' found under directory '_skbuild' 13.68 warning: no previously-included files matching '*' found under directory 'Ninja-src' 13.68 warning: no files found matching 'versioneer.py' 13.68 adding license file 'LICENSE_Apache_20' 13.69 adding license file 'AUTHORS.rst' 13.69 writing manifest file '/tmp/pip-modern-metadata-j639jt5h/ninja.egg-info/SOURCES.txt' 13.69 creating '/tmp/pip-modern-metadata-j639jt5h/ninja-1.11.1.1.dist-info' 13.75 Preparing metadata (pyproject.toml): finished with status 'done' 13.79 Building wheels for collected packages: ninja 13.79 Building wheel for ninja (pyproject.toml): started 13.80 Running command Building wheel for ninja (pyproject.toml) 14.69 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:308: UserWarning: git archive did not support describe output 14.69 warnings.warn("git archive did not support describe output") 14.69 /tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/setuptools_scm/git.py:327: UserWarning: unprocessed git archival found (no export subst applied) 14.69 warnings.warn("unprocessed git archival found (no export subst applied)") 14.84 14.84 14.84 -------------------------------------------------------------------------------- 14.85 -- Trying 'Ninja' generator 14.85 -------------------------------- 14.85 --------------------------- 14.85 ---------------------- 14.85 ----------------- 14.85 ------------ 14.85 ------- 14.85 -- 14.93 Not searching for unused variables given on the command line. 15.02 CMake Error: CMake was unable to find a build program corresponding to "Ninja". CMAKE_MAKE_PROGRAM is not set. You probably need to select a different build tool. 15.02 -- Configuring incomplete, errors occurred! 15.02 See also "/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_cmake_test_compile/build/CMakeFiles/CMakeOutput.log". 15.04 -- 15.04 ------- 15.04 ------------ 15.04 ----------------- 15.04 ---------------------- 15.04 --------------------------- 15.04 -------------------------------- 15.04 -- Trying 'Ninja' generator - failure 15.04 -------------------------------------------------------------------------------- 15.04 15.04 15.04 15.05 -------------------------------------------------------------------------------- 15.05 -- Trying 'Unix Makefiles' generator 15.05 -------------------------------- 15.05 --------------------------- 15.05 ---------------------- 15.05 ----------------- 15.05 ------------ 15.05 ------- 15.05 -- 15.13 Not searching for unused variables given on the command line. 15.85 -- The C compiler identification is GNU 10.2.1 15.92 -- Detecting C compiler ABI info 16.74 -- Detecting C compiler ABI info - done 16.84 -- Check for working C compiler: /usr/bin/cc - skipped 16.84 -- Detecting C compile features 16.84 -- Detecting C compile features - done 17.50 -- The CXX compiler identification is GNU 10.2.1 17.56 -- Detecting CXX compiler ABI info 18.37 -- Detecting CXX compiler ABI info - done 18.46 -- Check for working CXX compiler: /usr/bin/c++ - skipped 18.47 -- Detecting CXX compile features 18.47 -- Detecting CXX compile features - done 18.48 -- Configuring done 18.48 -- Generating done 18.49 -- Build files have been written to: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_cmake_test_compile/build 18.50 -- 18.50 ------- 18.50 ------------ 18.50 ----------------- 18.50 ---------------------- 18.50 --------------------------- 18.50 -------------------------------- 18.50 -- Trying 'Unix Makefiles' generator - success 18.50 -------------------------------------------------------------------------------- 18.50 18.51 Configuring Project 18.51 Working directory: 18.51 /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build 18.51 Command: 18.51 /usr/bin/cmake /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce -G 'Unix Makefiles' --no-warn-unused-cli -DCMAKE_INSTALL_PREFIX:PATH=/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-install -DPYTHON_VERSION_STRING:STRING=3.9.2 -DSKBUILD:INTERNAL=TRUE -DCMAKE_MODULE_PATH:PATH=/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/resources/cmake -DPYTHON_EXECUTABLE:PATH=/usr/bin/python3 -DPYTHON_INCLUDE_DIR:PATH=/usr/include/python3.9 -DPython_EXECUTABLE:PATH=/usr/bin/python3 -DPython_ROOT_DIR:PATH=/usr -DPython_FIND_REGISTRY:STRING=NEVER -DPython_INCLUDE_DIR:PATH=/usr/include/python3.9 -DPython3_EXECUTABLE:PATH=/usr/bin/python3 -DPython3_ROOT_DIR:PATH=/usr -DPython3_FIND_REGISTRY:STRING=NEVER -DPython3_INCLUDE_DIR:PATH=/usr/include/python3.9 -DCMAKE_BUILD_TYPE:STRING=Release 18.51 18.59 Not searching for unused variables given on the command line. 19.19 -- The C compiler identification is GNU 10.2.1 19.80 -- The CXX compiler identification is GNU 10.2.1 19.86 -- Detecting C compiler ABI info 20.64 -- Detecting C compiler ABI info - done 20.73 -- Check for working C compiler: /usr/bin/cc - skipped 20.73 -- Detecting C compile features 20.73 -- Detecting C compile features - done 20.75 -- Detecting CXX compiler ABI info 21.54 -- Detecting CXX compiler ABI info - done 21.63 -- Check for working CXX compiler: /usr/bin/c++ - skipped 21.63 -- Detecting CXX compile features 21.63 -- Detecting CXX compile features - done 21.64 -- ********************************************* 21.64 -- Ninja Python Distribution 21.64 -- 21.64 -- BUILD_VERBOSE : OFF 21.64 -- RUN_NINJA_TEST : ON 21.64 -- 21.64 -- ARCHIVE_DOWNLOAD_DIR : /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build 21.64 -- 21.64 -- src_archive : unix_source 21.64 -- <src_archive>_url : https://github.com/Kitware/ninja/archive/v1.11.1.g95dee.kitware.jobserver-1.tar.gz 21.65 -- <src_archive>_sha256 : 7ba84551f5b315b4270dc7c51adef5dff83a2154a3665a6c9744245c122dd0db 21.65 -- ********************************************* 21.80 -- download_ninja_source - URL: https://github.com/Kitware/ninja/archive/v1.11.1.g95dee.kitware.jobserver-1.tar.gz 21.80 -- SuperBuild - CMAKE_BUILD_TYPE: Release 21.90 -- Configuring done 21.93 -- Generating done 21.94 -- Build files have been written to: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build 22.32 Scanning dependencies of target download_ninja_source 22.41 [ 5%] Creating directories for 'download_ninja_source' 22.87 [ 11%] Performing download step (download, verify and extract) for 'download_ninja_source' 22.95 -- Downloading... 22.96 dst='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/v1.11.1.g95dee.kitware.jobserver-1.tar.gz' 22.96 timeout='none' 22.96 -- Using src='https://github.com/Kitware/ninja/archive/v1.11.1.g95dee.kitware.jobserver-1.tar.gz' 26.44 -- verifying file... 26.44 file='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/v1.11.1.g95dee.kitware.jobserver-1.tar.gz' 26.45 -- Downloading... done 26.65 -- extracting... 26.65 src='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/v1.11.1.g95dee.kitware.jobserver-1.tar.gz' 26.65 dst='/tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/Ninja-src' 26.65 -- extracting... [tar xfz] 26.77 -- extracting... [analysis] 26.77 -- extracting... [rename] 26.77 -- extracting... [clean up] 26.77 -- extracting... done 26.92 [ 16%] No update step for 'download_ninja_source' 27.10 [ 22%] No patch step for 'download_ninja_source' 27.29 [ 27%] No configure step for 'download_ninja_source' 27.50 [ 33%] No build step for 'download_ninja_source' 27.72 [ 38%] No install step for 'download_ninja_source' 27.93 [ 44%] Completed 'download_ninja_source' 28.17 [ 44%] Built target download_ninja_source 28.30 Scanning dependencies of target build_ninja 28.41 [ 50%] Creating directories for 'build_ninja' 28.90 [ 55%] No download step for 'build_ninja' 29.12 [ 61%] No update step for 'build_ninja' 29.33 [ 66%] No patch step for 'build_ninja' 29.51 [ 72%] Performing configure step for 'build_ninja' 29.60 loading initial cache file /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/build_ninja-prefix/tmp/build_ninja-cache-Release.cmake 30.32 -- The C compiler identification is GNU 10.2.1 30.95 -- The CXX compiler identification is GNU 10.2.1 31.02 -- Detecting C compiler ABI info 31.83 -- Detecting C compiler ABI info - done 31.92 -- Check for working C compiler: /usr/bin/cc - skipped 31.92 -- Detecting C compile features 31.93 -- Detecting C compile features - done 31.95 -- Detecting CXX compiler ABI info 32.84 -- Detecting CXX compiler ABI info - done 32.93 -- Check for working CXX compiler: /usr/bin/c++ - skipped 32.93 -- Detecting CXX compile features 32.94 -- Detecting CXX compile features - done 37.96 -- IPO / LTO enabled 37.97 -- Performing Test flag_no_deprecated 38.72 -- Performing Test flag_no_deprecated - Success 38.72 -- Performing Test flag_color_diag 39.47 -- Performing Test flag_color_diag - Success 39.57 -- Looking for fork 40.34 -- Looking for fork - found 40.34 -- Looking for pipe 41.10 -- Looking for pipe - found 41.15 -- Configuring done 41.35 -- Generating done 41.36 -- Build files have been written to: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build/Ninja-build 41.51 [ 77%] Performing build step for 'build_ninja' 41.82 [ 1%] Generating lexer.cc 41.95 [ 2%] Generating depfile_parser.cc 42.08 Scanning dependencies of target libninja-re2c 42.18 [ 4%] Building CXX object CMakeFiles/libninja-re2c.dir/depfile_parser.cc.o 44.05 [ 5%] Building CXX object CMakeFiles/libninja-re2c.dir/lexer.cc.o 46.17 [ 5%] Built target libninja-re2c 46.31 Scanning dependencies of target libninja 46.42 [ 7%] Building CXX object CMakeFiles/libninja.dir/src/build_log.cc.o 50.01 [ 8%] Building CXX object CMakeFiles/libninja.dir/src/build.cc.o 55.84 [ 10%] Building CXX object CMakeFiles/libninja.dir/src/clean.cc.o 59.49 [ 11%] Building CXX object CMakeFiles/libninja.dir/src/clparser.cc.o 61.63 [ 13%] Building CXX object CMakeFiles/libninja.dir/src/dyndep.cc.o 64.66 [ 14%] Building CXX object CMakeFiles/libninja.dir/src/dyndep_parser.cc.o 67.94 [ 15%] Building CXX object CMakeFiles/libninja.dir/src/debug_flags.cc.o 68.23 [ 17%] Building CXX object CMakeFiles/libninja.dir/src/deps_log.cc.o 71.21 [ 18%] Building CXX object CMakeFiles/libninja.dir/src/disk_interface.cc.o 73.30 [ 20%] Building CXX object CMakeFiles/libninja.dir/src/edit_distance.cc.o 74.99 [ 21%] Building CXX object CMakeFiles/libninja.dir/src/eval_env.cc.o 77.91 [ 23%] Building CXX object CMakeFiles/libninja.dir/src/graph.cc.o 82.09 [ 24%] Building CXX object CMakeFiles/libninja.dir/src/graphviz.cc.o 84.75 [ 26%] Building CXX object CMakeFiles/libninja.dir/src/json.cc.o 86.13 [ 27%] Building CXX object CMakeFiles/libninja.dir/src/line_printer.cc.o 87.81 [ 28%] Building CXX object CMakeFiles/libninja.dir/src/manifest_parser.cc.o 91.32 [ 30%] Building CXX object CMakeFiles/libninja.dir/src/metrics.cc.o 93.08 [ 31%] Building CXX object CMakeFiles/libninja.dir/src/missing_deps.cc.o 97.68 [ 33%] Building CXX object CMakeFiles/libninja.dir/src/parser.cc.o 99.55 [ 34%] Building CXX object CMakeFiles/libninja.dir/src/state.cc.o 103.4 [ 36%] Building CXX object CMakeFiles/libninja.dir/src/status.cc.o 106.4 [ 37%] Building CXX object CMakeFiles/libninja.dir/src/string_piece_util.cc.o 108.3 [ 39%] Building CXX object CMakeFiles/libninja.dir/src/tokenpool-gnu-make.cc.o 109.8 [ 40%] Building CXX object CMakeFiles/libninja.dir/src/util.cc.o 113.6 [ 42%] Building CXX object CMakeFiles/libninja.dir/src/version.cc.o 115.0 [ 43%] Building CXX object CMakeFiles/libninja.dir/src/subprocess-posix.cc.o 117.2 [ 44%] Building CXX object CMakeFiles/libninja.dir/src/tokenpool-gnu-make-posix.cc.o 117.9 [ 44%] Built target libninja 118.0 Scanning dependencies of target manifest_parser_perftest 118.1 [ 46%] Building CXX object CMakeFiles/manifest_parser_perftest.dir/src/manifest_parser_perftest.cc.o 121.1 [ 47%] Linking CXX executable manifest_parser_perftest 133.3 [ 47%] Built target manifest_parser_perftest 133.4 Scanning dependencies of target hash_collision_bench 133.5 [ 49%] Building CXX object CMakeFiles/hash_collision_bench.dir/src/hash_collision_bench.cc.o 135.6 [ 50%] Linking CXX executable hash_collision_bench 138.3 [ 50%] Built target hash_collision_bench 138.4 Scanning dependencies of target depfile_parser_perftest 138.5 [ 52%] Building CXX object CMakeFiles/depfile_parser_perftest.dir/src/depfile_parser_perftest.cc.o 140.3 [ 53%] Linking CXX executable depfile_parser_perftest 143.2 [ 53%] Built target depfile_parser_perftest 143.4 Scanning dependencies of target clparser_perftest 143.5 [ 55%] Building CXX object CMakeFiles/clparser_perftest.dir/src/clparser_perftest.cc.o 145.2 [ 56%] Linking CXX executable clparser_perftest 148.5 [ 56%] Built target clparser_perftest 148.6 Scanning dependencies of target canon_perftest 148.7 [ 57%] Building CXX object CMakeFiles/canon_perftest.dir/src/canon_perftest.cc.o 150.3 [ 59%] Linking CXX executable canon_perftest 152.7 [ 59%] Built target canon_perftest 152.8 Scanning dependencies of target ninja_test 152.9 [ 60%] Building CXX object CMakeFiles/ninja_test.dir/src/build_log_test.cc.o 156.3 [ 62%] Building CXX object CMakeFiles/ninja_test.dir/src/build_test.cc.o 166.6 [ 63%] Building CXX object CMakeFiles/ninja_test.dir/src/clean_test.cc.o 170.9 [ 65%] Building CXX object CMakeFiles/ninja_test.dir/src/clparser_test.cc.o 173.8 [ 66%] Building CXX object CMakeFiles/ninja_test.dir/src/depfile_parser_test.cc.o 177.3 [ 68%] Building CXX object CMakeFiles/ninja_test.dir/src/deps_log_test.cc.o 180.8 [ 69%] Building CXX object CMakeFiles/ninja_test.dir/src/disk_interface_test.cc.o 184.5 [ 71%] Building CXX object CMakeFiles/ninja_test.dir/src/dyndep_parser_test.cc.o 188.3 [ 72%] Building CXX object CMakeFiles/ninja_test.dir/src/edit_distance_test.cc.o 190.9 [ 73%] Building CXX object CMakeFiles/ninja_test.dir/src/graph_test.cc.o 195.9 [ 75%] Building CXX object CMakeFiles/ninja_test.dir/src/json_test.cc.o 198.6 [ 76%] Building CXX object CMakeFiles/ninja_test.dir/src/lexer_test.cc.o 201.3 [ 78%] Building CXX object CMakeFiles/ninja_test.dir/src/manifest_parser_test.cc.o 205.6 [ 79%] Building CXX object CMakeFiles/ninja_test.dir/src/missing_deps_test.cc.o 209.4 [ 81%] Building CXX object CMakeFiles/ninja_test.dir/src/ninja_test.cc.o 212.1 [ 82%] Building CXX object CMakeFiles/ninja_test.dir/src/state_test.cc.o 215.1 [ 84%] Building CXX object CMakeFiles/ninja_test.dir/src/string_piece_util_test.cc.o 218.0 [ 85%] Building CXX object CMakeFiles/ninja_test.dir/src/subprocess_test.cc.o 221.1 [ 86%] Building CXX object CMakeFiles/ninja_test.dir/src/test.cc.o 225.2 [ 88%] Building CXX object CMakeFiles/ninja_test.dir/src/tokenpool_test.cc.o 227.9 [ 89%] Building CXX object CMakeFiles/ninja_test.dir/src/util_test.cc.o 230.8 [ 91%] Linking CXX executable ninja_test 375.9 [ 91%] Built target ninja_test 376.0 [ 92%] Generating build/browse_py.h 376.3 Scanning dependencies of target ninja 376.4 [ 94%] Building CXX object CMakeFiles/ninja.dir/src/ninja.cc.o 382.0 [ 95%] Building CXX object CMakeFiles/ninja.dir/src/browse.cc.o 383.2 [ 97%] Linking CXX executable ninja 420.5 [ 97%] Built target ninja 420.6 Scanning dependencies of target build_log_perftest 420.7 [ 98%] Building CXX object CMakeFiles/build_log_perftest.dir/src/build_log_perftest.cc.o 423.6 [100%] Linking CXX executable build_log_perftest 435.9 [100%] Built target build_log_perftest 436.0 [ 83%] Stripping CMake executables 436.1 [ 88%] Running Ninja test suite 437.3 gmake[2]: *** [CMakeFiles/build_ninja.dir/build.make:143: build_ninja-prefix/src/build_ninja-stamp/build_ninja-run_ninja_test_suite] Error 130 437.3 gmake[1]: *** [CMakeFiles/Makefile2:97: CMakeFiles/build_ninja.dir/all] Error 2 437.3 gmake: *** [Makefile:149: all] Error 2 437.3 Traceback (most recent call last): 437.3 File "/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/setuptools_wrap.py", line 674, in setup 437.3 cmkr.make(make_args, install_target=cmake_install_target, env=env) 437.3 File "/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/cmaker.py", line 697, in make 437.3 self.make_impl(clargs=clargs, config=config, source_dir=source_dir, install_target=install_target, env=env) 437.3 File "/tmp/pip-build-env-174q2c58/overlay/lib/python3.9/site-packages/skbuild/cmaker.py", line 742, in make_impl 437.3 raise SKBuildError(msg) 437.3 437.3 An error occurred while building with CMake. 437.3 Command: 437.3 /usr/bin/cmake --build . --target install --config Release -- 437.3 Install target: 437.3 install 437.3 Source directory: 437.3 /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce 437.3 Working directory: 437.3 /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce/_skbuild/linux-armv7l-3.9/cmake-build 437.3 Please check the install target is valid and see CMake's output for more information. 437.3 437.4 error: subprocess-exited-with-error 437.4 437.4 × Building wheel for ninja (pyproject.toml) did not run successfully. 437.4 │ exit code: 1 437.4 ╰─> See above for output. 437.4 437.4 note: This error originates from a subprocess, and is likely not a problem with pip. 437.4 full command: /usr/bin/python3 /usr/local/lib/python3.9/dist-packages/pip/_vendor/pyproject_hooks/_in_process/_in_process.py build_wheel /tmp/tmpdw_4oh6w 437.4 cwd: /tmp/pip-install-db8u1c7g/ninja_caee880361904b2eb54024d544948cce 437.4 Building wheel for ninja (pyproject.toml): finished with status 'error' 437.4 ERROR: Failed building wheel for ninja 437.4 Failed to build ninja 437.4 ERROR: Could not build wheels for ninja, which is required to install pyproject.toml-based projects ------ Dockerfile:23 -------------------- 21 | RUN python3 -m pip install --upgrade scikit-build 22 | 23 | >>> RUN python3 -m pip install ninja --verbose -------------------- ERROR: failed to solve: process "/bin/sh -c python3 -m pip install ninja --verbose" did not complete successfully: exit code: 1 I’m building ninja in a RISC-V riscv64 container and it outputs the exactly same error. By my testing the latest version on ninja that can be successfully built is ** 1.10.2.3**. Errors start at 1.10.2.4. By the way, apt install ninja-build is also an option. I have the same issue on Raspberry Pi 4, with Python 3.10 in a pipenv. For me, 1.10.2.3 doesn't work either. Same issue appear when I try to install streamlit on RiscV Same issue when building for armhf / linux/arm32/v7 I don't know if building from sources got fixed on any of those platforms with 1.11.1.2 but binary armv7l wheels should now be retrieved instead of trying to build from sources. 1.11.1.2 is still broken on riscv on the NumPy CI with python3.10. @mattip, I have no trouble building in a riscv64 ghcr.io/mayeut/manylinux_2_35:2024.11.23-1 container. It seems NumPy CI is doing some clever things to do cross-compilation instead of native compilation through QEMU, I'll try to reproduce this workflow locally if I can. There's a first error that's related to setuptools_scm which suggest one of our build dependency isn't doing the right thing here - but the error does not seem to cause the build process to stop -, I guess from either one of those (or the combination), @henryiii should probably know what's going on under the hood in scikit-build-core: metadata.version.provider = "scikit_build_core.metadata.setuptools_scm" metadata.readme.provider = "scikit_build_core.metadata.fancy_pypi_readme" The other error suggests something wrong with the cross-compilation environment: /host/usr/riscv64-linux-gnu/include/wchar.h:35:10: fatal error: stddef.h: No such file or directory 35 | #include <stddef.h> That being said, for Numpy QEMU CI , you might want to filter out ninja from the build requirements as you're overriding it with the host ninja right after installation. Closing this issue as we could verify building from sources on armhf (where there's now binary wheels which should also reduce the number of users building from sources) & riscv (tested on ubuntu 22.04). Please re-open or open a new issue if the last version did not solve the issue for your use-case.
gharchive/issue
2023-11-06T18:15:55
2025-04-01T06:45:44.416508
{ "authors": [ "AnatolySh", "KumaTea", "fberndl", "gdmrw", "guerda", "mattip", "mayeut" ], "repo": "scikit-build/ninja-python-distributions", "url": "https://github.com/scikit-build/ninja-python-distributions/issues/226", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
548810719
Instructions to set up a development environment This is a tiny commit which adds a requirements-dev.txt containing the packages needed to run the suite and a short description how to set up a working environment. I always forget which packages are needed for the test suite and I run pytest a few times. This will make it a bit easier :wink: Yes, that I understood. My instructions are indeed aimed for what you have described in your last paragraph, I just skipped that it's cloned via git as I assumed it is anyways. Following that, maybe a tiny change of the line cd /path/to/... to git clone ... && cd .. is maybe enough, or what do you think? The main idea was to make it a bit simpler to get started with a green test suite. I just updated the setup.py to use extras_require with a "testing" entry. Not sure though if the tests_require is used somewhere in your CI or so. The tests_require is not explicitly relied upon in CI. I like this solution using extras_require! It's good to know that Python has a built-in concept of "optional dependencies." Alright, ready for merge then. Thanks!
gharchive/pull-request
2020-01-13T09:41:53
2025-04-01T06:45:44.437437
{ "authors": [ "jpivarski", "tamasgal" ], "repo": "scikit-hep/uproot", "url": "https://github.com/scikit-hep/uproot/pull/436", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
842855848
Some more memberwise stuff I've been trying to read some CMS edm collections, and played around a bit, finding at least one memberwise interpretation successful, with a header like >h: version >I: length >I: values_num_bytes, e.g. --+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+---+- 0 10 0 0 16 1 64 22 127 198 0 9 0 0 0 0 0 0 0 0 --- --- --- --- --- --- @ --- --- --- --- --- --- --- --- --- --- --- --- --- for a class that indeed was version 10 and had the right amount of bytes. This class had a very simple set of members (just one): _stl_container0 = uproot.containers.AsVector(False, numpy.dtype("u1")) base_names_versions = [] member_names = ['data_'] class_flags = {} @kratsg might be interested. I think we can close this for now. Perhaps in the future we can revisit. At least some work has been documented. Yeah, for now let's close this... Jim knows very well that I probably broke my brain trying to understand parts of this back then... and I'm happy to try to stab this again in the future when I have more spare cycles, or if we want to get a student working on this... I have a perfect project for them. Yeah, that's where it would be nice to get a student/fellow to jump in. I have ROOT files they can use with RooFit/RooStats stuff which are always memberwise-serialized and it's a good test bench to try and get uproot to read those out correctly. I agree; we'd need some concerted effort from someone who can concentrate on just this for at least a month. Knowing that you have a lot of sample files is a good thing. Even pedantically worse -- ROOT generates them in different formats pre 6.18 and post 6.20 -- where they're not backward-compatible readable even within ROOT... so it will be interesting to see if uproot can bridge this gap.
gharchive/pull-request
2021-03-28T23:18:18
2025-04-01T06:45:44.441462
{ "authors": [ "jpivarski", "kratsg", "nsmith-" ], "repo": "scikit-hep/uproot4", "url": "https://github.com/scikit-hep/uproot4/pull/314", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1116071174
Set baseurl to "/" to make CSS / JS loading work under https://blog.scikit-learn.org/ As discussed in #23. That should fix browsing from https://blog.scikit-learn.org/ . But that will break browsing from https://scikit-learn.org/blog/ (but maybe we don't care?) Edit: now https://scikit-learn.org/blog/ automatically redirects to https://blog.scikit-learn.org/. Since the redirect is active, let's merge to see if that fixes the rendering.
gharchive/pull-request
2022-01-27T11:02:51
2025-04-01T06:45:44.461323
{ "authors": [ "ogrisel" ], "repo": "scikit-learn/blog", "url": "https://github.com/scikit-learn/blog/pull/63", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
310845467
'NoneType' object is not iterable This uninformative error comes from LeveragingBagging, under for example, covtype.csv. See _test_file_stream_multiple_cfier.py ;the error is quite subtle, in fact it is not even explicitly an error: root - INFO - Prequential Evaluation root - INFO - Generating 7 targets. root - INFO - Pre-training on 100 samples. root - INFO - Evaluating... 'NoneType' object is not iterable root - INFO - Evaluation time: 0.000 s root - INFO - Total instances: 100 root - INFO - Global performance: root - INFO - Learner 0 - Accuracy : 0.000 The root cause seems to be in the KNN (used as default model in leveraging bag) when working with multi-label data. Good, that can probably be fixed. The thing is that sometimes the problems are very silent, .. and its difficult to see the problem. There is another one: "classes should include all valid labels that can be in y" which probably comes from sklearn -- maybe we should break the stream and return an error when these things happen, rather than continuing and producing a summary at the end as if all was well?
gharchive/issue
2018-04-03T13:50:51
2025-04-01T06:45:44.605601
{ "authors": [ "jacobmontiel", "jmread" ], "repo": "scikit-multiflow/scikit-multiflow", "url": "https://github.com/scikit-multiflow/scikit-multiflow/issues/19", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
489837421
PS: improve metrics [x] Segment registrations by AS and type (a given AS will only register a single type of segment, but this allows querying along either axis). (done by #3152) [x] Segment registrations by update (i.e. it already exists in pathdb) vs new. (done by #3152) [x] Segment requests by type (up/down/core) and number of segments and revocations returned. Have label to indicate status of CacheOnly flag. (done by #3232) [x] Segment requests errors by type (up/down/core) and error type. (details https://github.com/scionproto/scion/issues/3106#issuecomment-528821888) (done by #3232) [x] segsyncer metrics. (done by #3241) [x] Revocations received. (details https://github.com/scionproto/scion/issues/3106#issuecomment-531679988) [x] ~SCMP errors received.~ Replaced by #3107, for now. Eventually it would be nice to know what actions got an scmp error reply. [x] Crypto issues: [x] cert/trc request failures: timeouts vs not available [x] Signature verification failed (e.g. possible forgery) [x] Signature validation failed (e.g. signature from expired cert) [x] ~Dispatcher errors/reconnects.~ Replaced by #3107 [x] Implement https://github.com/scionproto/scion/issues/3121 for all periodic tasks Currently we have request counts handled generally in the messenger. We could expand the Handler interface to provide more information about a request and additional metric labels, so that we can still keep this in the messenger. @lukedirtwalker : that will works for some stuff, but not others. E.g. knowing whether a registration is updating an existing segment or not requires the business logic of the PS. @kormat that can only be decided ~after the handler is done, so it could be in a map in the handler result (maybe?) It gets hard to handle this in a generic way. E.g. a path registration should indicate the type (up/down/core) in the "result", plus whether it's an update or new. A path request should indicate the type (up/down/core) and cache_only status in the "result". It seems likely to end up a real mess. As an example of what i'm thinking of: # Number of path registrations. "result" can be one of: new, updated, err_crypto, err_db pathsrv_registrations_total{elem="ps1-1-1", type="up", src_ia="1-34", result="updated"} 443 # Number of path requests. Incremented by one for every request. "result" can be one of: cached, fetched, err_crypto, err_db, err_timeout. pathsrv_requests_total{elem="ps1-1-1", type="core", cache_only="false", dst_isd="3", result="fetched"} 11 # Number of segments in replies to path requests. Could be a histogram later if we really wanted more insight. Incremented by the number of segments in a reply. pathsrv_requests_reply_segs_total(elem="ps1-1-1", type="core", cache_only="false", dst_isd="3"} 32 @kormat question: Revocations received. should that also include revocations received as part of a SegReply ? Yeah, i think we want a metric for revocations received, with a label to say whether it's: part of a SegReply a top-level PathMgmt.sRevInfo an SCMP revocation and another label to say what the result is (expired, crypto not available, invalid, ok, etc). segsyncer metrics: # Number of segments received from segment syncing "result" can be one of: new, updated, err_crypto, err_db ps_sync_regs_total{elem="ps1-1-1", type="up", src_ia="1-34", result="updated"} 93 # Number of segment syncer executions ps_segsyncer_pushes_total{elem="ps1-1-1", dst_ia="1-32", result="ok_succsess"} 31 All metrics should be in by now. If there is something missing we should create a new issue.
gharchive/issue
2019-09-05T15:27:06
2025-04-01T06:45:44.636090
{ "authors": [ "kormat", "lukedirtwalker" ], "repo": "scionproto/scion", "url": "https://github.com/scionproto/scion/issues/3106", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
3141275
Added a Sylvester equation solver to scipy.linalg I've added a function to provide a solution to the Sylvester equation (AX + XB = Q) to the linear algebra module. The code is currently grouped with "basic" linear algebra routines as this seems to be the only applicable place to include such a solver. The routine relies on some Schur decomposition calls and a LAPACK call. If including the code in this location is sub-optimal, maybe another location in SciPy could be suggested. I did a quick read-through: the code is very clear, well-written, and well documented, and the test-coverage is great. Nice work! As far as including this in basic.py: I think it's fine. If we end up adding more specialized solvers in the future, it may be better to factor them out into a "solvers" submodule, but I don't think that's necessary right now. One more thing: you should add a brief description of the routine to the doc string in scipy/linalg/info.py. I added a doc string, but I think I tripped up adding a merge from scipy/master. Hopefully we're still ok. Looks pretty good to me too. The committer probably should rebase to get rid of the merge, and move the "Computes a solution to the Sylvester matrix equation ..." part of the docstring to its Notes section. @jakevdp: Note that you can also use git commit --amend --author="..." to change the author field of the commit (the committer field retains your name).
gharchive/issue
2012-02-08T14:17:14
2025-04-01T06:45:44.644983
{ "authors": [ "ArmstrongJ", "jakevdp", "pv" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/154", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1110250610
BUG: Documentation of stft misses the import statement for numpy Describe your issue. The sample code given for stft in the signal module misses the following line: import numpy as np Reproducing Code Example from scipy import signal import matplotlib.pyplot as plt rng = np.random.default_rng() Error message NameError: name 'np' is not defined SciPy/NumPy/Python version information 1.7.3 1.21.4 sys.version_info(major=3, minor=9, micro=9, releaselevel='final', serial=0) Have a look at https://github.com/scipy/scipy/issues/13049 @ilayn, thanks for the very quick response. Sorry I missed that issue. I was just looking for bugs related to stft and didn't expect it is a general issue. One can then probably also close my related PR.
gharchive/issue
2022-01-21T09:33:21
2025-04-01T06:45:44.648424
{ "authors": [ "hoehnp", "ilayn" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/15446", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1316022151
BUG: malloc(0) in NI_InitFilterOffsets Describe your issue. In NI_InitFilterOffsets, it's possible to end up with a calculation where malloc(0) would be called, due to one of the multiplication parameters being 0 (e.g., footprint). Passing 0 to malloc is implementation-defined and may return some pointer (not necessary NULL), but any attempts to dereference that pointer would be invalid. Here's an example patch that detects this issue and a test that fails (from my WIP dev branch): https://github.com/nkaretnikov/scipy/commit/c44143f16b527d21dc865ea026ccddddd2698fd7 https://github.com/nkaretnikov/scipy/commit/a99cbbdd62570bc61d52da6e4f2dce2943e0b31f Reproducing Code Example see above, note that the code is slightly different than in main. but here's an inlined version: diff --git a/scipy/ndimage/src/ni_support.c b/scipy/ndimage/src/ni_support.c index de7bace3a..481a7ffbb 100644 --- a/scipy/ndimage/src/ni_support.c +++ b/scipy/ndimage/src/ni_support.c @@ -583,14 +583,24 @@ int NI_InitFilterOffsets(PyArrayObject *array, npy_bool *footprint, offsets_size *= (ashape[ii] < fshape[ii] ? ashape[ii] : fshape[ii]); /* allocate offsets data: */ *offsets_bytes = offsets_size * footprint_size * sizeof(npy_intp); + if (*offsets_bytes == 0) { + PyErr_SetString(PyExc_RuntimeError, "*offsets_bytes is 0"); + goto exit; + } *offsets = malloc(*offsets_bytes); if (!*offsets) { PyErr_NoMemory(); goto exit; } + if (coordinate_offsets) { *coordinate_offsets_bytes = offsets_size * rank * footprint_size * sizeof(npy_intp); + if (*coordinate_offsets_bytes == 0) { + PyErr_SetString( + PyExc_RuntimeError, "*coordinate_offsets_bytes is 0"); + goto exit; + } *coordinate_offsets = malloc(*coordinate_offsets_bytes); if (!*coordinate_offsets) { PyErr_NoMemory(); # XXX: This ends up calling NI_BinaryErosion, which calls NI_InitFilterOffsets. # In the latter, footprint_size is computed and set to 0, so malloc(0) is called # for offsets. Then in NI_BinaryErosion, the offsets pointer is passed to # NI_FilterNext2 where it would be adjusted, but this is prevented by the size # check at the start of NI_FilterNext2, so the whole thing errors out. @pytest.mark.skip(reason="footprint_size equals 0") def test_binary_hit_or_miss_input_as_output(): rstate = numpy.random.RandomState(123) data = rstate.randint(low=0, high=2, size=100).astype(bool) # input data is not modified data_orig = data.copy() expected = ndimage.binary_hit_or_miss(data) assert_array_equal(data, data_orig) # data should now contain the expected result ndimage.binary_hit_or_miss(data, output=data) assert_array_equal(expected, data) commit 4277f939b1d63b2212a3903eb500b3ab16bc33a8 (upstream-https/main, upstream-https/HEAD) Merge: 7d9301f7c 33ea45b44 Author: Ralf Gommers <ralf.gommers@gmail.com> Date: Sat Jul 2 13:16:40 2022 +0200 Merge pull request #16444 from sabonerune/fix/messagestream_tempfile BUG: fix fail to open tempfile in messagestream.pyx (#8850) Thanks @nkaretnikov. This reproducer doesn't fail for me without your patch. Is it supposed to? Also, is the shown commit for messagestream.pyx relevant? @rgommers i've created a pr here: https://github.com/scipy/scipy/pull/16697. it's based on top of the current main, so should be easy to repro. i've also adjusted the comment there to better match the behavior in main. about the bug: passing 0 to malloc is valid as long as you pass the resulting pointer to free iiuc. that's the only guarantee you have. dereferencing this malloc pointer is not valid (since it would be either NULL or would be pointing to some address backed by 0 bytes, this is implementation-defined). the test doesn't seem to crash with asan if i try without these checks. however, it might be just because asan fails to catch this. i discovered this on my branch which implements extensive size checking. without this patch on that branch, you should get an error later due to accessing out of bounds memory. iirc, the pointer is adjusted (incremented or decremented) and then dereferenced. but it doesn't make sense to do with a pointer pointing to 0 bytes. now, as i mention above, using malloc like this seems problematic, so better to just fail early instead of relying on other code to notice. it's all due to one of the arguments used in multiplication being 0 (the footprint). here's a SO discussion on this: https://stackoverflow.com/questions/2022335/whats-the-point-of-malloc0
gharchive/issue
2022-07-24T22:14:48
2025-04-01T06:45:44.655617
{ "authors": [ "nkaretnikov", "rgommers" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/16694", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2090727004
ENH: Adding COBYQA to scipy.optimize? Is your feature request related to a problem? Please describe. As discussed in https://github.com/scipy/scipy/issues/18118, COBYLA is the only derivative-free optimization solver available in SciPy. COBYQA is a solver we developed with my colleague Zaikun Zhang from The Hong Kong Polytechnic University. It is designed to supersede COBYLA as a general solver. Describe the solution you'd like. I want to propose the inclusion of the COBYQA solver in scipy.optimize. A Python implementation of COBYQA is available at https://github.com/cobyqa/cobyqa, and the related Python package is available on PyPI. Numerical experiments we conducted show the clear superiority of COBYQA over COBYLA in general. I would be happy to share our thoughts on this inclusion. Describe alternatives you've considered. No response Additional context (e.g. screenshots, GIFs) No response Thanks @ragonneau, can you post this on the mailing list as this is required for new features? That being said I think I am +1 Pinging also @zaikunzhang Thanks @ragonneau for the proposal. Is there sign of general adoption yet? We have strict academic criteria for inclusion in SciPy. Usually we don't follow trends and add new things but wait until there is a large consensus around adoption of new methods. https://docs.scipy.org/doc/scipy/dev/core-dev/index.html#core-dev-guide Based on that alone I would be -1. I'd like to understand the lineage and the status of this method. If it's completely new, I'd second Pamphile's opinion: it should mature as a separate package. However, is it a completely new method? There is COBYLA by Powell; then there's PRIMA by Zhang & Powell, and now here's COBYQA by Ragonneau & Zhang. So PRIMA is a streamlined and modernized implementation of COBYLA, and we sort of planned to replace the latter by the former. Now PRIMA is still in FORTRAN so https://github.com/scipy/scipy/issues/18566 applies. So is COBYQA separate if related to COBYLA/PRIMA, or can it be thought of an improved version of the same method? If it is, we can I guess think of formulating a plan to skip PRIMA and go straight to COBYQA maybe. Also cross-ref gh-1477 and gh-18118. I'd like to understand the lineage and the status of this method. If it's completely new, I'd second Pamphile's opinion: it should mature as a separate package. However, is it a completely new method? There is COBYLA by Powell; then there's PRIMA by Zhang & Powell, and now here's COBYQA by Ragonneau & Zhang. So PRIMA is a streamlined and modernized implementation of COBYLA, and we sort of planned to replace the latter by the former. Now PRIMA is still in FORTRAN so #18566 applies. No, PRIMA is not in FORTRAN. PRIMA is in Fortran, which I regard as a different language that conveniently benefits from the abundant libraries in FORTRAN without bearing the historical burden. So is COBYQA separate if related to COBYLA/PRIMA, or can it be thought of an improved version of the same method? If it is, we can I guess think of formulating a plan to skip PRIMA and go straight to COBYQA maybe. COBYQA is not COBYLA. COBYQA is a completely new and different algorithm, using more advanced techniques than COBYLA. This particularly includes SQP and derivative-free symmetric Broyden update. Thank you @zaikunzhang It is in Fortran, which I regard as a different language that conveniently benefits from the abundant libraries in FORTRAN without bearing the historical burden. OK, let's keep the F77 vs modern Fortran discussion in gh-18118 or elsewhere. COBYQA is not COBYLA. COBYQA is a completely new and different algorithm, using more advanced techniques than COBYLA. This particularly includes SQP and derivative-free symmetric Broyden update. Tom @ragonneau and I (Zaikun Zhang) developed COBYQA as part of Tom's Ph.D. thesis. Thanks. So is COBYQA in some sense a superset of COBYLA? As in, can it be expected to be strictly better than COBYLA, would domain experts (you) recommend it over cobyla/prima? shgo was also a very new solver when it was added to scipy so I would not directly reject COBYQA. How does COBYQA perform for example on the CUTEST benchmark compared to what scipy currently has to offer? Not sure about the quality of our benchmark suite there though. Regarding COBYLA I would strongly argue against removing it for COBYLA. In my experience, no optimizer is strictly better than another (no free lunch theorem etc. ..) and COBYLA is a famous and widely used algorithm. Thanks. So is COBYQA in some sense a superset of COBYLA? As in, can it be expected to be strictly better than COBYLA, would domain experts (you) recommend it over cobyla/prima? No, COBYQA is not a superset of COBYLA. They are different algorithms, just like GMRES is not a superset of CG or vice versa. From a mathematical perspective, the techniques used in COBYQA are more advanced. We have developed COBYQA as a successor of COBYLA. COBYQA is a package under development, but COBYLA has been stabilized for almost 30 years, the last major change being the modernization I made. Note that "under development" does not mean that it is not ready for production use, but means that new features and new improvements will continue to be made (recall that the Linux kernel is under development). No algorithm can always outperform another one on all problems. We, as algorithm developers, try to develop algorithms that perform better than others on most test problems we have. This does not mean that our algorithm will outperform others on a particular problem. However, a particular user normally only cares about a single particular problem. If a user asks my advice for solving a nonlinearly constrained optimization problem, I would suggest trying both COBYLA and COBYQA. Both have their advantages. Note that it is improper to compare PRIMA with COBYQA, the former being a package of five solvers while the latter being a particular solver. For unconstrained, bound-constrained, and linearly-constrained problems, the solvers in PRIMA are the state of the art, and the performance of COBYQA is comparable when the number of variables is at most 50. We have not tested larger problems systematically due to the speed limitation of Python, but PRIMA algorithms are tested on larger problems every day. COBYQA has its own advantages. For example, it respects the bound constraints (if any), meaning that it never evaluates the objective/constraint functions at points outside the bounds. This is the next improvement I will make to LINCOA and COBYLA. COBYQA still has much potential to be fulfilled and explored. I believe it can still be improved to obtain high-precision solutions. Derivative-free optimization methods (aka, zeroth-order methods) is attracting more and more attention due to new applications arising from machine learning and AI. My suggestion is to make both PRIMA solvers (including COBYLA) and COBYLA available in SciPy --- this will be of great value to both SciPy and the community. Thank you. Regarding COBYLA I would strongly argue against removing it for COBYQA. In my experience, no optimizer is strictly better than another (no free lunch theorem etc. ..) and COBYLA is a famous and widely used algorithm. I hold the same opinion. shgo was also a very new solver when it was added to scipy so I would not directly reject COBYQA. How does COBYQA perform for example on the CUTEST benchmark compared to what scipy currently has to offer? Not sure about the quality of our benchmark suite there though. We have our own benchmark suite that one can run (see in the benchmark folder). If, as shgo, the method proves itself to be competitive, then by all means my -1 would become +1 😃 Running the benchmark would in any case be the next step for me. Running the benchmark would in any case be the next step for me. Agreed, it would be great to see the result of those. @ragonneau you can find those in benchmarks/benchmarks/optimize.py. It would be very useful to see the results of those benchmarks that include COBYLA now (and other solvers), and add COBYQA to that. We have our own benchmark suite that one can run (see in the benchmark folder). Those include benchmarks from CUTEST, which may be what @dschmitz89 meant. A Python implementation of COBYQA is available at https://github.com/cobyqa/cobyqa, I had a quick browse of this package; the code generally looks in good shape. It's not so easy to immediately see how much of it would be included in SciPy, since there's models, problem sets, and some duplicated code from SciPy. But it looks like it'd be a few thousand lines of code. Those will presumably land in a single file in the SciPy implementation (e.g., scipy/optimize/_cobyla.py). You'll probably have to put that together anyway to be able to run the benchmarks. It'll be nice to look at that once it's ready. Those will presumably land in a single file in the SciPy implementation (e.g., scipy/optimize/_cobyla.py). Thank you @rgommers, I suppose you meant scipy/optimize/_cobyqa.py. Yes indeed - that was a typo Agreed, it would be great to see the result of those. @ragonneau you can find those in benchmarks/benchmarks/optimize.py. It would be very useful to see the results of those benchmarks that include COBYLA now (and other solvers), and add COBYQA to that. I ran SciPy benchmarks locally after including COBYQA. The results for BenchDFO: [ 2.27%] ··· ============================== ============== =========== ======================== -- result type --------------------------------------------- ------------------------------------ DFO benchmark problem number solver mean_nfev min_obj ============================== ============== =========== ======================== 0 COBYLA 182.0 36.00000001629636 0 COBYQA 38.0 36.0 0 SLSQP 21.0 35.99999999999999 0 Powell 213.0 35.99999999999998 0 nelder-mead 1390.0 36.000000004916636 0 L-BFGS-B 30.0 36.00000000029089 0 BFGS 40.0 36.00000000000747 0 trust-constr 30.0 36.000000000065775 1 COBYLA 239.0 36.00000002161123 1 COBYQA 43.0 36.00000000000001 1 SLSQP 21.0 36.0 1 Powell 356.0 35.99999999999997 1 nelder-mead 1096.0 36.00000001672908 1 L-BFGS-B 50.0 36.00000000000111 1 BFGS 70.0 36.00000000020062 1 trust-constr 40.0 36.00000000004839 2 COBYLA 93.0 8.38460155098156 2 COBYQA 102.0 8.380281696617237 2 SLSQP 23.0 8.380281690140846 2 Powell 203.0 8.380281690140844 2 nelder-mead 276.0 8.380314322959702 2 L-BFGS-B 32.0 8.380281690152536 2 BFGS 252.0 8.380281698131483 2 trust-constr 128.0 8.380281690145555 3 COBYLA 107.0 8.400780446868142 3 COBYQA 121.0 8.380281704183599 3 SLSQP 23.0 8.380284602260867 3 Powell 182.0 8.380281690140844 3 nelder-mead 320.0 8.380314069260285 3 L-BFGS-B 40.0 8.380281690152602 3 BFGS 372.0 8.380281690224777 3 trust-constr 168.0 8.380281690143562 4 COBYLA 103.0 9.89172925972701 4 COBYQA 116.0 9.880597101883852 4 SLSQP 23.0 9.880597014925373 4 Powell 209.0 9.880597014925371 4 nelder-mead 267.0 9.880630904843933 4 L-BFGS-B 112.0 9.880597014925657 4 BFGS 341.0 9.880597015302904 4 trust-constr 136.0 9.880597014929531 5 COBYLA 124.0 9.880673204658704 5 COBYQA 109.0 9.880597794292857 5 SLSQP 23.0 9.880597015430164 5 Powell 139.0 9.880597014925371 5 nelder-mead 310.0 9.880645653536046 5 L-BFGS-B 40.0 9.880597014932862 5 BFGS 380.0 9.880597014940303 5 trust-constr 200.0 9.880597014935532 6 COBYLA 1000.0 0.2761343670249867 6 COBYQA 136.0 3.4577885926458176e-11 6 SLSQP 109.0 3.534301293924672e-05 6 Powell 607.0 1.7404243721438573e-26 6 nelder-mead 159.0 8.177661197416674e-10 6 L-BFGS-B 126.0 4.006329767353906e-07 6 BFGS 117.0 3.403318352733089e-11 6 trust-constr 177.0 8.189074233317919e-09 7 COBYLA 1000.0 18.15444498891401 7 COBYQA 342.0 1.6575428203999043e-08 7 SLSQP 295.0 2.4772818584112955e-05 7 Powell 950.0 1.1093356479670479e-31 7 nelder-mead 325.0 4.3387935016943286e-10 7 L-BFGS-B 33.0 18.35859455687261 7 BFGS 390.0 2.1519670672414852e-11 7 trust-constr 372.0 2.8716495252807545e-11 8 COBYLA 1000.0 0.0063088149736903664 8 COBYQA 87.0 4.72095533231052e-08 8 SLSQP 112.0 2.467522680839185e-06 8 Powell 60.0 0.0 8 nelder-mead 142.0 0.0003575878654431348 8 L-BFGS-B 120.0 1.2458244845337688e-08 8 BFGS 136.0 3.645847385308141e-12 8 trust-constr 128.0 4.664490098037049e-12 9 COBYLA 1000.0 1.302791737849277 9 COBYQA 171.0 2.6707378731301188e-08 9 SLSQP 114.0 3.1523978560448592e-06 9 Powell 54.0 0.0 9 nelder-mead 110.0 0.0001185672468711772 9 L-BFGS-B 140.0 1.0981915721898683e-05 9 BFGS 160.0 4.3547185576347985e-12 9 trust-constr 172.0 5.033947368965333e-11 10 COBYLA 1000.0 0.0019940963089381952 10 COBYQA 312.0 1.2655194279446566e-10 10 SLSQP 99.0 9.231453061284653e-05 10 Powell 908.0 1.692602785645291e-13 10 nelder-mead 305.0 1.3905860499424258e-06 10 L-BFGS-B 100.0 3.5705209681058525e-05 10 BFGS 165.0 3.266335996489636e-07 10 trust-constr 190.0 3.424058956351133e-07 11 COBYLA 1000.0 0.007588035153884722 11 COBYQA 232.0 1.9523309860714916e-08 11 SLSQP 179.0 3.44391715908274e-05 11 Powell 1577.0 3.013054708612935e-22 11 nelder-mead 380.0 1.2207288155937388e-06 11 L-BFGS-B 165.0 2.5692617247172744e-05 11 BFGS 270.0 1.7148717367764208e-07 11 trust-constr 355.0 2.2298540818792096e-07 12 COBYLA 1000.0 63.35600768650094 12 COBYQA 73.0 48.98425367984218 12 SLSQP 28.0 48.9842541402506 12 Powell 118.0 48.98425367924135 12 nelder-mead 120.0 48.98425367981376 12 L-BFGS-B 54.0 48.98425585393734 12 BFGS 27.0 48.98425367924926 12 trust-constr 60.0 48.98425367924055 13 COBYLA 1000.0 52.5180366249726 13 COBYQA 90.0 48.98425367987076 13 SLSQP 47.0 48.984285104912686 13 Powell 232.0 48.984253679248496 13 nelder-mead 119.0 1.7922698821282258e-09 13 L-BFGS-B 90.0 48.984313597884025 13 BFGS 75.0 48.98425367924157 13 trust-constr 87.0 48.98425367924004 14 COBYLA 1000.0 0.010846968025621654 14 COBYQA 104.0 0.008214877370595158 14 SLSQP 36.0 0.010926876593972496 14 Powell 435.0 0.008214880387696202 14 nelder-mead 226.0 0.008214877316410097 14 L-BFGS-B 36.0 0.009974372219725124 14 BFGS 92.0 0.008214877322487959 14 trust-constr 96.0 0.008214931318406598 15 COBYLA 1000.0 8.3527643805625 15 COBYQA 119.0 0.008214877325203946 15 SLSQP 80.0 0.010852444332488186 15 Powell 212.0 1.7749269756254313 15 nelder-mead 346.0 0.008214877318251797 15 L-BFGS-B 92.0 0.010443321667281222 15 BFGS 136.0 0.008214877306852426 15 trust-constr 1924.0 0.008214941243166055 16 COBYLA 1000.0 0.00038308147268715565 16 COBYQA 213.0 0.0003075056071085903 16 SLSQP 22.0 0.0005011638818990982 16 Powell 517.0 0.00030750560950841397 16 nelder-mead 260.0 0.00030750561104950664 16 L-BFGS-B 45.0 0.0004998454317253484 16 BFGS 165.0 0.00030750717208645955 16 trust-constr 170.0 0.000309855462194756 17 COBYLA 49.0 7191620.061242092 17 COBYQA 240.0 125887.6768230767 17 SLSQP 113.0 1417869686.359142 17 Powell 267.0 106966.8745748054 17 nelder-mead 600.0 6409.119264481873 17 L-BFGS-B 16.0 6977375.029202724 17 BFGS 1768.0 120.82601972635028 17 trust-constr 3376.0 22723.636975175406 18 COBYLA 1000.0 0.010975991274795866 18 COBYQA 631.0 0.0022876786037281716 18 SLSQP 145.0 0.0031019739183846718 18 Powell 1742.0 0.0022876970870087574 18 nelder-mead 1200.0 0.004794624522076027 18 L-BFGS-B 133.0 0.009260117102741666 18 BFGS 266.0 0.002287670084190521 18 trust-constr 371.0 0.0022877933510461667 19 COBYLA 1000.0 0.01449258959907267 19 COBYQA 1000.0 0.003041678435983581 19 SLSQP 334.0 0.002387635461912692 19 Powell 1840.0 0.017136978287569753 19 nelder-mead 1200.0 0.11412012392582763 19 L-BFGS-B 224.0 0.01039282675002245 19 BFGS 518.0 0.002287670112758827 19 trust-constr 539.0 0.0022876807830514287 20 COBYLA 1000.0 0.012314747076097223 20 COBYQA 1000.0 4.926796490034533e-05 20 SLSQP 435.0 3.7403380991915257e-05 20 Powell 3507.0 0.00013999662219714502 20 nelder-mead 1275.0 0.0037644989112963437 20 L-BFGS-B 270.0 0.0028624144428293146 20 BFGS 410.0 3.7976544143093916e-05 20 trust-constr 540.0 3.9115677318496754e-05 21 COBYLA 1000.0 0.19021806725341522 21 COBYQA 1000.0 0.12473258117608635 21 SLSQP 975.0 0.005234314850263207 21 Powell 5037.0 0.10605836851573695 21 nelder-mead 1800.0 0.5377026960946113 21 L-BFGS-B 500.0 0.0002881772390133559 21 BFGS 1000.0 2.1010434078081007e-06 21 trust-constr 1260.0 3.997821365625117e-05 22 COBYLA 1000.0 0.10003177690766717 22 COBYQA 1000.0 0.00013082934380459935 22 SLSQP 613.0 0.007622482850835413 22 Powell 9020.0 0.0009381552314367413 22 nelder-mead 2400.0 0.008256161684722775 22 L-BFGS-B 715.0 0.00033578796198918936 22 BFGS 936.0 1.8927701687374778e-06 22 trust-constr 1105.0 1.3321166924342236e-05 23 COBYLA 1000.0 0.2362971492708929 23 COBYQA 1000.0 0.10741551690232379 23 SLSQP 1331.0 0.10106796888625674 23 Powell 4778.0 6.567880194260053 23 nelder-mead 2400.0 5.580001849914321 23 L-BFGS-B 793.0 0.011328602404693662 23 BFGS 2041.0 0.000278772211425114 23 trust-constr 1612.0 1.3460798054702056e-05 24 COBYLA 1000.0 0.0027178520030068 24 COBYQA 177.0 6.38970543076227e-14 24 SLSQP 85.0 0.039023423059962446 24 Powell 1077.0 0.07583186991605789 24 nelder-mead 480.0 0.07558874075499751 24 L-BFGS-B 100.0 0.000234110576993369 24 BFGS 108.0 4.1709385349620864e-09 24 trust-constr 132.0 2.214996479723849e-06 25 COBYLA 32.0 256.1647690452771 25 COBYQA 73.0 124.36218407850306 25 SLSQP 16.0 2020.0 25 Powell 303.0 124.3621843304955 25 nelder-mead 72.0 124.3621848396066 25 L-BFGS-B 72.0 214.29894954503868 25 BFGS 147.0 124.36218235561685 25 trust-constr 180.0 124.36218235562204 26 COBYLA 678.0 85822.20465825836 26 COBYQA 161.0 85822.20162724197 26 SLSQP 130.0 85822.20163050466 26 Powell 270.0 85822.24229503315 26 nelder-mead 333.0 85822.20162974804 26 L-BFGS-B 70.0 85822.28170932793 26 BFGS 195.0 85822.20162635628 26 trust-constr 220.0 85822.2016263596 27 COBYLA 928.0 85822.20748496873 27 COBYQA 421.0 85822.20162810793 27 SLSQP 252.0 85822.20162647964 27 Powell 260.0 85822.29989083631 27 nelder-mead 397.0 85822.20164251165 27 L-BFGS-B 155.0 85822.2134844625 27 BFGS 435.0 85822.20162635644 27 trust-constr 390.0 85822.20162635874 28 COBYLA 274.0 4.567933585235919e-06 28 COBYQA 202.0 1.793038448544232e-09 28 SLSQP 73.0 0.0001275291143854763 28 Powell 2017.0 2.62902981284408e-22 28 nelder-mead 630.0 4.168900076322192e-09 28 L-BFGS-B 112.0 9.673157105942231e-07 28 BFGS 140.0 1.3774502648136339e-10 28 trust-constr 7000.0 0.019837220256553954 29 COBYLA 169.0 7.228441475939319e-07 29 COBYQA 173.0 1.2167995824458087e-08 29 SLSQP 58.0 0.00017792262972978623 29 Powell 2050.0 6.554533948298135e-23 29 nelder-mead 472.0 4.392761093121637e-07 29 L-BFGS-B 112.0 2.0258935358234505e-05 29 BFGS 176.0 6.121604659655201e-11 29 trust-constr 152.0 2.653738494450361e-10 30 COBYLA 1000.0 0.0036544588617460697 30 COBYQA 328.0 0.0035168812562413423 30 SLSQP 72.0 0.006287589283329275 30 Powell 2782.0 0.0035168774842111537 30 nelder-mead 813.0 0.0037430185611118747 30 L-BFGS-B 180.0 0.0036331210839248406 30 BFGS 261.0 0.0035168737405634758 30 trust-constr 333.0 0.0035168745891778416 31 COBYLA 507.0 9.563260675050992e-06 31 COBYQA 359.0 9.022403180824982e-08 31 SLSQP 113.0 7.559376195674753e-05 31 Powell 4760.0 1.4405961280325035e-21 31 nelder-mead 1008.0 1.993137661280346e-05 31 L-BFGS-B 190.0 0.0001645285146463783 31 BFGS 290.0 4.2026392909231023e-11 31 trust-constr 5000.0 4.6482458943424287e-10 32 COBYLA 1000.0 0.004996093419091057 32 COBYQA 428.0 0.004772756479272126 32 SLSQP 76.0 0.00748298302440139 32 Powell 2196.0 0.00654899679530417 32 nelder-mead 1142.0 0.004782384226390245 32 L-BFGS-B 143.0 0.007512546418004671 32 BFGS 363.0 0.006503957324546606 32 trust-constr 11000.0 0.03376312998006939 33 COBYLA 1000.0 0.004790522744041185 33 COBYQA 492.0 0.002800054880834453 33 SLSQP 80.0 0.006136919804702198 33 Powell 4493.0 0.0027998978593152275 33 nelder-mead 1799.0 0.0031707681652494613 33 L-BFGS-B 312.0 0.002825355695283585 33 BFGS 504.0 0.0027997619178927627 33 trust-constr 12000.0 0.026740538342589985 34 COBYLA 1000.0 0.002477075965308623 34 COBYQA 765.0 7.087734211593793e-08 34 SLSQP 47.0 1.8202225661529484e-05 34 Powell 2981.0 1.3549012557932144e-22 34 nelder-mead 2000.0 0.3406697389548668 34 L-BFGS-B 55.0 2.991968759920641e-05 34 BFGS 132.0 1.873957153532047e-13 34 trust-constr 154.0 1.7082620401841644e-08 35 COBYLA 1000.0 2.775269864105493 35 COBYQA 386.0 0.0660477478502633 35 SLSQP 74.0 1.1056629223458196 35 Powell 658.0 0.05064107286626231 35 nelder-mead 747.0 5.46489474798272e-05 35 L-BFGS-B 108.0 1.1059653777113296 35 BFGS 125.0 0.6345412274548067 35 trust-constr 198.0 1.105952334795245 36 COBYLA 1000.0 0.3000350186116832 36 COBYQA 1000.0 0.041631120736703206 36 SLSQP 426.0 0.04140440668157865 36 Powell 3143.0 0.040143302727005024 36 nelder-mead 2200.0 0.07609312836138543 36 L-BFGS-B 684.0 0.042042079864747596 36 BFGS 756.0 0.040137738738389564 36 trust-constr 1092.0 0.040137794964173 37 COBYLA 298.0 1.7898166135669258 37 COBYQA 232.0 1.7898135873783405 37 SLSQP 228.0 1.789825242678684 37 Powell 265.0 1.7898135868820382 37 nelder-mead 809.0 1.7898135880745636 37 L-BFGS-B 396.0 1.7898136300054583 37 BFGS 252.0 1.789813586881127 37 trust-constr 960.0 1.7898135869084242 38 COBYLA 289.0 10.238983941422939 38 COBYQA 424.0 10.238973723645511 38 SLSQP 216.0 10.238994699907447 38 Powell 494.0 10.238973850652153 38 nelder-mead 1058.0 10.238973442379468 38 L-BFGS-B 135.0 10.239986312974672 38 BFGS 396.0 10.23897342140286 38 trust-constr 405.0 10.238973421449707 39 COBYLA 704.0 18.281166339083896 39 COBYQA 379.0 18.281162006790357 39 SLSQP 276.0 18.281172707363083 39 Powell 603.0 18.281163008134268 39 nelder-mead 2000.0 18.281162945170095 39 L-BFGS-B 165.0 18.28315259151096 39 BFGS 517.0 18.28116175363231 39 trust-constr 638.0 18.281161753760827 40 COBYLA 864.0 22.26060556377675 40 COBYQA 560.0 22.260591789088142 40 SLSQP 398.0 22.2606981682437 40 Powell 682.0 22.260594329663107 40 nelder-mead 2200.0 22.260916512172354 40 L-BFGS-B 180.0 22.26368109072446 40 BFGS 564.0 22.260591734938853 40 trust-constr 792.0 22.260591734991067 41 COBYLA 807.0 26.2727789431198 41 COBYQA 622.0 26.272766516221093 41 SLSQP 367.0 26.272774470672505 41 Powell 715.0 26.272777250417114 41 nelder-mead 2400.0 26.27279966391945 41 L-BFGS-B 195.0 26.27565349718534 41 BFGS 611.0 26.27276639682053 41 trust-constr 949.0 26.272766396968294 42 COBYLA 1000.0 0.07310066404205529 42 COBYQA 1000.0 0.00014177578856341126 42 SLSQP 337.0 0.002869140620444522 42 Powell 472.0 0.0006099279648976476 42 nelder-mead 218.0 0.002941046450586494 42 L-BFGS-B 66.0 0.027570873338690554 42 BFGS 1680.0 1.2038440956761933e-07 42 trust-constr 2676.0 1.5612798805036442e-05 43 COBYLA 1000.0 0.07978793620432727 43 COBYQA 1000.0 0.00020598371863355666 43 SLSQP 395.0 0.002744641508948131 43 Powell 565.0 0.00026251356957190366 43 nelder-mead 326.0 0.005665898451296298 43 L-BFGS-B 371.0 0.0021694202909624477 43 BFGS 3715.0 8.540661203993601e-07 43 trust-constr 4347.0 2.5085999407170378e-05 44 COBYLA 1000.0 0.08233794111808837 44 COBYQA 1000.0 0.0006650365315197859 44 SLSQP 535.0 0.0032067490009361457 44 Powell 816.0 0.00026012887967490284 44 nelder-mead 723.0 8.272905977631895e-05 44 L-BFGS-B 126.0 0.027542394225624045 44 BFGS 6159.0 2.2593468907395e-06 44 trust-constr 4968.0 2.6883195826390666e-05 45 COBYLA 149.0 0.02029939321743818 45 COBYQA 42.0 0.00015000751307292886 45 SLSQP 84.0 6.360763241040102e-07 45 Powell 227.0 8.949262200449805e-22 45 nelder-mead 375.0 4.867800022545828e-05 45 L-BFGS-B 42.0 2.400899932123255e-10 45 BFGS 174.0 5.3673974914275e-10 45 trust-constr 78.0 2.454090575098533e-06 46 COBYLA 1000.0 1567939963474.8965 46 COBYQA 106.0 0.0007293992928959043 46 SLSQP 84.0 9.996680902431176e-10 46 Powell 278.0 1.1773407061309293e-20 46 nelder-mead 465.0 8.181505338496236e-05 46 L-BFGS-B 66.0 2.7791623937188346e-10 46 BFGS 210.0 5.368902177156296e-10 46 trust-constr 114.0 2.9953810817563102e-06 47 COBYLA 203.0 0.01914426426703756 47 COBYQA 55.0 5.4087818719220104e-05 47 SLSQP 156.0 8.591197491407533e-10 47 Powell 362.0 3.554425222840407e-20 47 nelder-mead 853.0 0.00010539806608652054 47 L-BFGS-B 63.0 1.7100578981890751e-09 47 BFGS 396.0 8.554467047410975e-10 47 trust-constr 81.0 4.0417068746088015e-06 48 COBYLA 279.0 0.0331710012331183 48 COBYQA 87.0 0.005073729292213983 48 SLSQP 139.0 0.003427873430343299 48 Powell 450.0 8.053464532704803e-20 48 nelder-mead 1117.0 0.00021983838347792528 48 L-BFGS-B 88.0 4.803599008795718e-10 48 BFGS 660.0 1.0683513286135924e-09 48 trust-constr 99.0 4.5226625371459055e-06 49 COBYLA 369.0 0.03428531322541776 49 COBYQA 95.0 0.0036452448361088343 49 SLSQP 304.0 2.269282789883771e-06 49 Powell 540.0 3.020724955433762e-19 49 nelder-mead 1411.0 0.00030925186830669545 49 L-BFGS-B 104.0 6.510760200351575e-10 49 BFGS 741.0 1.2849241618216414e-09 49 trust-constr 208.0 4.879304344714801e-06 50 COBYLA 1000.0 4476072203419.262 50 COBYQA 205.0 0.010356853034351174 50 SLSQP 174.0 16090.611615482758 50 Powell 562.0 2.8587659884100126e-19 50 nelder-mead 2400.0 9.64238414027846 50 L-BFGS-B 156.0 1.885624896893773e-09 50 BFGS 845.0 1.2851588507030082e-09 50 trust-constr 273.0 4.8786393752066874e-06 51 COBYLA 1000.0 0.045532842697210224 51 COBYQA 1000.0 0.0025772034067144288 51 SLSQP 321.0 0.003924596822050221 51 Powell 4594.0 1.0674322967281136e-24 51 nelder-mead 1600.0 0.001974655275796391 51 L-BFGS-B 162.0 0.018002244158092647 51 BFGS 540.0 2.5296977794668327e-10 51 trust-constr 837.0 5.2718565480438994e-08 52 COBYLA 1000.0 894783.8530217626 52 COBYQA 848.0 4.5059640773037355 52 SLSQP 996.0 267.22972398525724 52 Powell 5894.0 12.202607426064228 52 nelder-mead 1600.0 223.19648568816407 52 L-BFGS-B 306.0 4.536639136942162 52 BFGS 669.0 4.5348798664538466 52 trust-constr 432.0 4.544389712796571 ============================== ============== =========== ======================== The results for BenchSmoothUnbounded: [ 6.82%] ··· ====================== ============== ==================== ======================== -- result type ------------------------------------- --------------------------------------------- test function solver mean_nfev mean_time ====================== ============== ==================== ======================== rosenbrock_slow COBYLA 1000.0 0.10221092700958252 rosenbrock_slow COBYQA 152.8 0.41478211879730226 rosenbrock_slow Powell 887.2 0.09182374477386475 rosenbrock_slow nelder-mead 310.2 0.03188121318817139 rosenbrock_slow L-BFGS-B 148.0 0.017220544815063476 rosenbrock_slow BFGS 177.2 0.02521369457244873 rosenbrock_slow CG n/a n/a rosenbrock_slow TNC n/a n/a rosenbrock_slow SLSQP 155.1 0.022461724281311036 rosenbrock_slow Newton-CG n/a n/a rosenbrock_slow dogleg n/a n/a rosenbrock_slow trust-ncg n/a n/a rosenbrock_slow trust-exact n/a n/a rosenbrock_slow trust-krylov n/a n/a rosenbrock_slow trust-constr 218.0 0.08552446365356445 rosenbrock_nograd COBYLA 1000.0 0.019648361206054687 rosenbrock_nograd COBYQA 152.8 0.39280545711517334 rosenbrock_nograd Powell 887.2 0.016094064712524413 rosenbrock_nograd nelder-mead 310.2 0.007036185264587403 rosenbrock_nograd L-BFGS-B 148.0 0.005884408950805664 rosenbrock_nograd BFGS 177.2 0.009394955635070801 rosenbrock_nograd CG n/a n/a rosenbrock_nograd TNC n/a n/a rosenbrock_nograd SLSQP 155.1 0.007205891609191895 rosenbrock_nograd Newton-CG n/a n/a rosenbrock_nograd dogleg n/a n/a rosenbrock_nograd trust-ncg n/a n/a rosenbrock_nograd trust-exact n/a n/a rosenbrock_nograd trust-krylov n/a n/a rosenbrock_nograd trust-constr 218.0 0.07210166454315185 rosenbrock COBYLA 1000.0 0.019016504287719727 rosenbrock COBYQA 152.8 0.3881392240524292 rosenbrock Powell 887.2 0.015932464599609376 rosenbrock nelder-mead 310.2 0.006581926345825195 rosenbrock L-BFGS-B 92.75 0.0037383198738098144 rosenbrock BFGS 110.95 0.006753122806549073 rosenbrock CG 110.0 0.008317852020263672 rosenbrock TNC 84.3 0.0030642032623291017 rosenbrock SLSQP 102.3 0.005086040496826172 rosenbrock Newton-CG 64.5 0.007260346412658691 rosenbrock dogleg 18.4 0.00196533203125 rosenbrock trust-ncg 43.5 0.0044361591339111325 rosenbrock trust-exact 17.5 0.0029834747314453126 rosenbrock trust-krylov 35.7 0.004609107971191406 rosenbrock trust-constr 98.16666666666667 0.04931776523590088 rosenbrock_tight COBYLA 1000.0 0.01909213066101074 rosenbrock_tight COBYQA 180.2 0.4743908166885376 rosenbrock_tight Powell 1070.1 0.019453907012939455 rosenbrock_tight nelder-mead 407.7 0.008971905708312989 rosenbrock_tight L-BFGS-B 99.2 0.0039985418319702145 rosenbrock_tight BFGS 160.9 0.009589815139770507 rosenbrock_tight CG 127.8 0.009342217445373535 rosenbrock_tight TNC 90.4 0.003626441955566406 rosenbrock_tight SLSQP 108.15 0.005452108383178711 rosenbrock_tight Newton-CG 71.6 0.00818331241607666 rosenbrock_tight dogleg 19.2 0.002088189125061035 rosenbrock_tight trust-ncg 44.9 0.004395413398742676 rosenbrock_tight trust-exact 18.4 0.0029824018478393556 rosenbrock_tight trust-krylov 93.7 0.008433985710144042 rosenbrock_tight trust-constr 106.53333333333333 0.052931682268778486 simple_quadratic COBYLA 54.8 0.0005712509155273438 simple_quadratic COBYQA 22.2 0.03898210525512695 simple_quadratic Powell 47.2 0.0005118846893310547 simple_quadratic nelder-mead 152.8 0.0022516965866088865 simple_quadratic L-BFGS-B 7.5 0.0003270387649536133 simple_quadratic BFGS 7.5 0.0003994584083557129 simple_quadratic CG 3.6 0.00027928352355957033 simple_quadratic TNC 7.0 0.00021626949310302735 simple_quadratic SLSQP 6.0 0.0003117680549621582 simple_quadratic Newton-CG 2.0 0.0002607107162475586 simple_quadratic dogleg 3.0 0.00029482841491699217 simple_quadratic trust-ncg 3.0 0.00023899078369140624 simple_quadratic trust-exact 3.0 0.00047221183776855467 simple_quadratic trust-krylov 3.0 0.0003118276596069336 simple_quadratic trust-constr 6.0 0.0033010323842366535 asymmetric_quadratic COBYLA 64.2 0.0006988286972045898 asymmetric_quadratic COBYQA 23.5 0.043817758560180664 asymmetric_quadratic Powell 134.6 0.0012090682983398437 asymmetric_quadratic nelder-mead 162.9 0.0022761344909667967 asymmetric_quadratic L-BFGS-B 7.5 0.00037877559661865235 asymmetric_quadratic BFGS 7.5 0.0003869175910949707 asymmetric_quadratic CG 3.1 0.00026428699493408203 asymmetric_quadratic TNC 12.5 0.0002912759780883789 asymmetric_quadratic SLSQP 6.0 0.00032651424407958984 asymmetric_quadratic Newton-CG 2.0 0.0002610206604003906 asymmetric_quadratic dogleg 2.9 0.00026187896728515627 asymmetric_quadratic trust-ncg 2.9 0.0002459287643432617 asymmetric_quadratic trust-exact 2.9 0.0004597187042236328 asymmetric_quadratic trust-krylov 2.9 0.0003061056137084961 asymmetric_quadratic trust-constr 5.966666666666667 0.0032143274943033854 sin_1d COBYLA 24.7 0.0003074169158935547 sin_1d COBYQA 13.7 0.02740190029144287 sin_1d Powell 28.1 0.00037937164306640626 sin_1d nelder-mead 36.8 0.00053558349609375 sin_1d L-BFGS-B 9.45 0.00045931339263916016 sin_1d BFGS 10.2 0.0007171988487243653 sin_1d CG 5.5 0.00034372806549072267 sin_1d TNC 10.2 0.0002444028854370117 sin_1d SLSQP 6.95 0.00047876834869384763 sin_1d Newton-CG n/a n/a sin_1d dogleg n/a n/a sin_1d trust-ncg n/a n/a sin_1d trust-exact n/a n/a sin_1d trust-krylov n/a n/a sin_1d trust-constr 9.6 0.006979584693908691 booth COBYLA 67.1 0.0008264064788818359 booth COBYQA 30.9 0.06615271568298339 booth Powell 63.4 0.0007641792297363281 booth nelder-mead 122.9 0.0021659135818481445 booth L-BFGS-B 11.6 0.0005254864692687989 booth BFGS 15.4 0.001001286506652832 booth CG 6.4 0.00041408538818359374 booth TNC 6.8 0.000226593017578125 booth SLSQP 12.3 0.0006320714950561523 booth Newton-CG n/a n/a booth dogleg n/a n/a booth trust-ncg n/a n/a booth trust-exact n/a n/a booth trust-krylov n/a n/a booth trust-constr 14.8 0.008171749114990235 beale COBYLA 461.1 0.004681205749511719 beale COBYQA 206.9 0.45821413993835447 beale Powell 1821.9 0.019892191886901854 beale nelder-mead 141.1 0.002144122123718262 beale L-BFGS-B 60.7 0.0021002888679504395 beale BFGS 311.6 0.018624258041381837 beale CG 32.9 0.0018250226974487304 beale TNC 24.1 0.0005386114120483398 beale SLSQP 25.9 0.0011149048805236816 beale Newton-CG n/a n/a beale dogleg n/a n/a beale trust-ncg n/a n/a beale trust-exact n/a n/a beale trust-krylov n/a n/a beale trust-constr 476.35 0.2655073642730713 LJ COBYLA 868.8 0.02984781265258789 LJ COBYQA 696.3 4.859420919418335 LJ Powell 1900.0 0.05867128372192383 LJ nelder-mead 2281.8 0.08925626277923585 LJ L-BFGS-B 278.55 0.011390721797943116 LJ BFGS 722.75 0.03499552011489868 LJ CG 146.9 0.01970562934875488 LJ TNC 115.6 0.011374115943908691 LJ SLSQP 208.5 0.00989387035369873 LJ Newton-CG n/a n/a LJ dogleg n/a n/a LJ trust-ncg n/a n/a LJ trust-exact n/a n/a LJ trust-krylov n/a n/a LJ trust-constr 1083.05 0.30552700757980344 ====================== ============== ==================== ======================== For the numbered problems, the objective function value from COBYQA is always better, and the histogram of function evaluation counts looks better. It looks like something went awry with the named problems. COBQA's objective value is always worse; it's almost like it's solving a different problem. It looks like something went awry with the named problems (BenchSmoothUnbounded)? COBQA's objective value is always worse. Or are the other algorithms really always better on smooth problems? Hi @mdhaber , if you are referring to the second table posted by Tom @ragonneau , you may note that the column is mean_time rather than min_obj. Ah. It's tough to compare the performance without knowing the objective function values. I can create an algorithm that runs pretty quickly but doesn't solve the problem : ) But I think the first set is sufficient. Agree. I do not know the SciPy benchmark, I did not run it, and I do not understand why the data in the two tables are different. Maybe Tom @ragonneau should check. @ragonneau I suggest you edit your results. They are difficult to read. In addition, there seems no reason to use a layout different from the other one. Thanks. @ragonneau I suggest you edit your results. They are difficult to read. In addition, there seems no reason to use a table different from the other one. Thanks. I believe the layout of the table was made by asv. When the last column has more than two subcolumns, it creates a column that is difficult to read, as in my previous comment. If an asv expert is in the thread, I'd be happy to learn how to improve that. In the meantime, I reran the experiment with only mean_nfev and mean_obj: [100.00%] ··· ====================== ============== ==================== ======================== -- result type ------------------------------------- --------------------------------------------- test function solver mean_nfev mean_obj ====================== ============== ==================== ======================== rosenbrock_slow COBYLA 1000.0 1.040669274800954 rosenbrock_slow COBYQA 152.8 3.890437153553753e-08 rosenbrock_slow Powell 887.2 1.577792702290088e-23 rosenbrock_slow nelder-mead 310.2 1.948083288418953e-09 rosenbrock_slow L-BFGS-B 148.0 2.395468489328531e-06 rosenbrock_slow BFGS 177.2 4.675037804518409e-11 rosenbrock_slow CG n/a n/a rosenbrock_slow TNC n/a n/a rosenbrock_slow SLSQP 155.1 1.0577238854081483e-05 rosenbrock_slow Newton-CG n/a n/a rosenbrock_slow dogleg n/a n/a rosenbrock_slow trust-ncg n/a n/a rosenbrock_slow trust-exact n/a n/a rosenbrock_slow trust-krylov n/a n/a rosenbrock_slow trust-constr 218.0 9.51616929722639e-10 rosenbrock_nograd COBYLA 1000.0 1.040669274800954 rosenbrock_nograd COBYQA 152.8 3.890437153553753e-08 rosenbrock_nograd Powell 887.2 1.577792702290088e-23 rosenbrock_nograd nelder-mead 310.2 1.948083288418953e-09 rosenbrock_nograd L-BFGS-B 148.0 2.395468489328531e-06 rosenbrock_nograd BFGS 177.2 4.675037804518409e-11 rosenbrock_nograd CG n/a n/a rosenbrock_nograd TNC n/a n/a rosenbrock_nograd SLSQP 155.1 1.0577238854081483e-05 rosenbrock_nograd Newton-CG n/a n/a rosenbrock_nograd dogleg n/a n/a rosenbrock_nograd trust-ncg n/a n/a rosenbrock_nograd trust-exact n/a n/a rosenbrock_nograd trust-krylov n/a n/a rosenbrock_nograd trust-constr 218.0 9.51616929722639e-10 rosenbrock COBYLA 1000.0 1.040669274800954 rosenbrock COBYQA 152.8 3.890437153553753e-08 rosenbrock Powell 887.2 1.577792702290088e-23 rosenbrock nelder-mead 310.2 1.948083288418953e-09 rosenbrock L-BFGS-B 92.75 2.1047384326635483e-06 rosenbrock BFGS 110.95 2.5838983525240928e-11 rosenbrock CG 110.0 2.184879891252899e-09 rosenbrock TNC 84.3 0.41353307449601323 rosenbrock SLSQP 102.3 1.2278847681581272e-05 rosenbrock Newton-CG 64.5 0.00028386319748745415 rosenbrock dogleg 18.4 1.0473748283949003e-12 rosenbrock trust-ncg 43.5 9.096654341078131e-10 rosenbrock trust-exact 17.5 2.3688425209734387e-11 rosenbrock trust-krylov 35.7 1.2080822443269492e-09 rosenbrock trust-constr 98.16666666666667 7.582424644998224e-10 rosenbrock_tight COBYLA 1000.0 1.0395001250506932 rosenbrock_tight COBYQA 180.2 1.0833609443328238e-15 rosenbrock_tight Powell 1070.1 5.649905225233614e-24 rosenbrock_tight nelder-mead 407.7 1.7872342256810196e-17 rosenbrock_tight L-BFGS-B 99.2 2.4807079357316176e-10 rosenbrock_tight BFGS 160.9 1.6841960135232725e-11 rosenbrock_tight CG 127.8 5.959714896391864e-19 rosenbrock_tight TNC 90.4 0.41043093540666387 rosenbrock_tight SLSQP 108.15 5.646132513964207e-10 rosenbrock_tight Newton-CG 71.6 8.645626438121326e-19 rosenbrock_tight dogleg 19.2 7.554726540980743e-22 rosenbrock_tight trust-ncg 44.9 6.432714849941195e-24 rosenbrock_tight trust-exact 18.4 9.805261192705719e-20 rosenbrock_tight trust-krylov 93.7 1.227821735143473e-17 rosenbrock_tight trust-constr 106.53333333333333 1.2124258427696534e-11 simple_quadratic COBYLA 54.8 9.518126940933508e-09 simple_quadratic COBYQA 22.2 8.872232632322827e-31 simple_quadratic Powell 47.2 1.2855981106468597e-31 simple_quadratic nelder-mead 152.8 2.2619242867800513e-09 simple_quadratic L-BFGS-B 7.5 2.4856928645318576e-16 simple_quadratic BFGS 7.5 1.6821868699593401e-15 simple_quadratic CG 3.6 3.154879136346819e-10 simple_quadratic TNC 7.0 2.0304535898501623e-07 simple_quadratic SLSQP 6.0 1.336048763662199e-16 simple_quadratic Newton-CG 2.0 0.0 simple_quadratic dogleg 3.0 2.78107340254829e-32 simple_quadratic trust-ncg 3.0 0.0 simple_quadratic trust-exact 3.0 2.76371887415049e-32 simple_quadratic trust-krylov 3.0 0.0 simple_quadratic trust-constr 6.0 9.386350782614601e-17 asymmetric_quadratic COBYLA 64.2 -0.24999998850473912 asymmetric_quadratic COBYQA 23.5 -0.25 asymmetric_quadratic Powell 134.6 -0.25 asymmetric_quadratic nelder-mead 162.9 -0.24999999810601786 asymmetric_quadratic L-BFGS-B 7.5 -0.24999999999999883 asymmetric_quadratic BFGS 7.5 -0.24999999999999784 asymmetric_quadratic CG 3.1 -0.25 asymmetric_quadratic TNC 12.5 -0.24999858212319187 asymmetric_quadratic SLSQP 6.0 -0.24999999999999972 asymmetric_quadratic Newton-CG 2.0 -0.25 asymmetric_quadratic dogleg 2.9 -0.25 asymmetric_quadratic trust-ncg 2.9 -0.25 asymmetric_quadratic trust-exact 2.9 -0.25 asymmetric_quadratic trust-krylov 2.9 -0.25 asymmetric_quadratic trust-constr 5.966666666666667 -0.24999999999999986 sin_1d COBYLA 24.7 -0.9999999947997484 sin_1d COBYQA 13.7 -0.9999999999766593 sin_1d Powell 28.1 -0.9999999999215394 sin_1d nelder-mead 36.8 -0.9999999998495948 sin_1d L-BFGS-B 9.45 -0.9999999936365207 sin_1d BFGS 10.2 -0.9999999996362824 sin_1d CG 5.5 -0.9999999999451038 sin_1d TNC 10.2 -0.9999999999816668 sin_1d SLSQP 6.95 -0.9999874938865606 sin_1d Newton-CG n/a n/a sin_1d dogleg n/a n/a sin_1d trust-ncg n/a n/a sin_1d trust-exact n/a n/a sin_1d trust-krylov n/a n/a sin_1d trust-constr 9.6 -0.9999999995302697 booth COBYLA 67.1 9.727602527729466e-08 booth COBYQA 30.9 1.1072140768523674e-09 booth Powell 63.4 2.1062586169401015e-29 booth nelder-mead 122.9 1.219380568865664 booth L-BFGS-B 11.6 2.1341422667527374e-08 booth BFGS 15.4 1.4028567484760426e-10 booth CG 6.4 8.662008283688798e-26 booth TNC 6.8 4.8330558571088694e-15 booth SLSQP 12.3 8.848768281252701e-08 booth Newton-CG n/a n/a booth dogleg n/a n/a booth trust-ncg n/a n/a booth trust-exact n/a n/a booth trust-krylov n/a n/a booth trust-constr 14.8 5.502755360622469e-11 beale COBYLA 461.1 1.0645878833664317 beale COBYQA 206.9 1.2864101661603227 beale Powell 1821.9 0.41411155100942254 beale nelder-mead 141.1 3.970084546666909e-10 beale L-BFGS-B 60.7 1.1948511188692428 beale BFGS 311.6 0.13604038140884195 beale CG 32.9 1.5004900363633769 beale TNC 24.1 0.01830610047951784 beale SLSQP 25.9 1.5104357514520017 beale Newton-CG n/a n/a beale dogleg n/a n/a beale trust-ncg n/a n/a beale trust-exact n/a n/a beale trust-krylov n/a n/a beale trust-constr 476.35 1.4994513515047054 LJ COBYLA 868.8 -4.290638261230069 LJ COBYQA 696.3 -5.999999907301164 LJ Powell 1900.0 -5.7224243446285765 LJ nelder-mead 2281.8 -3.8851133797531867 LJ L-BFGS-B 278.55 -3.2094944991218455 LJ BFGS 722.75 -5.999999999973245 LJ CG 146.9 -5.999999999936375 LJ TNC 115.6 -5.487077119375702 LJ SLSQP 208.5 -1.9506977348879857 LJ Newton-CG n/a n/a LJ dogleg n/a n/a LJ trust-ncg n/a n/a LJ trust-exact n/a n/a LJ trust-krylov n/a n/a LJ trust-constr 1083.05 -4.104444644769407 ====================== ============== ==================== ======================== Based on the benchmarks, I would consider COBYQA a great improvement over COBYLA and for our derivative-free solvers in general. As it is implemented in Python, it should not expose extraordinary maintenance burden as well. On how to proceed: as the current trend seems to be git submodules, let's try that here as well? Based on the benchmarks, I would consider COBYQA a great improvement over COBYLA and for our derivative-free solvers in general. Also it should be noted that currently COBYLA is our only derivative-free local solver that works with constraints. As it is implemented in Python, COBYQA should not cause extraordinary maintenance burden as well. On how to proceed: as the current trend seems to be git submodules, l would advocate for that here as well. But let's wait for a few more opinions :). For the moment, we use setuptools to create the package. Would it be useful if we switch to meson? For the moment, we use setuptools to create the package. Would it be helpful if we switch to meson? It's pure Python, so it really doesn't matter. as the current trend seems to be git submodules, l would advocate for that here as well. It depends a bit on plans I'd say. E.g., is the standalone package expected to continue to be developed for years into the future? Or once it's stable in SciPy, that will become the main implementation? It depends a bit on plans I'd say. E.g., is the standalone package expected to continue to be developed for years into the future? Or once it's stable in SciPy, that will become the main implementation? The goal is to maintain cobyqa as a standalone package and to continue to develop it in the future, possibly to add new features, etc. I believe this implies adding it as a submodule in scipy/_lib? I tried on my side; my scipy/optimize/_cobyqa_py.py looks like this: import importlib.util import sys import numpy as np from ._optimize import _check_unknown_options __all__ = [] def _minimize_cobyqa(fun, x0, args=(), bounds=None, constraints=(), callback=None, disp=False, maxfev=1000, maxiter=1000, target=-np.inf, feasibility_tol=1e-8, radius_init=1.0, radius_final=1e-6, **unknown_options): """ Minimize a scalar function of one or more variables using the Constrained Optimization BY Quadratic Approximations (COBYQA) algorithm. Options ------- disp : bool Set to True to print information about the optimization procedure. maxfev : int Maximum number of function evaluations. maxiter : int Maximum number of iterations. target : float Target value for the objective function. The optimization procedure is terminated when the objective function value of a nearly feasible point is less than or equal to this target. feasibility_tol : float Tolerance for the constraint violation. radius_init : float Initial trust-region radius. Typically, this value should be in the order of one tenth of the greatest expected change to the variables. radius_final : float Final trust-region radius. It should indicate the accuracy required in the final values of the variables. """ # Import the cobyqa module. cobyqa_spec = importlib.util.spec_from_file_location( 'cobyqa', '../_lib/cobyqa/cobyqa/__init__.py') cobyqa_module = importlib.util.module_from_spec(cobyqa_spec) sys.modules['cobyqa'] = cobyqa_module cobyqa_spec.loader.exec_module(cobyqa_module) # Run the COBYQA method. _check_unknown_options(unknown_options) options = { 'disp': bool(disp), 'maxfev': int(maxfev), 'maxiter': int(maxiter), 'target': float(target), 'feasibility_tol': float(feasibility_tol), 'radius_init': float(radius_init), 'radius_final': float(radius_final), } return cobyqa_module.minimize(fun, x0, args, bounds, constraints, callback, options) The problem: I am unsure how to modify meson.build so that the Python files in scipy/_lib/cobyqa/cobyqa are included. There should be a cleaner way. +1 to what Robert said on using relative imports and avoiding importlib. That's not even much of a compromise I'd say; relative imports within a package are idiomatic. The problem: I am unsure how to modify meson.build so that the Python files in scipy/_lib/cobyqa/cobyqa are included. I think it only requires something like this at the bottom of scipy/_lib/meson.build: py3.install_sources( [ 'cobyqa/__init__.py', 'cobyqa/cobyqa.py', ], subdir: 'scipy/_lib/cobyqa', ) If there's some kind of hiccup, then I'm happy to help on a PR or branch, please feel free to ping me. +1 to what Robert said on using relative imports and avoiding importlib. That's not even much of a compromise I'd say; relative imports within a package are idiomatic. I already pushed an update to the COBYQA repo, removing all the absolute imports. I think it only requires something like this at the bottom of scipy/_lib/meson.build: py3.install_sources( [ 'cobyqa/__init__.py', 'cobyqa/cobyqa.py', ], subdir: 'scipy/_lib/cobyqa', ) If there's some kind of hiccup, then I'm happy to help on a PR or branch, please feel free to ping me. Great! I will work next week on a PR, I will contact you if I face any trouble by then. Thanks 👍 I just created a PR. I would be happy to have some feedback and improve it if necessary. As a side note, on my challenging noisy function COBYQA showed systematically better convergence to global minimum vs COBYLA and all other optimization functions available in scipy. Looking forward to see it merged in! As a side note, on my challenging noisy function COBYQA showed systematically better convergence to global minimum vs COBYLA and all other optimization functions available in scipy. Looking forward to see it merged in! That's great to hear! I'm delighted this work is useful!
gharchive/issue
2024-01-19T14:54:04
2025-04-01T06:45:44.710652
{ "authors": [ "BarsMonster", "dschmitz89", "ev-br", "ilayn", "j-bowhay", "mdhaber", "ragonneau", "rgommers", "tupui", "zaikunzhang" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/19918", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2397965503
BUG: signal: crash in signaltools on free-threaded Python, macOS arm64 In the cron wheel builds on main, there was a failure today in the arm64 macOS 14.0 wheel build. The test suite run exited in signal/tests/test_signaltools.py with: 13615 Trace/BPT trap: 5 Full CI log and raw output (doesn't show anything more informative it looks like): ... 2024-07-09T10:14:17.6854550Z + /bin/sh -c 'bash /Users/runner/work/scipy/scipy/tools/wheels/cibw_test_command.sh /Users/runner/work/scipy/scipy' 2024-07-09T10:14:17.6901730Z ++ python '-cimport sysconfig; print(bool(sysconfig.get_config_var('\''Py_GIL_DISABLED'\'')))' 2024-07-09T10:14:17.7121210Z + FREE_THREADED_BUILD=True 2024-07-09T10:14:17.7121520Z + [[ True == \T\r\u\e ]] 2024-07-09T10:14:17.7122050Z + export PYTHON_GIL=0 2024-07-09T10:14:17.7122330Z + PYTHON_GIL=0 2024-07-09T10:14:17.7122800Z + python -c 'import sys; import scipy; sys.exit(not scipy.test())' 2024-07-09T10:14:25.2555010Z ============================= test session starts ============================== 2024-07-09T10:14:25.2556010Z platform darwin -- Python 3.13.0b2, pytest-8.2.2, pluggy-1.5.0 2024-07-09T10:14:25.2556740Z rootdir: /private/var/folders/h9/l1shxhdd69nct08ylzq0n6q00000gn/T/cibw-run-9m42f5g1/cp313t-macosx_arm64/test_cwd 2024-07-09T10:14:25.2557380Z plugins: hypothesis-6.105.1, cov-5.0.0, xdist-3.6.1 2024-07-09T10:14:25.2557760Z collected 64837 items / 11739 deselected / 53098 selected 2024-07-09T10:14:25.2558000Z 2024-07-09T10:14:25.2607920Z _lib/tests/test__gcutils.py ...... [ 0%] 2024-07-09T10:14:25.2682730Z _lib/tests/test__pep440.py ......... [ 0%] 2024-07-09T10:14:25.2700990Z _lib/tests/test__testutils.py .. [ 0%] 2024-07-09T10:14:25.7786640Z _lib/tests/test__threadsafety.py .. [ 0%] 2024-07-09T10:14:26.5096540Z _lib/tests/test__util.py ........................ [ 0%] 2024-07-09T10:14:26.5173400Z _lib/tests/test_array_api.py ssssssssssssss [ 0%] 2024-07-09T10:14:26.5402140Z _lib/tests/test_bunch.py ............................ [ 0%] 2024-07-09T10:14:26.9057650Z _lib/tests/test_ccallback.py ss.. [ 0%] 2024-07-09T10:14:26.9207020Z _lib/tests/test_deprecation.py . [ 0%] 2024-07-09T10:14:27.0253750Z _lib/tests/test_public_api.py .......................................... [ 0%] 2024-07-09T10:14:27.1161830Z ........................................................... [ 0%] 2024-07-09T10:14:27.1171160Z _lib/tests/test_scipy_version.py . [ 0%] 2024-07-09T10:14:27.1208080Z _lib/tests/test_tmpdirs.py ... [ 0%] 2024-07-09T10:14:27.2026800Z cluster/tests/test_disjoint_set.py ..................... [ 0%] 2024-07-09T10:14:27.2528020Z cluster/tests/test_hierarchy.py ........................................ [ 0%] 2024-07-09T10:14:27.3202600Z .........................................sss........... [ 0%] 2024-07-09T10:14:27.7931330Z cluster/tests/test_vq.py ............................... [ 0%] 2024-07-09T10:14:27.7990700Z constants/tests/test_codata.py ....... [ 0%] 2024-07-09T10:14:27.8054610Z constants/tests/test_constants.py ....... [ 0%] 2024-07-09T10:14:28.4682470Z datasets/tests/test_data.py ..... [ 0%] 2024-07-09T10:14:28.5271930Z fft/_pocketfft/tests/test_basic.py ..................................... [ 0%] 2024-07-09T10:14:28.6184940Z ........................................................................ [ 0%] 2024-07-09T10:14:28.6871320Z ........................................................................ [ 1%] 2024-07-09T10:14:28.7571510Z ........................................................................ [ 1%] 2024-07-09T10:14:28.8555130Z ........................................................................ [ 1%] 2024-07-09T10:14:29.1293840Z ........................................................................ [ 1%] 2024-07-09T10:14:29.5032370Z .............................................................. [ 1%] 2024-07-09T10:14:29.5362690Z fft/_pocketfft/tests/test_real_transforms.py ........................... [ 1%] 2024-07-09T10:14:29.6031280Z ........................................................................ [ 1%] 2024-07-09T10:14:29.6720620Z ........................................................................ [ 1%] 2024-07-09T10:14:29.7397130Z ........................................................................ [ 2%] 2024-07-09T10:14:29.8047270Z ........................................................................ [ 2%] 2024-07-09T10:14:30.1155580Z ........................................................................ [ 2%] 2024-07-09T10:14:30.3840320Z ........................................................................ [ 2%] 2024-07-09T10:14:30.5076490Z ........................................................................ [ 2%] 2024-07-09T10:14:30.5846320Z ........................................................................ [ 2%] 2024-07-09T10:14:30.6565160Z ........................................................................ [ 2%] 2024-07-09T10:14:30.7249790Z ........................................................................ [ 2%] 2024-07-09T10:14:30.8180390Z ........................................................................ [ 3%] 2024-07-09T10:14:30.8948810Z ........................................................................ [ 3%] 2024-07-09T10:14:31.2646790Z ........................................................................ [ 3%] 2024-07-09T10:14:31.5765810Z ........................................................................ [ 3%] 2024-07-09T10:14:31.6457940Z ........................................................................ [ 3%] 2024-07-09T10:14:31.7139080Z ........................................................................ [ 3%] 2024-07-09T10:14:31.7755250Z ........................................................................ [ 3%] 2024-07-09T10:14:31.8335770Z ........................................................................ [ 4%] 2024-07-09T10:14:31.8937290Z ........................................................................ [ 4%] 2024-07-09T10:14:31.9552580Z ........................................................................ [ 4%] 2024-07-09T10:14:32.0153930Z ........................................................................ [ 4%] 2024-07-09T10:14:32.0742110Z ........................................................................ [ 4%] 2024-07-09T10:14:32.1342620Z ........................................................................ [ 4%] 2024-07-09T10:14:32.1948590Z ........................................................................ [ 4%] 2024-07-09T10:14:32.2564340Z ........................................................................ [ 4%] 2024-07-09T10:14:32.3222580Z ........................................................................ [ 5%] 2024-07-09T10:14:32.3705760Z ................................................. [ 5%] 2024-07-09T10:14:32.4125240Z fft/tests/test_backend.py .............................................. [ 5%] 2024-07-09T10:14:32.4125710Z [ 5%] 2024-07-09T10:14:32.5495700Z fft/tests/test_basic.py ................................................ [ 5%] 2024-07-09T10:14:32.7200660Z ........................................................................ [ 5%] 2024-07-09T10:14:33.7593160Z ................... [ 5%] 2024-07-09T10:14:33.8039110Z fft/tests/test_fftlog.py ........................................ [ 5%] 2024-07-09T10:14:33.8371100Z fft/tests/test_helper.py ............................... [ 5%] 2024-07-09T10:14:33.9503940Z fft/tests/test_multithreading.py ....................................... [ 5%] 2024-07-09T10:14:33.9837950Z ................ [ 5%] 2024-07-09T10:14:34.0240480Z fft/tests/test_real_transforms.py ...................................... [ 5%] 2024-07-09T10:14:34.0973000Z ........................................................................ [ 6%] 2024-07-09T10:14:34.1705940Z ........................................................................ [ 6%] 2024-07-09T10:14:34.2434150Z ........................................................................ [ 6%] 2024-07-09T10:14:34.3157850Z ........................................................................ [ 6%] 2024-07-09T10:14:34.3873810Z ........................................................................ [ 6%] 2024-07-09T10:14:34.4592810Z ........................................................................ [ 6%] 2024-07-09T10:14:34.5306500Z ........................................................................ [ 6%] 2024-07-09T10:14:34.6022770Z ........................................................................ [ 6%] 2024-07-09T10:14:34.6741650Z ........................................................................ [ 7%] 2024-07-09T10:14:34.7459530Z ........................................................................ [ 7%] 2024-07-09T10:14:34.8142910Z ........................................................................ [ 7%] 2024-07-09T10:14:34.8838600Z ........................................................................ [ 7%] 2024-07-09T10:14:34.9501190Z ........................................................................ [ 7%] 2024-07-09T10:14:35.0187220Z ........................................................................ [ 7%] 2024-07-09T10:14:35.0876930Z ........................................................................ [ 7%] 2024-07-09T10:14:35.1595300Z ........................................................................ [ 8%] 2024-07-09T10:14:35.2356300Z ........................................................................ [ 8%] 2024-07-09T10:14:35.3065520Z ........................................................................ [ 8%] 2024-07-09T10:14:35.3771920Z ........................................................................ [ 8%] 2024-07-09T10:14:35.4550810Z ........................................................................ [ 8%] 2024-07-09T10:14:35.5306390Z ........................................................................ [ 8%] 2024-07-09T10:14:35.6093410Z ........................................................................ [ 8%] 2024-07-09T10:14:35.6897450Z ........................................................................ [ 8%] 2024-07-09T10:14:35.7699340Z ........................................................................ [ 9%] 2024-07-09T10:14:35.8533310Z ........................................................................ [ 9%] 2024-07-09T10:14:35.9350320Z ........................................................................ [ 9%] 2024-07-09T10:14:36.0177350Z ........................................................................ [ 9%] 2024-07-09T10:14:36.0906920Z ........................................................................ [ 9%] 2024-07-09T10:14:36.1640020Z ........................................................................ [ 9%] 2024-07-09T10:14:36.2412880Z ........................................................................ [ 9%] 2024-07-09T10:14:36.3201230Z ........................................................................ [ 10%] 2024-07-09T10:14:36.3957790Z ........................................................................ [ 10%] 2024-07-09T10:14:36.4712290Z ........................................................................ [ 10%] 2024-07-09T10:14:36.5504580Z ........................................................................ [ 10%] 2024-07-09T10:14:36.6275450Z ........................................................................ [ 10%] 2024-07-09T10:14:36.6996650Z ........................................................................ [ 10%] 2024-07-09T10:14:36.7691050Z ........................................................................ [ 10%] 2024-07-09T10:14:36.8371680Z ........................................................................ [ 11%] 2024-07-09T10:14:36.9076270Z ........................................................................ [ 11%] 2024-07-09T10:14:36.9742240Z ........................................................................ [ 11%] 2024-07-09T10:14:37.0389130Z ........................................................................ [ 11%] 2024-07-09T10:14:37.0908880Z .......................................................... [ 11%] 2024-07-09T10:14:37.2353020Z fftpack/tests/test_basic.py ......x..................................... [ 11%] 2024-07-09T10:14:37.3292180Z ........................................................................ [ 11%] 2024-07-09T10:14:37.4028710Z ........................................................................ [ 11%] 2024-07-09T10:14:37.6076430Z ........................................................................ [ 12%] 2024-07-09T10:14:37.8148540Z .................................... [ 12%] 2024-07-09T10:14:37.8188790Z fftpack/tests/test_helper.py .... [ 12%] 2024-07-09T10:14:37.8864940Z fftpack/tests/test_pseudo_diffs.py ............................. [ 12%] 2024-07-09T10:14:38.1308650Z fftpack/tests/test_real_transforms.py .................................. [ 12%] 2024-07-09T10:14:38.8570470Z ........................................................................ [ 12%] 2024-07-09T10:14:38.9197940Z ........................................................................ [ 12%] 2024-07-09T10:14:38.9647940Z .............................................. [ 12%] 2024-07-09T10:14:39.5883850Z integrate/_ivp/tests/test_ivp.py ....................................... [ 12%] 2024-07-09T10:14:39.7089890Z ............... [ 12%] 2024-07-09T10:14:39.7182320Z integrate/_ivp/tests/test_rk.py .......... [ 12%] 2024-07-09T10:14:41.5625470Z integrate/tests/test__quad_vec.py .......................... [ 12%] 2024-07-09T10:14:41.6248260Z integrate/tests/test_banded_ode_solvers.py . [ 12%] 2024-07-09T10:14:41.7764800Z integrate/tests/test_bvp.py ................ [ 12%] 2024-07-09T10:14:41.8280720Z integrate/tests/test_integrate.py ...................................... [ 12%] 2024-07-09T10:14:41.8478350Z ......... [ 12%] 2024-07-09T10:14:41.8761530Z integrate/tests/test_odeint_jac.py .. [ 12%] 2024-07-09T10:14:42.0584870Z integrate/tests/test_quadpack.py ....................................... [ 12%] 2024-07-09T10:14:43.0279540Z ..sssssssssssssssssssssssssssssssssssssssssss........... [ 13%] 2024-07-09T10:14:43.1334810Z integrate/tests/test_quadrature.py ..................................... [ 13%] 2024-07-09T10:14:43.2340040Z ........................................................................ [ 13%] 2024-07-09T10:14:43.3278030Z ........................................................................ [ 13%] 2024-07-09T10:14:43.4222890Z ........................................................................ [ 13%] 2024-07-09T10:14:43.4286090Z ....... [ 13%] 2024-07-09T10:14:43.5115680Z integrate/tests/test_tanhsinh.py ....................................... [ 13%] 2024-07-09T10:14:43.6967280Z .........ssss........................................................... [ 13%] 2024-07-09T10:14:44.0721280Z .............................ss..ss......ss............................. [ 13%] 2024-07-09T10:14:44.2780360Z .................................ss..................................... [ 14%] 2024-07-09T10:14:44.4808950Z ........................................ [ 14%] 2024-07-09T10:14:44.5441160Z interpolate/tests/test_bsplines.py ..................................... [ 14%] 2024-07-09T10:14:44.6400190Z ........................................................................ [ 14%] 2024-07-09T10:14:45.0000690Z .......................x................................................ [ 14%] 2024-07-09T10:14:45.0590470Z .............................. [ 14%] 2024-07-09T10:14:45.1476160Z interpolate/tests/test_fitpack.py .....................................s [ 14%] 2024-07-09T10:14:45.1579000Z ....s.. [ 14%] 2024-07-09T10:14:45.3876110Z interpolate/tests/test_fitpack2.py ..................................... [ 14%] 2024-07-09T10:14:45.5359100Z ................................ [ 14%] 2024-07-09T10:14:45.5366470Z interpolate/tests/test_gil.py s [ 14%] 2024-07-09T10:14:45.6146460Z interpolate/tests/test_interpnd.py ..................... [ 14%] 2024-07-09T10:14:45.7301320Z interpolate/tests/test_interpolate.py .................................. [ 14%] 2024-07-09T10:14:46.0830570Z ........................................................................ [ 14%] 2024-07-09T10:14:46.4576170Z ........................ [ 14%] 2024-07-09T10:14:46.4888780Z interpolate/tests/test_ndgriddata.py .................... [ 15%] 2024-07-09T10:14:46.4956010Z interpolate/tests/test_pade.py .... [ 15%] 2024-07-09T10:14:46.7352530Z interpolate/tests/test_polyint.py ...................................... [ 15%] 2024-07-09T10:14:46.8566020Z ........................................................................ [ 15%] 2024-07-09T10:14:46.8740440Z ................. [ 15%] 2024-07-09T10:14:46.8957080Z interpolate/tests/test_rbf.py .......... [ 15%] 2024-07-09T10:14:47.0738270Z interpolate/tests/test_rbfinterp.py .................................... [ 15%] 2024-07-09T10:14:47.2650430Z ........................................................................ [ 15%] 2024-07-09T10:14:47.5113650Z ........................................................................ [ 15%] 2024-07-09T10:14:47.5241690Z ..... [ 15%] 2024-07-09T10:14:48.0257510Z interpolate/tests/test_rgi.py ....................s..................... [ 15%] 2024-07-09T10:14:49.7347350Z ..................................ss.......ss..................s.......s [ 15%] 2024-07-09T10:14:50.5596830Z s.......ss............................................s.......ss........ [ 15%] 2024-07-09T10:14:50.6196080Z ..s......................... [ 16%] 2024-07-09T10:14:50.6303250Z io/_harwell_boeing/tests/test_fortran_format.py ........... [ 16%] 2024-07-09T10:14:50.6412170Z io/_harwell_boeing/tests/test_hb.py .. [ 16%] 2024-07-09T10:14:50.8372350Z io/arff/tests/test_arffread.py ........................... [ 16%] 2024-07-09T10:14:50.8392440Z io/matlab/tests/test_byteordercodes.py .. [ 16%] 2024-07-09T10:14:50.9405920Z io/matlab/tests/test_mio.py ............................................ [ 16%] 2024-07-09T10:14:50.9769900Z ............................ [ 16%] 2024-07-09T10:14:50.9829330Z io/matlab/tests/test_mio5_utils.py ...... [ 16%] 2024-07-09T10:14:50.9840530Z io/matlab/tests/test_mio_funcs.py . [ 16%] 2024-07-09T10:14:50.9857970Z io/matlab/tests/test_mio_utils.py .. [ 16%] 2024-07-09T10:14:50.9872670Z io/matlab/tests/test_miobase.py . [ 16%] 2024-07-09T10:14:50.9899830Z io/matlab/tests/test_pathological.py .. [ 16%] 2024-07-09T10:14:51.6076130Z io/matlab/tests/test_streams.py ........... [ 16%] 2024-07-09T10:14:51.6357840Z io/tests/test_fortran.py ........... [ 16%] 2024-07-09T10:14:51.6975860Z io/tests/test_idl.py ................................................... [ 16%] 2024-07-09T10:14:51.7507160Z ................ [ 16%] 2024-07-09T10:14:51.8459390Z io/tests/test_mmio.py .................................................. [ 16%] 2024-07-09T10:14:52.0525170Z ........................................................................ [ 16%] 2024-07-09T10:14:52.1781320Z .............................. [ 16%] 2024-07-09T10:14:52.2297610Z io/tests/test_netcdf.py ......................... [ 16%] 2024-07-09T10:14:52.2471830Z io/tests/test_paths.py ........... [ 16%] 2024-07-09T10:14:52.3103000Z io/tests/test_wavfile.py ................s.............................. [ 16%] 2024-07-09T10:14:52.4051870Z ........................................................................ [ 17%] 2024-07-09T10:14:52.5090840Z ........................................................................ [ 17%] 2024-07-09T10:14:52.6753370Z ........................................................................ [ 17%] 2024-07-09T10:14:52.7456120Z .......................... [ 17%] 2024-07-09T10:14:52.7938530Z linalg/tests/test_basic.py ............................................. [ 17%] 2024-07-09T10:14:52.8787530Z ................................................................s....... [ 17%] 2024-07-09T10:14:52.9472940Z ........................................................................ [ 17%] 2024-07-09T10:14:53.4837720Z ........................................................................ [ 17%] 2024-07-09T10:14:54.3404830Z ........................................................................ [ 17%] 2024-07-09T10:14:54.4335140Z ........................................s............................... [ 18%] 2024-07-09T10:14:54.4576590Z ........................... [ 18%] 2024-07-09T10:14:54.5152150Z linalg/tests/test_blas.py .............................................. [ 18%] 2024-07-09T10:14:54.5625230Z .....x. [ 18%] 2024-07-09T10:14:54.5705570Z linalg/tests/test_cython_blas.py ...... [ 18%] 2024-07-09T10:14:54.5738780Z linalg/tests/test_cython_lapack.py ... [ 18%] 2024-07-09T10:14:54.5977860Z linalg/tests/test_cythonized_array_utils.py ............................ [ 18%] 2024-07-09T10:14:54.6176280Z ...................... [ 18%] 2024-07-09T10:14:54.7476860Z linalg/tests/test_decomp.py ............................................ [ 18%] 2024-07-09T10:14:54.8324210Z ........................................................................ [ 18%] 2024-07-09T10:14:54.9226810Z ........................................................................ [ 18%] 2024-07-09T10:14:58.3365110Z ........................................................................ [ 18%] 2024-07-09T10:14:58.4764580Z ........................................................................ [ 19%] 2024-07-09T10:14:58.5902640Z ........................................................................ [ 19%] 2024-07-09T10:14:59.3769020Z ..............................................x......................... [ 19%] 2024-07-09T10:14:59.4114420Z ................ [ 19%] 2024-07-09T10:14:59.4436090Z linalg/tests/test_decomp_cholesky.py .....s............................. [ 19%] 2024-07-09T10:14:59.4787890Z .................................... [ 19%] 2024-07-09T10:14:59.5673790Z linalg/tests/test_decomp_cossin.py ..................................... [ 19%] 2024-07-09T10:14:59.7591640Z ...................................................... [ 19%] 2024-07-09T10:14:59.8045200Z linalg/tests/test_decomp_ldl.py ............ [ 19%] 2024-07-09T10:14:59.8499530Z linalg/tests/test_decomp_lu.py ......................................... [ 19%] 2024-07-09T10:14:59.8973490Z ................................................... [ 19%] 2024-07-09T10:14:59.9048490Z linalg/tests/test_decomp_polar.py .. [ 19%] 2024-07-09T10:14:59.9945930Z linalg/tests/test_decomp_update.py ..................................... [ 19%] 2024-07-09T10:15:00.1746640Z ........................................................................ [ 20%] 2024-07-09T10:15:00.3120800Z ........................................................................ [ 20%] 2024-07-09T10:15:00.4651120Z ........................................................................ [ 20%] 2024-07-09T10:15:00.6254490Z ........................................................................ [ 20%] 2024-07-09T10:15:00.8040730Z ........................................................................ [ 20%] 2024-07-09T10:15:00.9568510Z ........................................................................ [ 20%] 2024-07-09T10:15:01.1552150Z ........................................................................ [ 20%] 2024-07-09T10:15:01.3566930Z ........................................................................ [ 20%] 2024-07-09T10:15:01.4763820Z .................................................... [ 21%] 2024-07-09T10:15:06.5056140Z linalg/tests/test_extending.py . [ 21%] 2024-07-09T10:15:06.5427880Z linalg/tests/test_fblas.py ............................................. [ 21%] 2024-07-09T10:15:06.6780290Z ........................................................................ [ 21%] 2024-07-09T10:15:06.6882240Z ............ [ 21%] 2024-07-09T10:15:07.5339200Z linalg/tests/test_interpolative.py ..................................... [ 21%] 2024-07-09T10:15:07.5444960Z .......... [ 21%] 2024-07-09T10:15:07.5941610Z linalg/tests/test_lapack.py ............................................ [ 21%] 2024-07-09T10:15:07.6923310Z ....................................................................s... [ 21%] 2024-07-09T10:15:07.7727540Z ........................................................................ [ 21%] 2024-07-09T10:15:07.8527350Z ........................................................................ [ 21%] 2024-07-09T10:15:07.9207110Z ........................................................................ [ 22%] 2024-07-09T10:15:07.9951510Z ........................................................................ [ 22%] 2024-07-09T10:15:08.0691010Z ........................................................................ [ 22%] 2024-07-09T10:15:08.1576350Z ........................................................................ [ 22%] 2024-07-09T10:15:08.2409530Z ........................................................................ [ 22%] 2024-07-09T10:15:08.3137500Z ........................................................................ [ 22%] 2024-07-09T10:15:08.3907690Z ........................................................................ [ 22%] 2024-07-09T10:15:08.4672670Z ........................................................................ [ 22%] 2024-07-09T10:15:08.5464580Z ........................................................................ [ 23%] 2024-07-09T10:15:08.6297740Z ........................................................................ [ 23%] 2024-07-09T10:15:08.7047530Z ........................................................................ [ 23%] 2024-07-09T10:15:08.7771680Z ........................................................................ [ 23%] 2024-07-09T10:15:08.8509100Z ........................................................................ [ 23%] 2024-07-09T10:15:08.9244920Z ........................................................................ [ 23%] 2024-07-09T10:15:09.0135440Z ........................................................................ [ 23%] 2024-07-09T10:15:09.0994090Z ........................................................................ [ 24%] 2024-07-09T10:15:09.1688880Z ........................................................................ [ 24%] 2024-07-09T10:15:09.2433480Z ........................................................................ [ 24%] 2024-07-09T10:15:09.3176420Z ........................................................................ [ 24%] 2024-07-09T10:15:09.3871150Z ........................................................................ [ 24%] 2024-07-09T10:15:09.4503830Z ........................................................................ [ 24%] 2024-07-09T10:15:09.6009720Z ........................................................................ [ 24%] 2024-07-09T10:15:09.6778920Z ........................................................................ [ 25%] 2024-07-09T10:15:09.6901010Z ........... [ 25%] 2024-07-09T10:15:10.1008090Z linalg/tests/test_matfuncs.py ................xxxxx..............x...... [ 25%] 2024-07-09T10:15:10.7697250Z .............X........................ [ 25%] 2024-07-09T10:15:10.7775130Z linalg/tests/test_matmul_toeplitz.py .... [ 25%] 2024-07-09T10:15:10.7912770Z linalg/tests/test_procrustes.py ........... [ 25%] 2024-07-09T10:15:11.0652420Z linalg/tests/test_sketches.py ..... [ 25%] 2024-07-09T10:15:11.0950730Z linalg/tests/test_solve_toeplitz.py ......X......................... [ 25%] 2024-07-09T10:15:11.3559020Z linalg/tests/test_solvers.py ........x....x....x.xx..................... [ 25%] 2024-07-09T10:15:11.3629790Z .... [ 25%] 2024-07-09T10:15:11.4569850Z linalg/tests/test_special_matrices.py .................................. [ 25%] 2024-07-09T10:15:11.5297410Z ........................................................................ [ 25%] 2024-07-09T10:15:11.5357540Z ...... [ 25%] 2024-07-09T10:15:11.6376640Z misc/tests/test_common.py ... [ 25%] 2024-07-09T10:15:11.6426770Z misc/tests/test_config.py .... [ 25%] 2024-07-09T10:15:11.6471870Z misc/tests/test_doccer.py ..... [ 25%] 2024-07-09T10:15:11.6554370Z ndimage/tests/test_c_api.py ... [ 25%] 2024-07-09T10:15:11.7703620Z ndimage/tests/test_datatypes.py .x [ 25%] 2024-07-09T10:15:11.8099170Z ndimage/tests/test_filters.py .......................................... [ 25%] 2024-07-09T10:15:11.8746060Z ........................................................................ [ 25%] 2024-07-09T10:15:11.9408870Z ........................................................................ [ 25%] 2024-07-09T10:15:12.0060500Z ........................................................................ [ 26%] 2024-07-09T10:15:12.0687140Z ........................................................................ [ 26%] 2024-07-09T10:15:12.1318650Z ........................................................................ [ 26%] 2024-07-09T10:15:12.1954470Z ........................................................................ [ 26%] 2024-07-09T10:15:12.2615740Z ........................................................................ [ 26%] 2024-07-09T10:15:12.3250940Z ........................................................................ [ 26%] 2024-07-09T10:15:12.3928960Z ........................................................................ [ 26%] 2024-07-09T10:15:12.4606520Z ........................................................................ [ 27%] 2024-07-09T10:15:12.5452470Z ........................................................................ [ 27%] 2024-07-09T10:15:12.6220390Z ........................................................................ [ 27%] 2024-07-09T10:15:12.7023650Z ........................................................................ [ 27%] 2024-07-09T10:15:12.7842420Z ........................................................................ [ 27%] 2024-07-09T10:15:12.8742090Z ........................................................................ [ 27%] 2024-07-09T10:15:12.9463760Z ........................................................................ [ 27%] 2024-07-09T10:15:13.0147990Z ........................................................................ [ 27%] 2024-07-09T10:15:13.0845290Z ........................................................................ [ 28%] 2024-07-09T10:15:13.1497550Z ........................................................................ [ 28%] 2024-07-09T10:15:13.2107280Z ........................................................................ [ 28%] 2024-07-09T10:15:13.2720800Z ........................................................................ [ 28%] 2024-07-09T10:15:13.3373750Z ........................................................................ [ 28%] 2024-07-09T10:15:13.4042240Z ........................................................................ [ 28%] 2024-07-09T10:15:13.4707980Z ........................................................................ [ 28%] 2024-07-09T10:15:13.5378050Z ........................................................................ [ 29%] 2024-07-09T10:15:13.6017690Z ........................................................................ [ 29%] 2024-07-09T10:15:14.0162760Z ......................................................... [ 29%] 2024-07-09T10:15:14.0562490Z ndimage/tests/test_fourier.py .......................................... [ 29%] 2024-07-09T10:15:14.0848060Z .................................. [ 29%] 2024-07-09T10:15:14.1189730Z ndimage/tests/test_interpolation.py .................................... [ 29%] 2024-07-09T10:15:14.1818590Z ........................................................................ [ 29%] 2024-07-09T10:15:14.2420070Z ........................................................................ [ 29%] 2024-07-09T10:15:14.3013030Z ........................................................................ [ 29%] 2024-07-09T10:15:14.3611000Z ........................................................................ [ 30%] 2024-07-09T10:15:14.4311920Z ........................................................................ [ 30%] 2024-07-09T10:15:14.5121480Z ........................................................................ [ 30%] 2024-07-09T10:15:14.5722300Z ........................................................................ [ 30%] 2024-07-09T10:15:14.6336240Z ........................................................................ [ 30%] 2024-07-09T10:15:14.6953700Z ........................................................................ [ 30%] 2024-07-09T10:15:14.7701420Z ........................................................................ [ 30%] 2024-07-09T10:15:14.8486430Z ........................................................................ [ 31%] 2024-07-09T10:15:14.9185810Z ........................................................................ [ 31%] 2024-07-09T10:15:14.9838040Z ........................................................................ [ 31%] 2024-07-09T10:15:14.9980120Z ................ [ 31%] 2024-07-09T10:15:15.1718370Z ndimage/tests/test_measurements.py ..................................... [ 31%] 2024-07-09T10:15:15.2533650Z ........................................................................ [ 31%] 2024-07-09T10:15:15.2672510Z ................ [ 31%] 2024-07-09T10:15:15.3084360Z ndimage/tests/test_morphology.py ....................................... [ 31%] 2024-07-09T10:15:15.3795670Z ........................................................................ [ 31%] 2024-07-09T10:15:15.4443610Z ........................................................................ [ 31%] 2024-07-09T10:15:15.5042860Z ........................................................................ [ 32%] 2024-07-09T10:15:15.5675510Z ........................................................................ [ 32%] 2024-07-09T10:15:15.6323520Z ........................................................................ [ 32%] 2024-07-09T10:15:15.6942970Z ........................................................................ [ 32%] 2024-07-09T10:15:15.7532250Z ........................................................................ [ 32%] 2024-07-09T10:15:15.8135430Z ........................................................................ [ 32%] 2024-07-09T10:15:15.8747790Z ........................................................................ [ 32%] 2024-07-09T10:15:15.9362540Z ........................................................................ [ 32%] 2024-07-09T10:15:15.9974380Z ........................................................................ [ 33%] 2024-07-09T10:15:16.0146030Z ................. [ 33%] 2024-07-09T10:15:16.0230610Z ndimage/tests/test_ni_support.py .......... [ 33%] 2024-07-09T10:15:16.0489620Z ndimage/tests/test_splines.py .................. [ 33%] 2024-07-09T10:15:16.0801220Z odr/tests/test_odr.py ................s.. [ 33%] 2024-07-09T10:15:16.0919140Z optimize/_trustregion_constr/tests/test_canonical_constraint.py ...... [ 33%] 2024-07-09T10:15:16.1229470Z optimize/_trustregion_constr/tests/test_projections.py .......... [ 33%] 2024-07-09T10:15:16.1436180Z optimize/_trustregion_constr/tests/test_qp_subproblem.py ............... [ 33%] 2024-07-09T10:15:16.1563350Z ........ [ 33%] 2024-07-09T10:15:16.1774570Z optimize/_trustregion_constr/tests/test_report.py .s [ 33%] 2024-07-09T10:15:18.1183910Z optimize/tests/test__basinhopping.py ................................. [ 33%] 2024-07-09T10:15:18.4623840Z optimize/tests/test__differential_evolution.py ......................... [ 33%] 2024-07-09T10:15:22.9791320Z .............................s....... [ 33%] 2024-07-09T10:15:25.3800790Z optimize/tests/test__dual_annealing.py ................................. [ 33%] 2024-07-09T10:15:26.8269010Z ............. [ 33%] 2024-07-09T10:15:26.8392760Z optimize/tests/test__linprog_clean_inputs.py ............. [ 33%] 2024-07-09T10:15:26.9077750Z optimize/tests/test__numdiff.py ................................ [ 33%] 2024-07-09T10:15:26.9533350Z optimize/tests/test__remove_redundancy.py .............................. [ 33%] 2024-07-09T10:15:27.0364490Z ...................................... [ 33%] 2024-07-09T10:15:27.0545480Z optimize/tests/test__root.py ............... [ 33%] 2024-07-09T10:15:28.4213460Z optimize/tests/test__shgo.py ..................ss...................s... [ 33%] 2024-07-09T10:15:29.7282670Z ............................ [ 33%] 2024-07-09T10:15:29.8347330Z optimize/tests/test__spectral.py .... [ 33%] 2024-07-09T10:15:29.8976010Z optimize/tests/test_bracket.py ......................................... [ 34%] 2024-07-09T10:15:30.0832880Z ........................................................................ [ 34%] 2024-07-09T10:15:30.1558930Z ........................................................................ [ 34%] 2024-07-09T10:15:30.2331280Z ........................................................................ [ 34%] 2024-07-09T10:15:30.3187730Z ........................................................................ [ 34%] 2024-07-09T10:15:30.6099050Z ........................................................................ [ 34%] 2024-07-09T10:15:30.7122890Z ................................................................... [ 34%] 2024-07-09T10:15:31.1437480Z optimize/tests/test_chandrupatla.py .................................... [ 34%] 2024-07-09T10:15:31.3160680Z ........................................................................ [ 35%] 2024-07-09T10:15:31.3628760Z ........................... [ 35%] 2024-07-09T10:15:31.3758290Z optimize/tests/test_cobyla.py s..... [ 35%] 2024-07-09T10:15:31.7728800Z optimize/tests/test_cobyqa.py ........ [ 35%] 2024-07-09T10:15:32.7723680Z optimize/tests/test_constraint_conversion.py ........ [ 35%] 2024-07-09T10:15:32.7888780Z optimize/tests/test_constraints.py ................ [ 35%] 2024-07-09T10:15:32.7930720Z optimize/tests/test_cython_optimize.py ..... [ 35%] 2024-07-09T10:15:32.8319130Z optimize/tests/test_differentiable_functions.py ..............x. [ 35%] 2024-07-09T10:15:32.9012880Z optimize/tests/test_differentiate.py ................................... [ 35%] 2024-07-09T10:15:33.2916880Z ........................................................................ [ 35%] 2024-07-09T10:15:33.4273250Z ...............................................XX................... [ 35%] 2024-07-09T10:15:33.5483090Z optimize/tests/test_direct.py ....................................ss.... [ 35%] 2024-07-09T10:15:33.5655220Z ..................... [ 35%] 2024-07-09T10:15:36.2213890Z optimize/tests/test_extending.py . [ 35%] 2024-07-09T10:15:36.2360910Z optimize/tests/test_hessian_update_strategy.py ..... [ 35%] 2024-07-09T10:15:36.2642970Z optimize/tests/test_isotonic_regression.py ............................. [ 35%] 2024-07-09T10:15:36.2786870Z ................ [ 35%] 2024-07-09T10:15:36.3305210Z optimize/tests/test_lbfgsb_hessinv.py .. [ 35%] 2024-07-09T10:15:36.3327930Z optimize/tests/test_lbfgsb_setulb.py .. [ 35%] 2024-07-09T10:15:36.9080380Z optimize/tests/test_least_squares.py ................................... [ 35%] 2024-07-09T10:15:38.1155430Z ........................................................................ [ 35%] 2024-07-09T10:15:38.1291890Z ........... [ 35%] 2024-07-09T10:15:38.1900150Z optimize/tests/test_linear_assignment.py ...................... [ 36%] 2024-07-09T10:15:38.2060650Z optimize/tests/test_linesearch.py ........... [ 36%] 2024-07-09T10:15:38.4155820Z optimize/tests/test_linprog.py ......................................... [ 36%] 2024-07-09T10:15:38.6611000Z .......................................s.ss............................. [ 36%] 2024-07-09T10:15:39.0405520Z ...............................................s.s...................... [ 36%] 2024-07-09T10:15:39.3690070Z ....................................................s...s.ss............ [ 36%] 2024-07-09T10:15:39.4870250Z ................................................................ss...... [ 36%] 2024-07-09T10:15:39.8017070Z .....................................................................ssX [ 36%] 2024-07-09T10:15:40.2077170Z ........................................................................ [ 36%] 2024-07-09T10:15:40.3654850Z ....sssX................................................................ [ 37%] 2024-07-09T10:15:40.4783920Z .............s.ssss..................................................... [ 37%] 2024-07-09T10:15:40.6751150Z ..........................s.sss..............................x.......... [ 37%] 2024-07-09T10:15:40.9290110Z ..................................s..................................... [ 37%] 2024-07-09T10:15:42.1105030Z ........x............................................s.................. [ 37%] 2024-07-09T10:15:42.1674890Z .s........................... [ 37%] 2024-07-09T10:15:42.1775110Z optimize/tests/test_lsq_common.py .......... [ 37%] 2024-07-09T10:15:43.1575130Z optimize/tests/test_lsq_linear.py ..........s........s... [ 37%] 2024-07-09T10:15:43.2816280Z optimize/tests/test_milp.py .x.........s [ 37%] 2024-07-09T10:15:43.5628820Z optimize/tests/test_minimize_constrained.py .....................s...... [ 37%] 2024-07-09T10:15:44.1901520Z ..........s.............................ssssssssssssssssssssssssssssssss [ 37%] 2024-07-09T10:15:45.1899620Z ss.....................s................s............................... [ 38%] 2024-07-09T10:15:46.3663690Z ..s.......x........s.................................s.......x........s. [ 38%] 2024-07-09T10:15:46.7592070Z .....................x........X [ 38%] 2024-07-09T10:15:46.8312760Z optimize/tests/test_minpack.py ......................................... [ 38%] 2024-07-09T10:15:46.9424900Z .................................................... [ 38%] 2024-07-09T10:15:46.9556820Z optimize/tests/test_nnls.py ....... [ 38%] 2024-07-09T10:15:47.1980860Z optimize/tests/test_nonlin.py ....................................... [ 38%] 2024-07-09T10:15:47.6160690Z optimize/tests/test_optimize.py .........................x.............. [ 38%] 2024-07-09T10:15:48.1526430Z ...........x.........................x.........................x........ [ 38%] 2024-07-09T10:15:49.6752560Z .....................................................x.................. [ 38%] 2024-07-09T10:15:50.1991420Z ........................................................................ [ 39%] 2024-07-09T10:15:50.3351140Z ........................................................................ [ 39%] 2024-07-09T10:15:58.4262230Z ......................................................s.ss.ss.ss.ss.ss.s [ 39%] 2024-07-09T10:15:58.5769120Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s [ 39%] 2024-07-09T10:15:58.7756750Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s.................................... [ 39%] 2024-07-09T10:15:58.9463060Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s [ 39%] 2024-07-09T10:15:59.2881390Z s.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.ss.s.................. [ 39%] 2024-07-09T10:15:59.3500620Z .... [ 39%] 2024-07-09T10:15:59.3950930Z optimize/tests/test_quadratic_assignment.py .............. [ 39%] 2024-07-09T10:15:59.3976070Z optimize/tests/test_regression.py ... [ 39%] 2024-07-09T10:15:59.4530400Z optimize/tests/test_slsqp.py ..................................... [ 39%] 2024-07-09T10:15:59.5016050Z optimize/tests/test_tnc.py .................... [ 39%] 2024-07-09T10:15:59.5339520Z optimize/tests/test_trustregion.py ........ [ 39%] 2024-07-09T10:15:59.8097390Z optimize/tests/test_trustregion_exact.py .......... [ 40%] 2024-07-09T10:15:59.8149600Z optimize/tests/test_trustregion_krylov.py ..... [ 40%] 2024-07-09T10:15:59.8613950Z optimize/tests/test_zeros.py ........................................... [ 40%] 2024-07-09T10:16:00.0879160Z ........................................................................ [ 40%] 2024-07-09T10:16:00.1109110Z ...................... [ 40%] 2024-07-09T10:16:00.1167350Z signal/tests/test_array_tools.py ...... [ 40%] 2024-07-09T10:16:00.1469750Z signal/tests/test_bsplines.py ......................XXXXXXX.. [ 40%] 2024-07-09T10:16:00.1739540Z signal/tests/test_cont2discrete.py ........................ [ 40%] 2024-07-09T10:16:00.2598200Z signal/tests/test_czt.py ............................................... [ 40%] 2024-07-09T10:16:00.3353940Z ........................................................................ [ 40%] 2024-07-09T10:16:00.3776900Z .......................................... [ 40%] 2024-07-09T10:16:00.4092980Z signal/tests/test_dltisys.py ................................ [ 40%] 2024-07-09T10:16:00.4527410Z signal/tests/test_filter_design.py ..................................... [ 40%] 2024-07-09T10:16:00.7298940Z ........................................................................ [ 40%] 2024-07-09T10:16:01.8037600Z ........................................................................ [ 41%] 2024-07-09T10:16:02.0839580Z .............................................................. [ 41%] 2024-07-09T10:16:02.1196630Z signal/tests/test_fir_filter_design.py ................................. [ 41%] 2024-07-09T10:16:02.2388470Z .... [ 41%] 2024-07-09T10:16:02.3864560Z signal/tests/test_ltisys.py ............................................ [ 41%] 2024-07-09T10:16:02.4179210Z ..................................... [ 41%] 2024-07-09T10:16:02.4248210Z signal/tests/test_max_len_seq.py .. [ 41%] 2024-07-09T10:16:02.4607540Z signal/tests/test_peak_finding.py ...................................... [ 41%] 2024-07-09T10:16:02.5206730Z ............... [ 41%] 2024-07-09T10:16:02.5257490Z signal/tests/test_result_type.py ..... [ 41%] 2024-07-09T10:16:02.5515660Z signal/tests/test_savitzky_golay.py .............. [ 41%] 2024-07-09T10:16:02.6127120Z signal/tests/test_short_time_fft.py .................................... [ 41%] 2024-07-09T10:16:02.8062180Z ........................................................................ [ 41%] 2024-07-09T10:16:03.1699940Z ........................................ [ 41%] 2024-07-09T10:16:03.2585410Z signal/tests/test_signaltools.py ....................................... [ 41%] 2024-07-09T10:16:03.3188100Z ........................................................................ [ 42%] 2024-07-09T10:16:03.5789530Z ........................................................................ [ 42%] 2024-07-09T10:16:03.6483680Z ........................................................................ [ 42%] 2024-07-09T10:16:03.7216200Z ........................................................................ [ 42%] 2024-07-09T10:16:03.7953200Z ........................................................................ [ 42%] 2024-07-09T10:16:03.8685520Z ........................................................................ [ 42%] 2024-07-09T10:16:03.9417370Z ........................................................................ [ 42%] 2024-07-09T10:16:04.0166170Z ........................................................................ [ 43%] 2024-07-09T10:16:04.1058930Z ........................................................................ [ 43%] 2024-07-09T10:16:04.2119420Z ........................................................................ [ 43%] 2024-07-09T10:16:04.3251090Z ........................................................................ [ 43%] 2024-07-09T10:16:04.4468720Z ........................................................................ [ 43%] 2024-07-09T10:16:04.5551980Z ........................................................................ [ 43%] 2024-07-09T10:16:04.6493940Z ........................................................................ [ 43%] 2024-07-09T10:16:04.7529190Z ........................................................................ [ 43%] 2024-07-09T10:16:04.8595980Z ........................................................................ [ 44%] 2024-07-09T10:16:04.9487580Z ........................................................................ [ 44%] 2024-07-09T10:16:04.9614450Z .....ss..../Users/runner/work/scipy/scipy/tools/wheels/cibw_test_command.sh: line 11: 13615 Trace/BPT trap: 5 python -c "import sys; import scipy; sys.exit(not scipy.test())" 2024-07-09T10:16:05.8676330Z ##[endgroup] 2024-07-09T10:16:05.8678070Z ✕ 121.80s 2024-07-09T10:16:05.8724980Z ##[error]Command ['/bin/sh', '-c', 'bash /Users/runner/work/scipy/scipy/tools/wheels/cibw_test_command.sh /Users/runner/work/scipy/scipy'] failed with code 133. 2024-07-09T10:16:05.8731550Z 2024-07-09T10:16:05.9060960Z ##[error]Process completed with exit code 1. 2024-07-09T10:16:05.9228700Z Post job cleanup. 2024-07-09T10:16:05.9299200Z Post job cleanup. 2024-07-09T10:16:06.2955200Z [command]/opt/homebrew/bin/git version 2024-07-09T10:16:06.3255860Z git version 2.45.2 2024-07-09T10:16:06.3296050Z Copying '/Users/runner/.gitconfig' to '/Users/runner/work/_temp/955ef7cb-151e-4e43-9d1d-139ce053d906/.gitconfig' 2024-07-09T10:16:06.3306900Z Temporarily overriding HOME='/Users/runner/work/_temp/955ef7cb-151e-4e43-9d1d-139ce053d906' before making global git config changes 2024-07-09T10:16:06.3307910Z Adding repository directory to the temporary git global config as a safe directory 2024-07-09T10:16:06.3309400Z [command]/opt/homebrew/bin/git config --global --add safe.directory /Users/runner/work/scipy/scipy 2024-07-09T10:16:06.3536820Z [command]/opt/homebrew/bin/git config --local --name-only --get-regexp core\.sshCommand 2024-07-09T10:16:06.4118110Z [command]/opt/homebrew/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'core\.sshCommand' && git config --local --unset-all 'core.sshCommand' || :" 2024-07-09T10:16:06.5565800Z Entering 'doc/source/_static/scipy-mathjax' 2024-07-09T10:16:06.5729790Z Entering 'scipy/_lib/array_api_compat' 2024-07-09T10:16:06.5886670Z Entering 'scipy/_lib/boost_math' 2024-07-09T10:16:06.6096750Z Entering 'scipy/_lib/cobyqa' 2024-07-09T10:16:06.6270050Z Entering 'scipy/_lib/highs' 2024-07-09T10:16:06.6455590Z Entering 'scipy/_lib/pocketfft' 2024-07-09T10:16:06.6631580Z Entering 'scipy/_lib/unuran' 2024-07-09T10:16:06.6809800Z Entering 'scipy/sparse/linalg/_propack/PROPACK' 2024-07-09T10:16:06.6968700Z [command]/opt/homebrew/bin/git config --local --name-only --get-regexp http\.https\:\/\/github\.com\/\.extraheader 2024-07-09T10:16:06.7004900Z http.https://github.com/.extraheader 2024-07-09T10:16:06.7012330Z [command]/opt/homebrew/bin/git config --local --unset-all http.https://github.com/.extraheader 2024-07-09T10:16:06.7057840Z [command]/opt/homebrew/bin/git submodule foreach --recursive sh -c "git config --local --name-only --get-regexp 'http\.https\:\/\/github\.com\/\.extraheader' && git config --local --unset-all 'http.https://github.com/.extraheader' || :" 2024-07-09T10:16:06.7621440Z Entering 'doc/source/_static/scipy-mathjax' 2024-07-09T10:16:06.7718540Z http.https://github.com/.extraheader 2024-07-09T10:16:06.7813910Z Entering 'scipy/_lib/array_api_compat' 2024-07-09T10:16:06.7932730Z http.https://github.com/.extraheader 2024-07-09T10:16:06.8060740Z Entering 'scipy/_lib/boost_math' 2024-07-09T10:16:06.8171270Z http.https://github.com/.extraheader 2024-07-09T10:16:06.8279550Z Entering 'scipy/_lib/cobyqa' 2024-07-09T10:16:06.8394170Z http.https://github.com/.extraheader 2024-07-09T10:16:06.8489320Z Entering 'scipy/_lib/highs' 2024-07-09T10:16:06.8592700Z http.https://github.com/.extraheader 2024-07-09T10:16:06.8706510Z Entering 'scipy/_lib/pocketfft' 2024-07-09T10:16:06.8956800Z http.https://github.com/.extraheader 2024-07-09T10:16:06.9190440Z Entering 'scipy/_lib/unuran' 2024-07-09T10:16:06.9285650Z http.https://github.com/.extraheader 2024-07-09T10:16:06.9469110Z Entering 'scipy/sparse/linalg/_propack/PROPACK' 2024-07-09T10:16:06.9562200Z http.https://github.com/.extraheader 2024-07-09T10:16:06.9945710Z Cleaning up orphan processes 2024-07-09T10:16:07.7179740Z Terminate orphan process: pid (4062) (diskimages-help) Based on the test output (......ss....), and assuming that locally the test order within test_signaltools.py is unchanged from what it was in CI, the crash happened in: scipy/signal/tests/test_signaltools.py::TestMedFilt::test_medfilt2d_parallel[float32] which certainly looks likely since it's calling medfilt2d () under ThreadPoolExecutor(max_workers=4): https://github.com/scipy/scipy/blob/81c53d48a290b604ec5faa34c0a7d48537b487d6/scipy/signal/tests/test_signaltools.py#L1171-L1180 The C code that ends up being called there is: _sigtools_median2d: https://github.com/scipy/scipy/blob/81c53d48a290b604ec5faa34c0a7d48537b487d6/scipy/signal/_sigtoolsmodule.c#L954 _f_medfilt2d: https://github.com/scipy/scipy/blob/81c53d48a290b604ec5faa34c0a7d48537b487d6/scipy/signal/_medianfilter.c#L130 I'd expect this to be reproducible locally. I wonder if the following patch still crashes: $ git diff diff --git a/scipy/signal/_signaltools.py b/scipy/signal/_signaltools.py index 13caf1718f..5eb51f3596 100644 --- a/scipy/signal/_signaltools.py +++ b/scipy/signal/_signaltools.py @@ -2004,8 +2004,8 @@ def medfilt2d(input, kernel_size=3): # checking dtype.type, rather than just dtype, is necessary for # excluding np.longdouble with MS Visual C. - if image.dtype.type not in (np.ubyte, np.float32, np.float64): - return medfilt(image, kernel_size) +# if image.dtype.type not in (np.ubyte, np.float32, np.float64): + return medfilt(image, kernel_size) if kernel_size is None: kernel_size = [3] * 2 what the patch does, it utlimately calls ndimage.rank_filter instead of multipack-era C implementation. Maybe instead of debugging those, it's easier (and certainly better for the code health overall) to use ndimage filtering across signal filtering functions. (We did that for medfilt recently, got no complaints so far) I tried to reproduce this one locally, however, all tests in signal/tests/test_signaltools.py are passing successfully under PYTHON_GIL=0 I was able to trigger the crash. I installed pytest-repeat using pip and then ran the following command: PYTHON_GIL=0 lldb $(pyenv which python) -- dev.py test -- 'scipy.signal.tests.test_signaltools::TestMedFilt' --count=1000 And after some iterations iterations, hit a crash with the following python traceback on the segfaulting thread: File "/Users/goldbaum/Documents/scipy/build-install/lib/python3.13/site-packages/scipy/signal/_signaltools.py", line 1951 in medfilt2d File "/Users/goldbaum/Documents/scipy/build-install/lib/python3.13/site-packages/scipy/signal/tests/test_signaltools.py", line 1168 in apply File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/concurrent/futures/thread.py", line 58 in run File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/concurrent/futures/thread.py", line 92 in _worker File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/threading.py", line 990 in run File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/threading.py", line 1039 in _bootstrap_inner File "/Users/goldbaum/.pyenv/versions/3.13.0b3t/lib/python3.13/threading.py", line 1010 in _bootstrap and the following C traceback, (lldb) bt * thread #3, stop reason = EXC_BAD_ACCESS (code=261, address=0xd50323ff) * frame #0: 0x00000001842c1698 libsystem_platform.dylib`setjmp + 36 frame #1: 0x00000001006a30f4 _sigtools.cpython-313t-darwin.so`_sigtools_median2d(__NPY_UNUSED_TAGGEDdummy='0x171e12c30', args=('0x53580023728', '0x53580023730')) at _sigtoolsmodule.c:980:9 frame #2: 0x0000000100a878dc libpython3.13t.dylib`cfunction_call + 180 frame #3: 0x0000000100a23368 libpython3.13t.dylib`_PyObject_MakeTpCall + 324 frame #4: 0x0000000100b5efd0 libpython3.13t.dylib`_PyEval_EvalFrameDefault + 9684 frame #5: 0x0000000100a26818 libpython3.13t.dylib`method_vectorcall + 328 frame #6: 0x0000000100c512c0 libpython3.13t.dylib`thread_run + 128 frame #7: 0x0000000100be6f84 libpython3.13t.dylib`pythread_wrapper + 28 frame #8: 0x0000000184292f94 libsystem_pthread.dylib`_pthread_start + 136 Which is in a call to setjmp. I suspect there are hairy thread safety issues around use of longjmp/setjmp. I think what's happening is described more or less in this StackOverflow: https://stackoverflow.com/questions/38234546/setjmp-longjmp-between-threads-to-handle-timeout There is a static jmp_buf in _sigtoolsmodule.c which defines a spot to handle malloc failures. It's only every used in one spot. IMO, get rid of it and just use a normal goto fail block instead. Unless I'm misunderstanding why this code uses setjmp/longjmp at all. Unless I'm misunderstanding why this code uses setjmp/longjmp at all. I think it shouldn't indeed. The check_malloc should also be cleaned up. There's no callback mechanism here that would justify using setjmp/longjmp. Rather, it only avoids a single level of regular error handling it looks like (f_medfilt2d is void rather than int). gh-21160 fixes this issue.
gharchive/issue
2024-07-09T11:38:56
2025-04-01T06:45:44.745229
{ "authors": [ "andfoy", "ev-br", "ngoldbaum", "rgommers" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/21142", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2526940707
BUG: Windows Arm build fails Describe your issue. Can't install scipy on win arm machine for native Python. No pre-built wheels, from source compilation fails. Reproducing Code Example pip install scipy Error message Downloading scipy-1.14.1.tar.gz (58.6 MB) --------------------------------------- 58.6/58.6 MB 24.2 MB/s eta 0:00:00 Installing build dependencies: started Installing build dependencies: finished with status 'done' Getting requirements to build wheel: started Getting requirements to build wheel: finished with status 'done' Installing backend dependencies: started Installing backend dependencies: finished with status 'done' Preparing metadata (pyproject.toml): started Preparing metadata (pyproject.toml): finished with status 'error' error: subprocess-exited-with-error Preparing metadata (pyproject.toml) did not run successfully. exit code: 1 [39 lines of output] + meson setup C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh -Dbuildtype=release -Db_ndebug=if-release -Db_vscrt=md --native-file=C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh\meson-python-native-file.ini The Meson build system Version: 1.5.1 Source dir: C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c Build dir: C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh Build type: native build Project name: scipy Project version: 1.14.1 Activating VS 17.8.3 C compiler for the host machine: cl (msvc 19.38.33133 "Microsoft (R) C/C++ Optimizing Compiler Version 19.38.33133 for ARM64") C linker for the host machine: link link 14.38.33133.0 C++ compiler for the host machine: cl (msvc 19.38.33133 "Microsoft (R) C/C++ Optimizing Compiler Version 19.38.33133 for ARM64") C++ linker for the host machine: link link 14.38.33133.0 Cython compiler for the host machine: cython (cython 3.0.11) Host machine cpu family: aarch64 Host machine cpu: aarch64 Program python found: YES (X:\Github\PWABuilder\Windows-DevRel\Tools\ML-tests\.temp\env_scipy\Scripts\python.exe) Run-time dependency python found: YES 3.11 Program cython found: YES (C:\Users\gkhmyznikov\AppData\Local\Temp\pip-build-env-cw5uc0or\overlay\Scripts\cython.EXE) Compiler for C supports arguments -Wno-unused-but-set-variable: NO Compiler for C supports arguments -Wno-unused-function: NO Compiler for C supports arguments -Wno-conversion: NO Compiler for C supports arguments -Wno-misleading-indentation: NO Library m found: NO ..\meson.build:84:0: ERROR: Unknown compiler(s): [['ifort'], ['gfortran'], ['flang'], ['pgfortran'], ['g95']] The following exception(s) were encountered: Running `ifort --version` gave "[WinError 2] The system cannot find the file specified" Running `ifort -V` gave "[WinError 2] The system cannot find the file specified" Running `gfortran --version` gave "[WinError 2] The system cannot find the file specified" Running `gfortran -V` gave "[WinError 2] The system cannot find the file specified" Running `flang --version` gave "[WinError 2] The system cannot find the file specified" Running `flang -V` gave "[WinError 2] The system cannot find the file specified" Running `pgfortran --version` gave "[WinError 2] The system cannot find the file specified" Running `pgfortran -V` gave "[WinError 2] The system cannot find the file specified" Running `g95 --version` gave "[WinError 2] The system cannot find the file specified" Running `g95 -V` gave "[WinError 2] The system cannot find the file specified" A full log can be found at C:\Users\gkhmyznikov\AppData\Local\Temp\pip-install-1pan2rrf\scipy_d66596b99a184dd3a6302c554d0a9d7c\.mesonpy-m6g0z9mh\meson-logs\meson-log.txt [end of output] note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed System.Management.Automation.RemoteException Encountered error while generating package metadata. System.Management.Automation.RemoteException See above for output. System.Management.Automation.RemoteException SciPy/NumPy/Python version and system information scipy 1.14.1 python 3.11.9 win-arm64 I don't know if the build will succeed but at least to try, you need a fortran compiler on your system. That's what the error is saying. @ilayn any fortran compiler for win arm is available? I have no idea unfortunately. Not sure if we even support windows arm. It's not an architecture I am familiar with. Given that the root cause here is a missing Fortran compiler and the builds from @cgohlke show that a from source build on Windows on Arm works fine at the moment, let's close this issue. We can open a "provide wheels for Windows on Arm" issue once NumPy has wheels, it's not actionable now.
gharchive/issue
2024-09-15T14:31:51
2025-04-01T06:45:44.756371
{ "authors": [ "ilayn", "khmyznikov", "rgommers" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/21562", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
214562613
Add guidance on error/warning/exception handling patterns and best practices I could not easily find information on what to expect from scipy in terms error/warning/exception handling patterns and best practices, either as a consumer or a contributor. I looked here: http://scipy.github.io/devdocs/hacking.html and here: http://scipy.github.io/devdocs/dev/index.html Not sure what you mean. Maybe an example would help? So numpy has settings for how to handle floating point errors here. In a "typical" scipy routine, what should cause a warning vs. an exception, and what about returning a separate error value/boolean from a computation vs. simply having a nan for the output to signify something went wrong? How can the python logger be leveraged effectively, if desired? What design pattens allow a scipy consumer flexibility in detecting, handling, and logging warnings and errors? I have never seen any written down or spelled out policy for this. based on my experience: numpy still applies, e.g. Zero Division Warnings in any code depend on the numpy settings incorrect user input that is not usable raises an Exception, ValueError or similar, This can be either explicit in the scipy function or via numpy if for example shapes don't match. vectorized functions return np.nan in positions that cannot be computed to a specific value np.inf is mostly just a regular float and not a value to check for optional checks can raise if condition applies, e.g. isfinite check in linalg or nan_policy now in stats specific computational problems like nonconvergence in optimize or integration problems might just issue a specific warning basic principle: If the user input is ambiguous or there no useful computations can be performed, then raise. If there is a problem, but the computation still contains useful information, then warn or return nan and let the users check it. (e.g. nonconvergence or vectorized) Also a note on So numpy has settings for how to handle floating point errors As of 0.19 scipy.special has a similar style of error handling: https://scipy.github.io/devdocs/special.html#error-handling Also as this isn't a bug the mailing list would probably be a more appropriate place for this discussion. Thanks for your feedback here. I'm looking into the mailing list option for ideas around formulating best practices. The dev mailing list would be the appropriate location to discuss whether this can or should be spelled out. My impression is that it is largely an implicit policy. (related aside: statsmodels is suffering from "organic growth" in this area without systematic review or policy and insufficient unit tests for flags, warnings and exceptions if problems show up https://github.com/statsmodels/statsmodels/issues/3520 scipy is much better but I would expect that it still has some inconsistencies)
gharchive/issue
2017-03-16T00:07:25
2025-04-01T06:45:44.764410
{ "authors": [ "josef-pkt", "person142", "thunderfish24" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/7183", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
261725566
linalg.solve ill-condition warning For the linalg.solve, since version 19.1, we have the capability of printing the warning the ill conditioning of the matrix. However the result is not quite pretty due to the traceback inclusion. For example: I think the story got a bit uglier with our latest change that also prints the reported dlamch machine precision since now we actually have to look for the correct number. I think it's best if we print the truncated versions of those numbers anyhow. But should we switch to a warning printing scheme instead? I don't know yet any way to raise a warning without printing the traceback. One use case I'd like to retain is that if a user is solving iteratively a lot of equations s/he should still be able to catch a warning if the problem is ill-conditioned (which was the reason to emit a warning in the first place). Or we can emit a warning directly and then do a custom printing. Such that there is a clear distinction between the traceback and the actual warning message. Can we set the stacklevel so that it's in the calling function? Then the traceback is much more useful. In general, all uses of warnings.warn should have stacklevel set. I don't know actually. If it is too deep, let's say (I'm just making it up) something is computing a filter coefficient for conversion from state space to transfer function, to get the frequency response to draw Bode plot, which function should be the source the deepest or the calling function? If it is too deep, This is a problem for the caller, whether it is a user or a library. In other words, in your example, it's a problem for the state space -> transfer function code; and we should improve our filter code so that it does the right thing with this warning (avoids it or catches it). So I'd say just go up however many stacklevels (2?) to get to the one that called e.g. scipy.linalg.solve. I've played around with this a bit more and it turns out that setting the right value to the stacklevel is trickier. I've also asked in SO to see whether someone teaches me more about it. Essentially there are two problems : from warnings import warn, formatwarning def _warning_on_one_line(message, category, filename, lineno, file=None, line=None): return ' %s:%s: %s:%s' % (filename, lineno, category.__name__, message) def inner(): formatwarning = _warning_on_one_line warn('Some warning', RuntimeWarning, stacklevel=3) return def middle(): return inner() def outer(): return middle() outer() As you play with stacklevel argument you can basically cause the warning to report different lines in this code. But if you overdo it you start getting IPython's caller stack all the way up to 26 where it finally starts to say sys.1: RuntimeWarning so detecting the correct caller doesn't seem obvious to me. And I agree that I think we have to be a bit more systematic when including a potentially warning emitting code inside a function. Such that we can catch the warning and emit a more meaningful message. So it seems to me that we are stuck with stacklevel=1 for now. Second problem is how to format the warning message but that is kind how it is done as given in the above code. Do you have a way to overcome this issue? My idea is/was: So I'd say just go up however many stacklevels (2?) to get to the one that called e.g. scipy.linalg.solve. If we know where the warning is emitted, and we know the codepath to get there from scipy.linalg.solve, we can do this, right? ... in other words, in your example, if scipy.linalg.solve (in our codebase) is inner, you'd use stacklevel=1, if it's middle you'd use stacklevel=2, and if it's outer you'd use stacklevel=3 (maybe +/-1 to all of these, can't remember the level details). So it seems to me that we are stuck with stacklevel=1 for now. If you're saying that the warning is currently emitted in scipy.linalg.solve, then yeah I agree stacklevel=1 (assuming that's the one that shows scipy.linalg.solve's caller) is the right thing to do. The stacktrace indicates the point in the calling code where the warning is emitted. The stacklevel must correspond to the innermost public API function. This is what Numpy does, and it is sensible behavior. For a public API function, we write def public_api_func(): warnings.warn("some warning", stacklevel=2) For an internal function for which we know the distance to the API function: def _helper_func(): warnings.warn("some warning", stacklevel=3) def public_api_func(): _helper_func() If we know where the warning is emitted, and we know the codepath to get there from scipy.linalg.solve, we can do this, right? Yes indeed but depending on platform we might go too far and start addressing irrelevant parts of the current stack, that's what I meant by the 26 argument. For example if I use stacklevel=9 runfile('C:/Users/Ilhan Polat/Documents/pydump/warning_exp.py', wdir='C:/Users/Ilhan Polat/Documents/pydump') c:\users\ilhan polat\appdata\local\programs\python\python36\lib\site-packages\IPython\core\interactiveshell.py:2808: RuntimeWarning: Some warning if self.run_code(code, result): which is related to IPython internals So give me the top of the stack kind of argument won't be accurate to get the actual code path. The stacktrace indicates the point in the calling code where the warning is emitted. The stacklevel must correspond to the innermost public API function. This is what Numpy does, and it is sensible behavior. Ah OK now I get your arguments. Warnings emitted inside Scipy code (e.g. due to calling solve() on a singular matrix in the implementation of some routine) are probably implementation defect on our side, although usually probably minor one. Either we e.g. know the algorithm works correctly despite such singularity (in which case we should ensure somehow the warning doesn't get emitted), or the singularity implies a failure of the algorithm (in which case an error or a more specific warning would be more appropriate in the perfect world). I agree but now I'm a bit confused how to detect an ill-conditioned array. Because if we remove the warning, then it goes unnoticed. This is known to matlab users a lot with its silent exception handling and only blowing up when the algorithm cannot move any forward. If we emit the warning only pointing to where linalg.solve is called then we have the current situation where the warning seems to be unrelated to the current caller. Silencing the warning if it is not related is easy. Modifying the warning message is also doable by catching the warning. However I don't know what should be the way to go with this. Because if we remove the warning, then it goes unnoticed IIUC the propsal is to: Always emit the warning in the scipy.linalg.solve code when ill conditioned matrices are detected. In places in SciPy where we internally use scipy.linalg.solve, we should "do the right thing". In other words, we should deal with the warning safely so that no warning is ever emitted. This could mean "catch the warning and raise an error" in cases where rank deficiency is bad, or it could mean "catch the warning, and then use linalg.lstsq instead, because rank-deficiency is fine but we'd like the speed of solve if possible". Point (2) is something we need to decide / do on a case-by-case basis as it comes up, so I don't think we need to worry about it right now. OK, I'll just try to make the warning message a bit more structured then and leave the decision part for later then.
gharchive/issue
2017-09-29T18:27:33
2025-04-01T06:45:44.780991
{ "authors": [ "ilayn", "larsoner", "pv" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/7951", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
266012502
welch method unexpected results for low frequency When computing the PSD of the synthetic signal below using welch method I obtain an unexpected peak near 0 Hz with no physical meaning. The equivalent MATLAB code produces different results for that region. Python code example: import numpy as np import matplotlib.pyplot as plt from scipy.signal import welch, hanning x = np.linspace(0, 10, 100001) dt = x[1] - x[0] fs = 1 / dt a1 = 1 f1 = 500 a2 = 10 f2 = 2000 y = a1 * np.sin(2*np.pi*f1*x) + a2 * np.sin(2*np.pi*f2*x) datos = y nblock = 1024 overlap = 128 win = hanning(nblock, True) f, Pxxf = welch(datos, fs, window=win, noverlap=overlap, nfft=nblock, return_onesided=True) plt.semilogy(f, Pxxf, '-o') plt.grid() plt.show() MATLAB code example x = linspace(0, 10, 100001); dt = x(2) - x(1); Fs = 1 / dt; a1 = 1; f1 = 500; a2 = 10; f2 = 2000; y = a1 * sin(2*pi*f1*x) + a2 * sin(2*pi*f2*x); nblock=1024; overlap=128; windowsel=hann(nblock); [Pxx,f]=pwelch(y,windowsel,overlap,nblock,Fs,'onesided'); figure() semilogy(f,Pxx, '-o') Scipy/Numpy/Python version information: Scipy 0.19.1 NumPy 1.13.1 Python 3.6.2 final (serial 0) Matlab: Python I can reproduce with simplified scripts that only process a single window and shorter signals: fs = 10. n_samp = 32 t = np.arange(n_samp) / fs y = np.sin(2*np.pi*2*t) win = windows.hann(n_samp, False) f, Pxxf = welch(y, fs, window=win, noverlap=0, nfft=n_samp, return_onesided=True) plt.semilogy(f, Pxxf, '-o') plt.xlim([0, 5]) plt.ylim([1e-7, 1]) plt.show() MATLAB: clear all; close all; home; fs = 10; n_samp = 32; x = (0:n_samp-1) / fs; y = sin(2*pi*2*x); overlap=0; windowsel=hann(n_samp); [Pxx,f]=pwelch(y,windowsel,overlap,n_samp,fs,'onesided'); figure(); semilogy(f,Pxx, '-o'); But Octave at least agrees with our output: I'm not sure what MATLAB is doing to change/correct these values. I will have a look at the implementation of scipy and try to figure out where this behaviour might be coming from. Do you have any suggestion on that? I have also found this matlab example that might be relevant: https://es.mathworks.com/matlabcentral/answers/33653-psd-estimation-fft-vs-welch It compares results obtained using FFT with results using PSD and both show the same behaviour near 0 Hz. It may be a good starting point. Try using detrend=False. signal.welch defaults to subtracting the mean of the data (detrend='constant') which seems to be giving you this DC artifact. Thanks @ewmoore I can confirm detrend=False fixes it. Awesome! Thank you! It occurred to me that the root cause of this low-frequency energy showing up in the estimated PSD is that detrending takes place before windowing (check _fft_helper()). This means that the DC component of the windowed frame is not zero even though it has been detrended to have zero mean! To illustrate, I modified the example provided by @larsoner: fs = 10. n_samp = 32 t = np.arange(n_samp) / fs y = np.sin(2*np.pi*2*t) win = windows.hann(n_samp, False) windowed_y = y*win # apply window already now f, Pxxf = welch(windowed_y , fs, window='boxcar', noverlap=0) # do not apply it here plt.semilogy(f, Pxxf, '-o') plt.xlim([0, 5]) plt.ylim([1e-7, 1]) plt.show() This results in: Which has zero energy at DC, as expected given the default detrending method. @larsoner, do you think this qualifies as a bug that should be reported? Sorry for adding a comment years after this issue was marked as closed. (as a procedural note, feel free to open a new issue in case this doesn't receive a response)
gharchive/issue
2017-10-17T06:56:21
2025-04-01T06:45:44.791013
{ "authors": [ "AlexS12", "akukuq", "ewmoore", "larsoner", "lucascolley" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/issues/8045", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1142111325
CI: pin asv to avoid slowdowns in 0.5/0.5.1 [skip azp] [skip actions] Closes gh-15568 Okay, CI is green again, the end of almost 2 weeks of N emails/day about failures. So in it goes. Timing for the job is back to ~25 minutes for the asv run, rather than timing out at 40% after an hour.
gharchive/pull-request
2022-02-18T00:14:26
2025-04-01T06:45:44.792933
{ "authors": [ "rgommers" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/pull/15616", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
142766020
ENH: Add gesvd method for svd First pass at adding gesvd to complement gesdd method for svd. Closes #4994. @pv feel free to comment on if you think this approach makes sense. I hope to find time to get it working fully in the next couple of weeks. When adding something similar to lstsq we called the argument lapack_driver and just directly used the the names of the lapack functions as the option names. It would be good to be consistent. Sounds good, will change to lapack_driver='gesdd' | 'gesvd' Okay @pv @ewmoore this one should be good to go now. @@ master #5994 diff @@ ====================================== Files 238 238 Stmts 43803 43809 +6 Branches 8211 8213 +2 Methods 0 0 ====================================== + Hit 34230 34236 +6 Partial 2603 2603 Missed 6970 6970 Review entire Coverage Diff as of 360d271 Powered by Codecov. Updated on successful CI builds.
gharchive/pull-request
2016-03-22T20:13:10
2025-04-01T06:45:44.797398
{ "authors": [ "Eric89GXL", "codecov-io", "ewmoore" ], "repo": "scipy/scipy", "url": "https://github.com/scipy/scipy/pull/5994", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
275455661
Error response contains request id Fixes #525 Review Checklist Tests were added to cover all code changes Documentation was added / updated Code and tests follow standards in CONTRIBUTING.md Codecov Report Merging #1002 into master will increase coverage by 0.04%. The diff coverage is 87.5%. @@ Coverage Diff @@ ## master #1002 +/- ## ========================================== + Coverage 90.32% 90.36% +0.04% ========================================== Files 50 50 Lines 6716 6717 +1 ========================================== + Hits 6066 6070 +4 + Misses 650 647 -3
gharchive/pull-request
2017-11-20T18:34:40
2025-04-01T06:45:44.810764
{ "authors": [ "codecov-io", "hkethi002" ], "repo": "scitran/core", "url": "https://github.com/scitran/core/pull/1002", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2209650577
Inventory of related efforts Closes #51. @nikosft, thoughts? Can we discuss a little more what you want to add here? @SteveLasker, since you are touching up formatting: I am ok with this as a starting point (by @nikosft and I, thanks to him for his contributions so far). I am sure we can add to it over time incrementally. If you are ready to merge it, so am I. I marked it as ready last evening and hope to add more over time of course, but don't have anything in the immediate future here. 🚢 it when ready. Is there anything we can do to unblock this pull request? Is there anything we can do to unblock this pull request? I am happy with it if you are, we should try to merge it in short order. I probably shouldn't be reviewing and merging a PR I worked on. I am happy too :) To my understanding, related efforts are things that have the same goal as SCITT but are not SCITT. I think it is important to talk about how SCITT relates to other efforts and contextualize it. If not, we presume that information is elsewhere (I think it is not). Is there something we can to adjust or scope the content in the PR to finalize the review, or should we just close this PR? @SteveLasker, can we discuss at the tail end of next week's meeting how we intend to handle this?
gharchive/pull-request
2024-03-27T01:46:07
2025-04-01T06:45:44.814566
{ "authors": [ "aj-stein-nist", "nikosft" ], "repo": "scitt-community/scitt-web", "url": "https://github.com/scitt-community/scitt-web/pull/52", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
872640566
Creating a Shared Library Hello, When I try to use -DBUILD_SHARED_LIBS=on when using Intel compilers on Windows, I receive the following link errors: [870/887] Linking Fortran shared library src\cmumps.dll FAILED: src/cmumps.dll src/cmumps.lib cmd.exe /C "cmd.exe /C ""C:\Program Files\CMake\bin\cmake.exe" -E __create_def C:\Tools\MUMPS\build\src\CMakeFiles\cmumps.dir.\exports.def C:\Tools\MUMPS\build\src\CMakeFiles\cmumps.dir.\exports.def.objs && cd C:\Tools\MUMPS\build" && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_dll --intdir=src\CMakeFiles\cmumps.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\10.0.18362.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\10.0.18362.0\x64\mt.exe --manifests -- xilink /nologo src\CMakeFiles\cmumps.dir\mumps_c.c.obj src\CMakeFiles\cmumps.dir\cana_aux.F.obj src\CMakeFiles\cmumps.dir\cana_aux_par.F.obj src\CMakeFiles\cmumps.dir\cana_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_ELT_m.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_m.F.obj src\CMakeFiles\cmumps.dir\cfac_front_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_type2_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_sispointers_m.F.obj src\CMakeFiles\cmumps.dir\cfac_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_dynamic.F.obj src\CMakeFiles\cmumps.dir\cfac_omp_m.F.obj src\CMakeFiles\cmumps.dir\cfac_par_m.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_l0omp_m.F.obj src\CMakeFiles\cmumps.dir\clr_core.F.obj src\CMakeFiles\cmumps.dir\clr_stats.F.obj src\CMakeFiles\cmumps.dir\clr_type.F.obj src\CMakeFiles\cmumps.dir\cmumps_comm_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_config_file.F.obj src\CMakeFiles\cmumps.dir\cmumps_load.F.obj src\CMakeFiles\cmumps.dir\cmumps_lr_data_m.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc.F.obj src\CMakeFiles\cmumps.dir\cmumps_sol_es.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore_files.F.obj src\CMakeFiles\cmumps.dir\cmumps_struc_def.F.obj src\CMakeFiles\cmumps.dir\comp_tps_m.F.obj src\CMakeFiles\cmumps.dir\csol_lr.F.obj src\CMakeFiles\cmumps.dir\csol_omp_m.F.obj src\CMakeFiles\cmumps.dir\cstatic_ptr_m.F.obj src\CMakeFiles\cmumps.dir\cini_driver.F.obj src\CMakeFiles\cmumps.dir\cana_driver.F.obj src\CMakeFiles\cmumps.dir\cfac_driver.F.obj src\CMakeFiles\cmumps.dir\csol_driver.F.obj src\CMakeFiles\cmumps.dir\csol_distrhs.F.obj src\CMakeFiles\cmumps.dir\cend_driver.F.obj src\CMakeFiles\cmumps.dir\cana_aux_ELT.F.obj src\CMakeFiles\cmumps.dir\cana_dist_m.F.obj src\CMakeFiles\cmumps.dir\cana_LDLT_preprocess.F.obj src\CMakeFiles\cmumps.dir\cana_reordertree.F.obj src\CMakeFiles\cmumps.dir\carrowheads.F.obj src\CMakeFiles\cmumps.dir\cbcast_int.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_asm.F.obj src\CMakeFiles\cmumps.dir\cfac_b.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_distentry.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_lastrtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_alloc_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_compress_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_free_block_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack.F.obj src\CMakeFiles\cmumps.dir\cfac_process_band.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blfac_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto_LDLT.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto.F.obj src\CMakeFiles\cmumps.dir\cfac_process_bf.F.obj src\CMakeFiles\cmumps.dir\cfac_process_end_facto_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type3.F.obj src\CMakeFiles\cmumps.dir\cfac_process_maprow.F.obj src\CMakeFiles\cmumps.dir\cfac_process_master2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_message.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2son.F.obj src\CMakeFiles\cmumps.dir\cfac_process_rtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_root_parallel.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings.F.obj src\CMakeFiles\cmumps.dir\cfac_determinant.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScaleAbs.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScale_util.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_pool.F.obj src\CMakeFiles\cmumps.dir\cfac_type3_symmetrize.F.obj src\CMakeFiles\cmumps.dir\cini_defaults.F.obj src\CMakeFiles\cmumps.dir\cmumps_driver.F.obj src\CMakeFiles\cmumps.dir\cmumps_f77.F.obj src\CMakeFiles\cmumps.dir\cmumps_gpu.c.obj src\CMakeFiles\cmumps.dir\cmumps_iXamax.F.obj src\CMakeFiles\cmumps.dir\cana_mtrans.F.obj src\CMakeFiles\cmumps.dir\cooc_panel_piv.F.obj src\CMakeFiles\cmumps.dir\crank_revealing.F.obj src\CMakeFiles\cmumps.dir\csol_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd.F.obj src\CMakeFiles\cmumps.dir\csol_c.F.obj src\CMakeFiles\cmumps.dir\csol_fwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_fwd.F.obj src\CMakeFiles\cmumps.dir\csol_matvec.F.obj src\CMakeFiles\cmumps.dir\csol_root_parallel.F.obj src\CMakeFiles\cmumps.dir\ctools.F.obj src\CMakeFiles\cmumps.dir\ctype3_root.F.obj /out:src\cmumps.dll /implib:src\cmumps.lib /pdb:src\cmumps.pdb /dll /version:0.0 /machine:x64 /INCREMENTAL:NO /DEF:src\CMakeFiles\cmumps.dir.\exports.def src\mumps_common.lib PORD\pord.lib "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_scalapack_lp64.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_blacs_intelmpi_lp64.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_intel_lp64.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_sequential.lib" "C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_core.lib" "C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpifec.lib" "C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpi.lib" user32.lib && cd ." LINK: command "xilink /nologo src\CMakeFiles\cmumps.dir\mumps_c.c.obj src\CMakeFiles\cmumps.dir\cana_aux.F.obj src\CMakeFiles\cmumps.dir\cana_aux_par.F.obj src\CMakeFiles\cmumps.dir\cana_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_ELT_m.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_master_m.F.obj src\CMakeFiles\cmumps.dir\cfac_front_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LU_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_front_LDLT_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_front_type2_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_sispointers_m.F.obj src\CMakeFiles\cmumps.dir\cfac_lr.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_dynamic.F.obj src\CMakeFiles\cmumps.dir\cfac_omp_m.F.obj src\CMakeFiles\cmumps.dir\cfac_par_m.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_l0omp_m.F.obj src\CMakeFiles\cmumps.dir\clr_core.F.obj src\CMakeFiles\cmumps.dir\clr_stats.F.obj src\CMakeFiles\cmumps.dir\clr_type.F.obj src\CMakeFiles\cmumps.dir\cmumps_comm_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_config_file.F.obj src\CMakeFiles\cmumps.dir\cmumps_load.F.obj src\CMakeFiles\cmumps.dir\cmumps_lr_data_m.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc_buffer.F.obj src\CMakeFiles\cmumps.dir\cmumps_ooc.F.obj src\CMakeFiles\cmumps.dir\cmumps_sol_es.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore.F.obj src\CMakeFiles\cmumps.dir\cmumps_save_restore_files.F.obj src\CMakeFiles\cmumps.dir\cmumps_struc_def.F.obj src\CMakeFiles\cmumps.dir\comp_tps_m.F.obj src\CMakeFiles\cmumps.dir\csol_lr.F.obj src\CMakeFiles\cmumps.dir\csol_omp_m.F.obj src\CMakeFiles\cmumps.dir\cstatic_ptr_m.F.obj src\CMakeFiles\cmumps.dir\cini_driver.F.obj src\CMakeFiles\cmumps.dir\cana_driver.F.obj src\CMakeFiles\cmumps.dir\cfac_driver.F.obj src\CMakeFiles\cmumps.dir\csol_driver.F.obj src\CMakeFiles\cmumps.dir\csol_distrhs.F.obj src\CMakeFiles\cmumps.dir\cend_driver.F.obj src\CMakeFiles\cmumps.dir\cana_aux_ELT.F.obj src\CMakeFiles\cmumps.dir\cana_dist_m.F.obj src\CMakeFiles\cmumps.dir\cana_LDLT_preprocess.F.obj src\CMakeFiles\cmumps.dir\cana_reordertree.F.obj src\CMakeFiles\cmumps.dir\carrowheads.F.obj src\CMakeFiles\cmumps.dir\cbcast_int.F.obj src\CMakeFiles\cmumps.dir\cfac_asm_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_asm.F.obj src\CMakeFiles\cmumps.dir\cfac_b.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_distentry.F.obj src\CMakeFiles\cmumps.dir\cfac_distrib_ELT.F.obj src\CMakeFiles\cmumps.dir\cfac_lastrtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_alloc_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_compress_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_free_block_cb.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack_aux.F.obj src\CMakeFiles\cmumps.dir\cfac_mem_stack.F.obj src\CMakeFiles\cmumps.dir\cfac_process_band.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blfac_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto_LDLT.F.obj src\CMakeFiles\cmumps.dir\cfac_process_blocfacto.F.obj src\CMakeFiles\cmumps.dir\cfac_process_bf.F.obj src\CMakeFiles\cmumps.dir\cfac_process_end_facto_slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type1.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_contrib_type3.F.obj src\CMakeFiles\cmumps.dir\cfac_process_maprow.F.obj src\CMakeFiles\cmumps.dir\cfac_process_master2.F.obj src\CMakeFiles\cmumps.dir\cfac_process_message.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2slave.F.obj src\CMakeFiles\cmumps.dir\cfac_process_root2son.F.obj src\CMakeFiles\cmumps.dir\cfac_process_rtnelind.F.obj src\CMakeFiles\cmumps.dir\cfac_root_parallel.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings.F.obj src\CMakeFiles\cmumps.dir\cfac_determinant.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScaleAbs.F.obj src\CMakeFiles\cmumps.dir\cfac_scalings_simScale_util.F.obj src\CMakeFiles\cmumps.dir\cfac_sol_pool.F.obj src\CMakeFiles\cmumps.dir\cfac_type3_symmetrize.F.obj src\CMakeFiles\cmumps.dir\cini_defaults.F.obj src\CMakeFiles\cmumps.dir\cmumps_driver.F.obj src\CMakeFiles\cmumps.dir\cmumps_f77.F.obj src\CMakeFiles\cmumps.dir\cmumps_gpu.c.obj src\CMakeFiles\cmumps.dir\cmumps_iXamax.F.obj src\CMakeFiles\cmumps.dir\cana_mtrans.F.obj src\CMakeFiles\cmumps.dir\cooc_panel_piv.F.obj src\CMakeFiles\cmumps.dir\crank_revealing.F.obj src\CMakeFiles\cmumps.dir\csol_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_bwd.F.obj src\CMakeFiles\cmumps.dir\csol_c.F.obj src\CMakeFiles\cmumps.dir\csol_fwd_aux.F.obj src\CMakeFiles\cmumps.dir\csol_fwd.F.obj src\CMakeFiles\cmumps.dir\csol_matvec.F.obj src\CMakeFiles\cmumps.dir\csol_root_parallel.F.obj src\CMakeFiles\cmumps.dir\ctools.F.obj src\CMakeFiles\cmumps.dir\ctype3_root.F.obj /out:src\cmumps.dll /implib:src\cmumps.lib /pdb:src\cmumps.pdb /dll /version:0.0 /machine:x64 /INCREMENTAL:NO /DEF:src\CMakeFiles\cmumps.dir.\exports.def src\mumps_common.lib PORD\pord.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_scalapack_lp64.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_blacs_intelmpi_lp64.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_intel_lp64.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_sequential.lib C:\Program Files (x86)\Intel\oneAPI\mkl\latest\lib\intel64\mkl_core.lib C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpifec.lib C:\Tools\Microsoft MPI\SDK\Lib\x64\msmpi.lib user32.lib /MANIFEST /MANIFESTFILE:src\cmumps.dll.manifest" failed (exit code 1120) with the following output: Creating library src\cmumps.lib and object src\cmumps.exp LINK : warning LNK4098: defaultlib 'LIBCMT' conflicts with use of other libs; use /NODEFAULTLIB:library cooc_panel_piv.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U csol_bwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U csol_fwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U cfac_front_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U cfac_front_LU_type1.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U cfac_front_LU_type2.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_U ctools.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cfac_process_blocfacto.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cooc_panel_piv.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L csol_bwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L csol_fwd_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cfac_par_m.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cmumps_ooc_buffer.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cfac_process_blocfacto_LDLT.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cfac_front_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cfac_front_LDLT_type1.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cfac_front_LDLT_type2.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cfac_front_type2_aux.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_L cmumps_load.F.obj : error LNK2019: unresolved external symbol MUMPS_FUTURE_NIV2_mp_FUTURE_NIV2 referenced in function CMUMPS_LOAD_mp_CMUMPS_LOAD_UPDATE cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_ERR_STR_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_ERR_STR_OOC cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_DIM_ERR_STR_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_DIM_ERR_STR_OOC cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_MYID_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_MYID_OOC cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_ICNTL1 referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_ICNTL1 cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_LOW_LEVEL_STRAT_IO referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_LOW_LEVEL_STRAT_IO cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_VADDR referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_VADDR cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_STEP_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_STEP_OOC cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_INODE_SEQUENCE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_DO_IO_AND_CHBUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_INODE_SEQUENCE cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_NB_FILE_TYPE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_BUF_CLEAN_PENDING cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_NB_FILE_TYPE cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_ADDVIRTLIBRE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_END_OOC_BUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_ADDVIRTLIBRE cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_HBUF_SIZE referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_COPY_DATA_TO_BUFFER cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_HBUF_SIZE cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_STRAT_IO_ASYNC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_INIT_DB_BUFFER_PANEL cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_STRAT_IO_ASYNC cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_DIM_BUF_IO referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_OOC_INIT_DB_BUFFER_PANEL cmumps_ooc_buffer.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_KEEP_OOC referenced in function CMUMPS_OOC_BUFFER_mp_CMUMPS_INIT_OOC_BUF cmumps_ooc.F.obj : error LNK2001: unresolved external symbol MUMPS_OOC_COMMON_mp_KEEP_OOC cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_OOC_FCT_TYPE referenced in function CMUMPS_OOC_mp_CMUMPS_IS_THERE_FREE_SPACE cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_WITH_BUF referenced in function CMUMPS_OOC_mp_CMUMPS_NEW_FACTOR cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_PROCNODE_OOC referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_END_FACTO cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_SLAVEF_OOC referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_INIT_SOLVE cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_TYPEF_CB referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_INIT_SOLVE cmumps_ooc.F.obj : error LNK2019: unresolved external symbol MUMPS_OOC_COMMON_mp_SOLVE referenced in function CMUMPS_OOC_mp_CMUMPS_OOC_INIT_SOLVE cfac_process_band.F.obj : error LNK2019: unresolved external symbol MUMPS_FAC_DESCBAND_DATA_M_mp_INODE_WAITED_FOR referenced in function CMUMPS_PROCESS_DESC_BANDE src\cmumps.dll : fatal error LNK1120: 24 unresolved externals [873/887] Building Fortran object src\CMakeFiles\zmumps.dir\zmumps_save_restore.F.obj ninja: build stopped: subcommand failed. I imagine these symbols are supposed to be in mumps_common.dll/mumps_common.lib. When I run dumpbin /exports mumps_common.lib I see that some of the unresolved symbols above are in mumps_common.lib but others aren't. Any ideas on how to build a shared library in this context? P.S.: It is possible to build a shared library using MSYS2 and Mingw64 on Windows by specifying -DCMAKE_GNUtoMS=on -DBUILD_SHARED_LIBS=on but I can't use this approach because of an unrelated bug with gfortran and MSMPI. Thanks. I added a warning in 751f8252d028d573623ad1c8779212a102bd6864 that this setup (shared libs with MSVC/Intel on Windows) isn't supported. I don't have the time to fix that, but if you have a fix I could try it. It does work with MSYS2 / MinGW and BUILD_SHARED_LIBS=on I've refactored some build options to help shared builds on other platforms, which work. However, Windows + Intel oneAPI + shared libs still doesn't work with similar errors to above. I think this might be an issue for upstream MUMPS library. I simply am providing an easy CMake build instead of the Makefiles that upstream MUMPS provides.
gharchive/issue
2021-03-16T14:47:31
2025-04-01T06:45:44.880269
{ "authors": [ "ryagal", "scivision" ], "repo": "scivision/mumps", "url": "https://github.com/scivision/mumps/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
250175557
Finding selectors on hidden elements... Hi there ... new to a lot of dev stuff, and agouti in particular. Question on whether I can call Find methods using CSS selectors on elements that are hidden. Trying to create a selection object by calling the Find method on a file input field that is hidden. (Looking to ultimately call the UploadFile method on the selection object, but failing on the step to create the selection object.) any guidance would be much-appreciated! Hi @smithers1221, apologies for the delay! I recommend using agouti.RunScript to inject JavaScript into the page to manipulate the hidden input field. It's difficult to interact with hidden elements via the WebDriver, as WebDrivers are designed to replicate user actions (which can't happen on hidden elements).
gharchive/issue
2017-08-14T22:48:14
2025-04-01T06:45:44.882952
{ "authors": [ "sclevine", "smithers1221" ], "repo": "sclevine/agouti", "url": "https://github.com/sclevine/agouti/issues/115", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2367666386
Generators: do not hardcode archive exe path Resolve from archive contents instead. Done as part of #70
gharchive/issue
2024-06-22T07:30:27
2025-04-01T06:45:44.897478
{ "authors": [ "scop" ], "repo": "scop/wrun", "url": "https://github.com/scop/wrun/issues/60", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1454973988
Validation of fields is nearly non-existent This works and it shouldn't (look at the protocol) apiVersion: score.dev/v1b1 metadata: name: frontend containers: frontend: image: nginx:alpine service: ports: www: port: 80 protocol: Godzilla-growls-angry targetPort: 80 Translates with score-helm to: containers: frontend: image: name: nginx:alpine service: ports: - name: www port: 80 protocol: Godzilla-growls-angry targetPort: 80 type: ClusterIP This also "works" apiVersion: score.dev/v1b1 metadata: name: frontend container-with-massive-angry-typo: frontend: image: nginx:alpine service: ports: www: port: 80 protocol: Godzilla-growls-angry targetPort: 80 Outputs this: containers: {} service: ports: - name: www port: 80 protocol: Godzilla-growls-angry targetPort: 80 type: ClusterIP And it doesn't error at all, there is no validation as it exists with 0 status In the first section, you specified an invalid protocol for the port. In the second section, you used an incorrect field name for the container, which caused it to be ignored. Try this for the first: apiVersion: score.dev/v1b1 metadata: name: frontend containers: frontend: image: nginx:alpine service: ports: www: port: 80 protocol: TCP # Update the protocol to a valid value, such as TCP targetPort: 80 And this for the helm-part: apiVersion: score.dev/v1b1 metadata: name: frontend containers: # Update the field name to the correct name, which is "containers" frontend: image: nginx:alpine service: ports: www: port: 80 protocol: TCP targetPort: 80 In the first section, you specified an invalid protocol for the port. In the second section, you used an incorrect field name for the container, which caused it to be ignored. I think that's exactly @fernando-villalba's point: these errors should be automatically detected by Score implementations to help with troubleshooting. Duplicate of https://github.com/score-spec/spec/issues/6
gharchive/issue
2022-11-18T11:48:08
2025-04-01T06:45:44.904330
{ "authors": [ "adityassharma-ss", "dserodio", "fernando-villalba", "sujaya-sys" ], "repo": "score-spec/spec", "url": "https://github.com/score-spec/spec/issues/21", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
35743872
Grid images with In Home Streaming There have been reports that grid images don't appear when viewing a stream able game. Instead it displays the file path. If its not working for non-Ice shortcuts then I don't think there is anything I can do to fix the problem. As such, I am closing the task. I hope Valve fixes this issue though. I can understand why they would want the client computer to be able to put in their own images, but I dont see any reason why they wouldn't get the custom images from the host computer by default if it has them. Scott, since you're mucking with pysteam anyways, can you see if there's a way to get the client to automatically grab images for all the remote games it picks up? Steam lets you manually assign custom grid view images to streamed-only games from a client side, so there should be a way to access that in the code. I personally can't find it right now since it appears to not be in shortcuts.vdf or the games list there, but you may have more insight having worked with this library longer. Steam lets you manually assign custom grid view images to streamed-only games from a client side, so there should be a way to access that in the code. I personally can't find it right now since it appears to not be in shortcuts.vdf or the games list there, but you may have more insight having worked with this library longer. Ah, interesting. I have a theory on how this is working, and if I'm right then it should be totally doable to set grid images from a client machine. I won't elaborate too much more though because I haven't looked into it at all and I'm really just guessing. The bigger problem is - how do I get Ice on the client to know about the Ice that is running on the server? Ice wasn't really meant to handle that use case. I could provide a way to keep Ice running in the background and provide an API for a client to read the list of games (and even the images from the host!) but connecting the client Ice to the server Ice will be hard. Once I get a streaming box of my own set up I'll try to work on some basic support for Steam In Home Streaming. Please elaborate! With your familiarity, even a guess is better than my best attempt, and might point me in the right direction. I'm happy to dive into the code myself. Ice could definitely just run on both machines, and check whatever non-shortcuts.vdf file holds all those references, then assume they are all Ice games. I wish Tags were visible remotely, that would fix any confusion for that.
gharchive/issue
2014-06-15T06:55:20
2025-04-01T06:45:44.915042
{ "authors": [ "BlinksTale", "scottrice" ], "repo": "scottrice/Ice", "url": "https://github.com/scottrice/Ice/issues/202", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2022032323
Variable not in scope: pathParam In examples/basic.hs file Hello! I am new to Haskell and Scotty. I am processing examples in basic.hs and I encountered an error (look at the pic below). this is my cabal file, I think I am using the latest version. Name: scotty-examples Version: 0.1 Synopsis: scotty examples Homepage: https://github.com/scotty-web/scotty Bug-reports: https://github.com/scotty-web/scotty/issues License: BSD3 License-file: LICENSE Author: Andrew Farmer <xichekolas@gmail.com> Maintainer: Andrew Farmer <xichekolas@gmail.com> Copyright: (c) 2012-Present, Andrew Farmer and the Scotty contributors Category: Web Stability: experimental Build-type: Simple Cabal-version: >= 1.10 Description: Example programs using @scotty@ tested-with: GHC == 8.10.7 , GHC == 9.0.2 , GHC == 9.2.8 , GHC == 9.4.6 , GHC == 9.6.2 executable CarShop main-is: main.hs default-language: Haskell2010 hs-source-dirs: src build-depends: base >= 4.6 && < 5, http-types, mtl, random, scotty == 0.20.1, text, wai-extra GHC-options: -Wall -threaded I find it will work when using queryParam. I don't know why, but maybe it is better to fix the problem in the examples. Finally, if you consider it as a problem, maybe I can try to make an easy-fix pull request for you. Hi @fongfiafia , could you share the imports block in your main.hs ? 您好,来件已收到,我会尽快给您回复。 Hi @fongfiafia , could you share the imports block in your main.hs ? sure~ {-# LANGUAGE OverloadedStrings #-} module Main (main) where import Web.Scotty main :: IO () main = do scotty 8081 $ get "/user" $ do v <- pathParam "uid" html $ mconcat ["<h1>", v, "</h1>"]
gharchive/issue
2023-12-02T13:02:01
2025-04-01T06:45:44.926090
{ "authors": [ "fongfiafia", "ocramz" ], "repo": "scotty-web/scotty", "url": "https://github.com/scotty-web/scotty/issues/355", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2257430350
feat: add support for unqualified path expressions Part of #29. So, the tests are failing because we don't have Span::join is nightly-only, and we use the start span as a fallback 😿
gharchive/pull-request
2024-04-22T20:43:09
2025-04-01T06:45:44.943325
{ "authors": [ "scrabsha" ], "repo": "scrabsha/expandable", "url": "https://github.com/scrabsha/expandable/pull/40", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
136588782
Splash timeout on all requests. I have set up docker to work with splash. Docker starts up fine but even the simple pages like google.com time out. Follows the splash terminal output for splash startup. docker run -p 8050:8050 scrapinghub/splash 2016-02-26 03:39:07+0000 [-] Log opened. 2016-02-26 03:39:07.903064 [-] Splash version: 2.0.1 2016-02-26 03:39:07.903849 [-] Qt 5.5.1, PyQt 5.5.1, WebKit 538.1, sip 4.17, Twisted 15.5.0, Lua 5.2 2016-02-26 03:39:07.903997 [-] Python 3.4.3 (default, Oct 14 2015, 20:28:29) [GCC 4.8.4] 2016-02-26 03:39:07.904119 [-] Open files limit: 524288 2016-02-26 03:39:07.904248 [-] Open files limit increased from 524288 to 1048576 2016-02-26 03:39:08.112195 [-] Xvfb is started: ['Xvfb', ':1047', '-screen', '0', '1024x768x24'] 2016-02-26 03:39:08.216409 [-] proxy profiles support is enabled, proxy profiles path: /etc/splash/proxy-profiles 2016-02-26 03:39:08.364839 [-] verbosity=1 2016-02-26 03:39:08.365024 [-] slots=50 2016-02-26 03:39:08.365640 [-] Web UI: enabled, Lua: enabled (sandbox: enabled) 2016-02-26 03:39:08.368755 [-] Site starting on 8050 2016-02-26 03:39:08.369102 [-] Starting factory <twisted.web.server.Site object at 0x7fb0ee1c4860> Can you please help me with this. Docker starts up fine but even the simple pages like google.com time out how are you requesting those simple pages? Logs you posted dont contain any trace of requests being made to Splash, they just show Splash is started. If you posted your logs in full it means that requests you make dont reach Splash. Can you ping localhost:8050 and get demo page? When requests are timing out what error are you getting? Are they timing out with 504 status code or you dont get any status code? hey Pawl, Thanks. I started docker with this command: docker run -p 5023:5023 -p 8050:8050 -p 8051:8051 scrapinghub/splash telnet "ping" output is: telnet localhost 5023 Trying 127.0.0.1... Connected to localhost. Escape character is '^]'. Connection closed by foreign host. nmap command trace: nmap -p 8050 localhost Starting Nmap 7.00 ( https://nmap.org ) at 2016-02-26 21:50 PKT Nmap scan report for localhost (127.0.0.1) Host is up (0.00013s latency). PORT STATE SERVICE 8050/tcp open unknown Nmap done: 1 IP address (1 host up) scanned in 0.04 seconds Now I write in address line Firefox (44.0.2): localhost:8050 I write in text-field provided next to "Render" Button. "http://google.com"(its already there btw) and press the button. I am attaching compete logs of what the splash process outputs. If you look at the logs there is a problem that /etc/machine-id is not found. Well it is a problem in Ubuntu. I "patched" it, Can this be the problem? Anyways after processing it gives Error 504 HTTP Error 504 (Gateway Time-out) Type: GlobalTimeoutError Timeout exceeded rendering page { "info": { "timeout": 60 }, "description": "Timeout exceeded rendering page", "error": 504, "type": "GlobalTimeoutError" } Thanks again. :) splash.txt thanks for the detailed report. Is this happening for all splash endpoints or just execute? (when you press render on startpage it makes request to localhost:8050/execute). can you try following url? http://localhost:8050/render.json?url=http://httpbin.org this makes request to render.json. is it responding with 504 too? Yes it responds with 504 too {"description": "Timeout exceeded rendering page", "error": 504, "type": "GlobalTimeoutError", "info": {"timeout": 30}} attached are the logs splash-json.txt Thanks a lot. Ok now this works fine. The only thing I have changed is my internet connection. Now I have a faster internet and it does not timeout. The machine-id error is still there so it was not the case. docker run -p 8050:8050 scrapinghub/splash --max-timeout 3600 Hello Ahmed. I am retrieving the same mistake all the time Some scripts is working fine because it runs in less than 30 seconds Other script is always failing because it is slower. For this, each time that I execute it retrieves "GlobalTimeoutError" The server is deploy inside "docker container" with the following parameters: docker run -p 8050: 8050 --proxy-profiles-path / etc / splash / proxy-profiles --max-timeout 240 --filters-path / etc / splash / filters --js-profiles-path / etc / splash / Js-profiles --lua-sandbox-allowed-modules "populare; utils; popularetarj" -lua-package-path /etc/splash/lua_modules/?.lua --slots 500 -v 2 I feel that the configuration of "-max-timeout" is not working but I do not understand why. Whatever help is welcome. Thank you in advance Details: { "timestamp": 1491407822, "user-agent": null, "client_ip": "192.168.69.136", "method": "POST", "_id": 140363517724656, "path": "/execute", "load": [ 0.09, 0.04, 0.05 ], "status_code": 504, "error": { "error": 504, "info": { "timeout": 240 }, "type": "GlobalTimeoutError", "description": "Timeout exceeded rendering page" }, "active": 0, "maxrss": 1953936, "qsize": 0, "args": { "wait": "0.5", "uid": 140363517724656, "expand": "1", "lua_source": "local mymodule = require('popularetarj') function main(splash) splash.images_enabled = false splash.response_body_enabled = true args = {} args['user']='XXXX' args['pass']='XXXX' args['contract']='XXXX' args['fromDate']='05/03/2017' args['toDate']='05/04/2017' return mymodule.execute(splash, args) end", "images": "1", "timeout": "240.0", "url": "http://google.com" }, "fds": 28, "rendertime": 30.02491569519043 } @mgcuesta --max-timeout is only an upper bound of the timeout you can set per request. I didn't find a global config to set timeout for all requests though. If you want longer timeout you'll have to set them in each SplashRequest. @FreeTymeKiyan How to set for each SplashRequest?
gharchive/issue
2016-02-26T03:51:27
2025-04-01T06:45:44.962429
{ "authors": [ "AhmedRafikDjerah", "FreeTymeKiyan", "gotounix", "mgcuesta", "pawelmhm", "shafaq" ], "repo": "scrapinghub/splash", "url": "https://github.com/scrapinghub/splash/issues/394", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
896917627
use switch for platforms using switch over the plain string is more idiomatic I guess oh my, yep definitely nicer and more go looking than my hacky python :)! Thanks for this @hellt !
gharchive/pull-request
2021-05-20T14:38:03
2025-04-01T06:45:44.963849
{ "authors": [ "carlmontanari", "hellt" ], "repo": "scrapli/scrapligo", "url": "https://github.com/scrapli/scrapligo/pull/2", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1111639289
🛑 Animate Me is down In b4687f2, Animate Me (https://animate-me.atanas.info) was down: HTTP code: 0 Response time: 0 ms Resolved: Animate Me is back up in 1b024e8.
gharchive/issue
2022-01-22T17:51:16
2025-04-01T06:45:44.981220
{ "authors": [ "scriptex" ], "repo": "scriptex/uptime", "url": "https://github.com/scriptex/uptime/issues/1133", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1108577037
🛑 Material Tetris is down In dbb5eb0, Material Tetris (https://tetris.atanas.info) was down: HTTP code: 0 Response time: 0 ms Resolved: Material Tetris is back up in 3dbcc8d.
gharchive/issue
2022-01-19T21:26:21
2025-04-01T06:45:44.983840
{ "authors": [ "scriptex" ], "repo": "scriptex/uptime", "url": "https://github.com/scriptex/uptime/issues/640", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2337503924
add row_usages to ChunkProof struct Purpose or design rationale of this PR add row_usages to ChunkProof struct PR title Your PR title must follow conventional commits (as we are doing squash merge for each PR), so it must start with one of the following types: [ ] build: Changes that affect the build system or external dependencies (example scopes: yarn, eslint, typescript) [ ] ci: Changes to our CI configuration files and scripts (example scopes: vercel, github, cypress) [ ] docs: Documentation-only changes [ ] feat: A new feature [ ] fix: A bug fix [ ] perf: A code change that improves performance [ ] refactor: A code change that doesn't fix a bug, or add a feature, or improves performance [ ] style: Changes that do not affect the meaning of the code (white-space, formatting, missing semi-colons, etc) [ ] test: Adding missing tests or correcting existing tests Deployment tag versioning Has tag in common/version.go been updated or have you added bump-version label to this PR? [ ] No, this PR doesn't involve a new deployment, git tag, docker image tag [ ] Yes Breaking change label Does this PR have the breaking-change label? [ ] No, this PR is not a breaking change [ ] Yes Codecov Report All modified and coverable lines are covered by tests :white_check_mark: Project coverage is 54.31%. Comparing base (c862cb6) to head (86e573b). Additional details and impacted files @@ Coverage Diff @@ ## chore/libzkp-v0.11.0 #1366 +/- ## ===================================================== Coverage 54.31% 54.31% ===================================================== Files 156 156 Lines 11922 11922 ===================================================== Hits 6475 6475 Misses 4952 4952 Partials 495 495 Flag Coverage Δ coordinator 19.93% <ø> (ø) database 42.85% <ø> (ø) prover 26.55% <ø> (ø) Flags with carried forward coverage won't be shown. Click here to find out more. :umbrella: View full report in Codecov by Sentry. :loudspeaker: Have feedback on the report? Share it here. btw, we can merge this PR to develop branch and test the backward compatibility in sepolia. @amoylan2
gharchive/pull-request
2024-06-06T07:08:56
2025-04-01T06:45:45.015169
{ "authors": [ "amoylan2", "codecov-commenter", "colinlyguo" ], "repo": "scroll-tech/scroll", "url": "https://github.com/scroll-tech/scroll/pull/1366", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
417660381
Does it have a way to custom button icon? I want to change the icons. Anyone know how to do this? Thank you! https://github.com/scrumpy/tiptap/issues/193#issuecomment-461126841
gharchive/issue
2019-03-06T07:28:33
2025-04-01T06:45:45.017123
{ "authors": [ "philippkuehn", "tdkihrr" ], "repo": "scrumpy/tiptap", "url": "https://github.com/scrumpy/tiptap/issues/222", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
470177544
substraTEE-client must accept a SHA256 hash of the WASM to be executed Currently (M4), a WASM file has to be specified (or the default file is taken) and the SHA256 hash is calculated over the content. The client is finally separated from the worker and has no access to his internals/files. Therefore, the user must be able to enter the SHA256 hash of the WASM code he/she wants to have executed in the TEE. as of #61, WASM is no longer supported until we do #15
gharchive/issue
2019-07-19T07:20:44
2025-04-01T06:45:45.018369
{ "authors": [ "brenzi", "electronix" ], "repo": "scs/substraTEE-worker", "url": "https://github.com/scs/substraTEE-worker/issues/56", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1969948791
fix: send OK message for ephemeral events Hi, according to last changes in NIP01, the OK message MUST be sent in response to every EVENT message: https://github.com/nostr-protocol/nips/blob/4b9f13d983245e4dd166f102308afc28b8bb1603/01.md?plain=1#L153 Thanks for noticing and fixing this! Merged.
gharchive/pull-request
2023-10-31T08:55:11
2025-04-01T06:45:45.020245
{ "authors": [ "scsibug", "yukibtc" ], "repo": "scsibug/nostr-rs-relay", "url": "https://github.com/scsibug/nostr-rs-relay/pull/168", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
933807126
Add softsegs in derivatives of spine-generic-processed Description The purpose of this issue is to discuss where we will include the softsegs in the derivatives of spine-generic-processed and how we will name them. Suggestion We could inlcude new soft segmentations under /derivatives/labels_softseg/ and have the same structure than under /derivatives/labels/ spine-generic-processed │ ├──derivatives │ │ │ └── labels │   │ ├── sub-amu01 │   │ │   ├── anat │   │ │   │   ├── sub-amu01_acq-T1w_MTS_seg-manual.json │   │ │   │   ├── sub-amu01_acq-T1w_MTS_seg-manual.nii.gz │   │ │   │   ├── sub-amu01_acq-MTon_MTS_seg-manual.json │   │ │   │   ├── sub-amu01_acq-MTon_MTS_seg-manual.nii.gz │   │ │   │   ├── sub-amu01_T2star_seg-manual.json │   │ │   │   ├── sub-amu01_T2star_seg-manual.nii.gz │   │ │   │   ├── sub-amu01_T1w_seg-manual.json │   │ │   │   ├── sub-amu01_T1wr_seg-manual.nii.gz │   │ │   │   ├── sub-amu01_T2w_seg-manual.json │   │ │   │   └── sub-amu01_T2w_seg-manual.nii.gz │   │ │   └── dwi │   │ │   ├── sub-amu01_rec-average_dwi_seg-manual.json │   │ │   └── sub-amu01_rec-average_dwi_seg-manual.nii.gz │ │ ... │ └── labels_softseg ... We also have to decide how to name the softsegs, the suffix -manual isn't appropriate here since softsegs are created by averaging manual segmentations. Thank you for the heads up @sandrinebedard ! Tagging @kousu @alexfoias @copperwiring and @charleygros so they are aware I have generated the soft segmentations, I would be ready to add them to spine-generic-processed dataset. The plan for now would be to add them under derivatives/labels_softseg We could directly omit the suffix -manual since they are reconstructed, we don't need a suffix necessarily for the derivatives from what I understand sub-amu01_T2star_softseg.nii.gz or we could also use the entity desc to differentiate with normal softseg since they are artificially generated here Found this on probalistic segmentations : https://bids-specification.readthedocs.io/en/stable/05-derivatives/03-imaging.html#probabilistic-segmentations If you have any insights/suggestions feel free to tell me! also tagging @uzaymacar so that you are aware 😊! to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz? to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz? although a big problem with that is that it is misleading: the soft segmentation does not come from softseg but from the average of multiple manual segs. So the suffix "manual" still makes sense, no? to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz? although a big problem with that is that it is misleading: the soft segmentation does not come from softseg but from the average of multiple manual segs. So the suffix "manual" still makes sense, no? I would assume that the soft segmentations from softseg would be under /derivatives/labels/ and not /derivatives/labels_softseg, but yeah it may be still be confusing, we could absolutely add the suffix manual. It would also depend on what we call the soft segmentations from softseg. For now, for SC segmentation, we add -manual since it was manually QC-ed or corrected, if we do the same for softsegs, the same confusion would occur... to follow our convention maybe we would prefer sub-amu01_T2star_seg-softseg.nii.gz? although a big problem with that is that it is misleading: the soft segmentation does not come from softseg but from the average of multiple manual segs. So the suffix "manual" still makes sense, no? I would assume that the soft segmentations from softseg would be under /derivatives/labels/ and not /derivatives/labels_softseg, but yeah it may be still be confusing, we could absolutely add the suffix manual. It would also depend on what we call the soft segmentations from softseg. For now, for SC segmentation, we add -manual since it was manually QC-ed or corrected, if we do the same for softsegs, the same confusion would occur... given that the project is very specific to soft training, maybe we should in fact put those derivatives under derivatives/labels_softseg. This is not a bad idea. In that case we would not need the suffix "-manual" because we could add a README.md inside derivatives/labels_softseg/ that clearly explains how these labels were generated.
gharchive/issue
2021-06-30T15:04:17
2025-04-01T06:45:45.031354
{ "authors": [ "jcohenadad", "sandrinebedard" ], "repo": "sct-pipeline/contrast-agnostic-softseg-spinalcord", "url": "https://github.com/sct-pipeline/contrast-agnostic-softseg-spinalcord/issues/10", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
568337555
request: reliable means to interrogate sparsity It would be very helpful if the API had a standard way to query the anndata X sparsity, and sparse format, independent of the backed mode. In other words, a safe / reliable / performant equivalent equivalent of the following on memory-backed adata: scipy.sparse.issparse(adata.X) -- bool indicates if is sparse adata.X.get_format() -- return scipy sparse format, eg, 'csr' Is there a way to do this now? Or a hack to work around the lack of API today? One possible hack-around relies on the weak test for the format_str internal to spmatrix (these are overly wordy for clarity): from scipy import sparse def issparse(adata): if adata.isbacked: return hasattr(adata.X, 'format_str') else: return sparse.issparse(adata.X) def get_sparse_format(adata): if issparse(adata): if adata.isbacked: return getattr(adata.X, 'format_str', 'und') else: return adata.X.getformat() return False If you just want to test sparsity, and not if X is some other kind of array-like thing, you’d try scipy.sparse.issparse(adata.X) or isinstance(anndata.X, scipy.sparse.spmatrix). For the sparse format, you’d use scipy.sparse.isspmatrix_csr or isinstance(..., ...csr_matrix), right? This should actually work for BackedSparseMatrix, as all its subclasses actually inherit from the spmatrix subclasses: https://github.com/theislab/anndata/blob/4440b90ff3dff213b4c512478e21426cf210368d/anndata/_core/sparse_dataset.py#L117 https://github.com/theislab/anndata/blob/4440b90ff3dff213b4c512478e21426cf210368d/anndata/_core/sparse_dataset.py#L143 If you just had plain scipy.sparse classes and wanted to test sparsity, you’d use scipy.sparse.issparse(adata.X) or isinstance(anndata.X, scipy.sparse.spmatrix), right? Yes, that would be perfect. For the sparse format, you’d use scipy.sparse.isspmatrix_csr or isinstance(..., ...csr_matrix), right? I would probably use spmatrix.getformat(), but that is equivalent to scipy.sparse.isspmatrix_csr(), and both should work if adata.X is a subclass of the appropriate spmatrix. That said, in 0.7.1, it doesn't seem to work. The type of adata.X is a SparseMatrix, not a BackedSparseMatrix. >>> adata = anndata.read_h5ad('pbmc3k-CSC-gz.h5ad', backed='r') >>> adata AnnData object with n_obs × n_vars = 2638 × 1838 backed at 'pbmc3k-CSC-gz.h5ad' obs: 'n_genes', 'percent_mito', 'n_counts', 'louvain' var: 'n_cells' uns: 'draw_graph', 'louvain', 'louvain_colors', 'neighbors', 'pca', 'rank_genes_groups' obsm: 'X_pca', 'X_tsne', 'X_umap', 'X_draw_graph_fr' varm: 'PCs' >>> adata.X <HDF5 sparse dataset: format 'csc', shape (2638, 1838), type '<f4'> >>> type(adata.X) <class 'anndata._core.sparse_dataset.SparseDataset'> >>> scipy.sparse.isspmatrix(adata.X) False >>> scipy.sparse.isspmatrix_csc(adata.X) False >>> isinstance(adata.X, scipy.sparse.spmatrix) False >>> isinstance(adata.X, scipy.sparse.csc_matrix) False The SparseMatrix class has some intriguing parameters (eg, format_str), but I assume those are private? That’s very weird, X is supposed to be a matrix class, not a HDF5 dataset. LMK if you need a more complete test case. @bkmartinjr You should be able to get some idea from type(adata.X). Do you have a use case where it's important to know whether it's dense or sparse, but it's not important to know whether it's in memory or on disk? It might be nice to have this in the future, but I'd really like anndata to be more agnostic about the type of the array. That is, someone defines an array type and it works fine without us having to approve it. These trait detection methods should also be implemented in upstream packages (I've suggested this here: https://github.com/scipy/scipy/issues/11565). For right now, I think you could get away with: if isinstance(X, (scipy.sparse.spmatrix, SparseDataset)): # It's sparse elif isinstance(X, (np.ndarray, h5py.Dataset)): # It's dense @flying-sheep if the AnnData is backed and not a view, X is a h5py.Dataset or our SparseDataset. SparseDataset should get exported, and we should make its to_backed() method private. Why isn’t X a matrix-like class that produces matrices when sliced? You should be able to get some idea from type(adata.X) True, but doing that is what got this entire conversation started. AFAIK, there are no public types usable in that manner. As a hack, it is definitely an alternative. I had picked format_str instead of the type, as it seemed to be more broadly backward compatible. It might be nice to have this in the future, but I'd really like anndata to be more agnostic Fully agree with this, as long as you define the minimum interface that the array must support. If that abstract interface is sufficient, all would work fine. I consider sparsity to be necessary in that interface, as it is very hard to write high performance code, for large matrices, without knowing something about how you should partition and stride through it. It is more important, IMHO, than isbacked. For right now, I think you could get away with: This is similar to the hack I was using before, which broke when the internal class structure changed. TL;DR - I like your "agnostic" approach, but what it needs to work IMHO is: a "guaranteed" interface for all array-ish things that can be stashed in X if there are modes of behavior for X (like sparse/dense), interfaces that let users introspect these modes Coming in because I've ran into the exact same issue. Specifically I'm wondering if this will be fixed adata = anndata.read("file.h5ad", "r") type(adata.X) # anndata._core.sparse_dataset.SparseDataset adata = anndata.read("file.h5ad") type(adata.X) # scipy.sparse.csr.csr_matrix Why is the type of X SparseDataset when read in backed mode? With #219 (backed zarr), the incantation for this would get harder. what’s the plan here @ivirshup? Would still like to not implement our own solution, especially one we export, since I don't think we can reliably cover all cases. I would ideally like scipy to make issparse single dispatch, but idk if that's going to happen. Realistically I think the thing to do will be something like: SPARSE_TYPES = (scipy.sparse.sparray, scipy.sparse.spmatrix, ...) isinstance(x, SPARSE_TYPES)
gharchive/issue
2020-02-20T14:25:59
2025-04-01T06:45:45.061646
{ "authors": [ "bkmartinjr", "flying-sheep", "ivirshup", "pablo-gar" ], "repo": "scverse/anndata", "url": "https://github.com/scverse/anndata/issues/326", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
837525473
Opening file with "w" while it's open in backed mode elsewhere still deletes file contents [x] I have checked that this issue has not already been reported. [x] I have confirmed this bug exists on the latest version of scanpy. - Version '1.7.1' [ ] (optional) I have confirmed this bug exists on the master branch of scanpy. When I open adata in backed='r' mode in one script and then I try to modify the same adata/file from another script the adata file gets corrupted (its size becomes 0). I would expect writing to return an error before corrupting the file. Versions ----- anndata 0.7.4 scanpy 1.7.1 sinfo 0.3.1 ----- PIL 7.2.0 anndata 0.7.4 backcall 0.2.0 cairo 1.19.1 cffi 1.14.0 cloudpickle 1.3.0 colorama 0.4.4 cycler 0.10.0 cython_runtime NA dask 2.21.0 dateutil 2.8.0 decorator 4.4.2 future_fstrings NA get_version 2.1 google NA h5py 2.10.0 igraph 0.8.2 ipykernel 5.3.3 ipython_genutils 0.2.0 jedi 0.17.2 joblib 0.16.0 kiwisolver 1.2.0 legacy_api_wrap 0.0.0 leidenalg 0.8.1 llvmlite 0.35.0 louvain 0.6.1 matplotlib 3.3.0 mpl_toolkits NA natsort 7.0.1 numba 0.52.0 numexpr 2.7.1 numpy 1.19.4 packaging 20.9 pandas 1.0.5 parso 0.7.0 pexpect 4.8.0 pickleshare 0.7.5 pkg_resources NA prompt_toolkit 3.0.5 psutil 5.7.2 ptyprocess 0.6.0 pygments 2.6.1 pyparsing 2.4.7 pytz 2020.1 scanpy 1.7.1 scipy 1.4.1 setuptools_scm NA simplejson 3.17.2 sinfo 0.3.1 six 1.15.0 sklearn 0.23.1 storemagic NA tables 3.6.1 texttable 1.6.2 tlz 0.10.0 toolz 0.10.0 tornado 6.0.4 traitlets 4.3.3 typing_extensions NA wcwidth 0.2.5 yaml 5.3.1 zmq 19.0.1 zope NA ----- IPython 7.16.1 jupyter_client 6.1.6 jupyter_core 4.6.3 notebook 6.0.3 ----- Python 3.8.5 | packaged by conda-forge | (default, Jul 22 2020, 17:31:50) [GCC 7.5.0] Linux-3.10.0-1160.11.1.el7.x86_64-x86_64-with-glibc2.10 64 logical CPU cores, x86_64 ----- Session information updated at 2021-03-21 21:14 </details> @Hrovatin, I'd check out the issue at h5py for more info. As it pertains to AnnData though, what exactly were you trying to do? Did you want the file to be overwritten by the second process, or were you looking for it to update it? E.g. would you expect one of those processes to error? I had file open in one process in basked mode. I then decided that I need to make some stuff in another notebook and forgot about having the file open in the first notebook. When I tried to save the file from the second notebook it corrupted it. I would expect it at least to err out before corrupting it, so that I could close the file in the first notebook before trying to save it in the second notebook. I think not letting you modify a file while it is opened as basked somewhere else is the safest option. Any update on this? yet another of my adatas was deleted this way. @ivirshup Maybe a temporary solution would be info on how to close a backed file - now I usually just copy info from adata and delete the object immediately, but if there was a "close" function that would be nice to know. @ivirshup any updates on this? Alternatively, is there a good way to check if anndata file is open somewhere before saving? @Hrovatin, no updates yet. But reading this with fresh eyes, I might have something. What do you mean by corrupted? Is the data not just overwritten? Or does it fail to overwrite somehow? Looking some more and found an old issue I opened: https://github.com/h5py/h5py/issues/1864 It looks like this is basically an upstream bug (link to jira tracker, a little difficult to use), and it's unclear if/ when it will be fixed. It looks like there isn't a clean solution from python. One thing that could work is if the backed file was opened with h5py.File(..., locking=False). Then the file should just be overwritten. I'm not sure if this is a good default, since it changes behavior from hdf5. Ideally we would error before any data gets truncated. However, we could allow passing that argument through when reading a file in backed mode. Whatever you can do to prevent deletion of files would be very useful for me, especially now when we no longer have snapshots on the server. Because sometimes I forget to shut down a notebook where I have something backed opened and then data is gone. If you want to just prevent deletion in general, we could let you pass "w-" or "x" to write_h5ad. But this is specifically if you don't want to overwrite data regardless of whether any other process has it open. No, I would like to overwrite files, but prevent that they get corrupted if another process has it open. - I often update adatas so I want to change (rewrite) existing object on disk. Seems like this is a valid use case, and we’re just not clear how to enable it. Or is something still unclear about this?
gharchive/issue
2021-03-21T20:18:56
2025-04-01T06:45:45.071437
{ "authors": [ "Hrovatin", "flying-sheep", "ivirshup" ], "repo": "scverse/anndata", "url": "https://github.com/scverse/anndata/issues/522", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1262764397
suddenly unable to read .h5ad file there were no problem reading .h5ad file but all of the sudden, this error came up and unable to read .h5ad files,, someone have any idea?? Thank you!! It was probably written by a newer anndata version. Try updating anndata, maybe that helps. We have a PR that improves error messages in this regard. @ivirshup could you please merge or adapt and then merge #734? Thanks for your help! I have updated anndata and h5py and re-start my jupyter noteobok and it works!!
gharchive/issue
2022-06-07T06:02:05
2025-04-01T06:45:45.074384
{ "authors": [ "flying-sheep", "parkjooyoung99" ], "repo": "scverse/anndata", "url": "https://github.com/scverse/anndata/issues/780", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2732938152
Error in import squidpy: cannot import name 'SparseCSCView' from 'anndata._core.views' Description Hi, I'm trying to use squidpy but when I try to import it returns an error. It seems to search for a function in anndata which I couldn't find. I might just have misunderstood the error, and if that's the case, I apologize. Could you please help me understand or resolve this issue? Thanks in advance ... Minimal reproducible example #the package was downloaded from conda with other packages like sainsc. I'm running the code in this environment on Spyder 6. Anndata version=0.11.1 import squidpy as sq ... Traceback ImportError Traceback (most recent call last) Cell In[18], line 1 ----> 1 import squidpy as sq File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/__init__.py:5 1 from __future__ import annotations 3 from importlib import metadata ----> 5 from squidpy import datasets, gr, im, pl, read, tl 7 try: 8 md = metadata.metadata(__name__) File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/gr/__init__.py:5 1 """The graph module.""" 3 from __future__ import annotations ----> 5 from squidpy.gr._build import spatial_neighbors 6 from squidpy.gr._ligrec import ligrec 7 from squidpy.gr._nhood import centrality_scores, interaction_matrix, nhood_enrichment File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/gr/_build.py:31 29 from squidpy._docs import d, inject_docs 30 from squidpy._utils import NDArrayA ---> 31 from squidpy.gr._utils import ( 32 _assert_categorical_obs, 33 _assert_positive, 34 _assert_spatial_basis, 35 _save_data, 36 ) 38 __all__ = ["spatial_neighbors"] 41 @d.dedent 42 @inject_docs(t=Transform, c=CoordType) 43 def spatial_neighbors( (...) 56 copy: bool = False, 57 ) -> tuple[csr_matrix, csr_matrix] | None: File ~/anaconda3/envs/sainsc/lib/python3.12/site-packages/squidpy/gr/_utils.py:12 10 import pandas as pd 11 from anndata import AnnData ---> 12 from anndata._core.views import ArrayView, SparseCSCView, SparseCSRView 13 from anndata.utils import make_index_unique 14 from pandas import CategoricalDtype ImportError: cannot import name 'SparseCSCView' from 'anndata._core.views' (/home/onco-lab/anaconda3/envs/sainsc/lib/python3.12/site-packages/anndata/_core/views.py) ... Version Name Version Build Channel squidpy 1.5.0 pyhd8ed1ab_0 conda-forge Anndata version=0.11.1 ... https://github.com/scverse/squidpy/issues/915 I also had the same issue and could fix it by installing version 1.6.2.
gharchive/issue
2024-12-11T13:08:39
2025-04-01T06:45:45.080941
{ "authors": [ "HeesooSong", "diego-iac" ], "repo": "scverse/squidpy", "url": "https://github.com/scverse/squidpy/issues/928", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2638467631
scylla_node: ignore scylla-tools files if scylla-tools is missing This is a follow-up to https://github.com/scylladb/scylla-ccm/pull/619. Patch scylla_node.py so every access to java tools dir is conditional on the dir existing in the first place. With this PR, tests are passing with the java tools dropped form the unified package (https://github.com/scylladb/scylladb/pull/20740). Refs: https://github.com/scylladb/scylladb/issues/14856 @denesb are you sure you tested ontop of #619 ? since it was merged 4 hours ago, and you have a conflict with it. @denesb are you sure you tested ontop of #619 ? since it was merged 4 hours ago, and you have a conflict with it. Ah, seems to be the same thing. I guess my PR is not needed after all. Closing.
gharchive/pull-request
2024-11-06T15:39:20
2025-04-01T06:45:45.093034
{ "authors": [ "denesb", "fruch" ], "repo": "scylladb/scylla-ccm", "url": "https://github.com/scylladb/scylla-ccm/pull/623", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1425590323
Update files.rst Fixing typo Muchas gracias por la corrección! Saludos
gharchive/pull-request
2022-10-27T12:53:55
2025-04-01T06:45:45.281855
{ "authors": [ "quobit", "sdelquin" ], "repo": "sdelquin/aprendepython", "url": "https://github.com/sdelquin/aprendepython/pull/32", "license": "CC-BY-4.0", "license_type": "permissive", "license_source": "github-api" }
182256277
commit fails after file rename colemanbroaddus@myers-mac-10-display-1 ~/P/cell_tracker> gl status On branch master, repo-directory // Tracked files with modifications: ➜ these will be automatically considered for commit ➜ use gl untrack f if you don't want to track changes to file f ➜ if file f was committed before, use gl checkout f to discard local changes master.py truetree.py (new file) viewer.py (deleted) Untracked files: ➜ these won't be considered for commit ➜ use gl track f if you want to track changes to file f There are no untracked files to list colemanbroaddus@myers-mac-10-display-1 ~/P/cell_tracker> gl commit -m "Merged changes from nodeless." ✘ Some internal error occurred ➜ If you want to help, see http://gitless.com for info on how to report bugs and include the following information: 0.8.3 Traceback (most recent call last): File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.cli.gl", line 69, in main File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.cli.gl_commit", line 58, in main File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.core", line 1179, in create_commit File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.core", line 1152, in get_tree_and_update_index File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/gitless.core", line 1145, in update File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/pygit2.index", line 168, in remove File "/Users/santiago/Documents/code/gitless/build/gl/out00-PYZ.pyz/pygit2.errors", line 53, in check_error IOError: Index does not contain viewer.py at stage 0 This happened directly after a merge and resolve, although they did not include the file causing the error: (viewer.py renamed to truetree.py).
gharchive/issue
2016-10-11T12:54:59
2025-04-01T06:45:45.294422
{ "authors": [ "colemanbroad" ], "repo": "sdg-mit/gitless", "url": "https://github.com/sdg-mit/gitless/issues/63", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
725377654
Changing name and website In Competitive programming list,the name Code Fights is was changed to Code Signal in 2018.So I changed the website and name of CodeFights to CodeSignal. Thanks
gharchive/pull-request
2020-10-20T09:13:49
2025-04-01T06:45:45.305557
{ "authors": [ "nyilynnhtwe", "sdmg15" ], "repo": "sdmg15/Best-websites-a-programmer-should-visit", "url": "https://github.com/sdmg15/Best-websites-a-programmer-should-visit/pull/618", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
474405694
add images add images Please add some more information about the problem or feature request
gharchive/issue
2019-07-30T06:51:14
2025-04-01T06:45:45.314133
{ "authors": [ "Saileek-bidchat", "max-carroll" ], "repo": "sdoomz/react-google-picker", "url": "https://github.com/sdoomz/react-google-picker/issues/20", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
346998052
Make it possible to ignore usage of invalid SSL certificates Since #15 was closed the world has gotten better, with the merge of https://github.com/sfackler/rust-native-tls/pull/56 Since the native_tl crate now support this, it should be easy to implement it in reqwest. This probably requires bumping the native-tls crate to 0.2.0 from 0.1.5 though. Could we help to implement this feature? Certainly! To get this into 0.8.x, it'd probably be easiest to just inline the code that reqwest uses from hyper-tls and tokio-tls (since tokio-tls is stagnating/abandoned, and the code duplicate code wouldn't be much), and update it to work with native-tls 0.2. I started the #325 PR. Any policy to implement test for this feature? This feature is part of the v0.9 release.
gharchive/issue
2018-08-02T12:51:51
2025-04-01T06:45:45.398616
{ "authors": [ "adamtulinius", "seanmonstar", "yageek" ], "repo": "seanmonstar/reqwest", "url": "https://github.com/seanmonstar/reqwest/issues/324", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
477474073
Algorithm runs exceptionally slow With addition of 2 children per crossover, alg runs insanely slow. Looking into issue now Selection was chunking parents every 1 step, doubling the number of parents necessary. Fixed with last commit.
gharchive/issue
2019-08-06T16:17:50
2025-04-01T06:45:45.399592
{ "authors": [ "seanmor5" ], "repo": "seanmor5/genex", "url": "https://github.com/seanmor5/genex/issues/17", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
536783457
implementation DatasourceProxy without AlibabaDruid Why you need it? DatasourceProxy design reason is adapt to more difference JDBC Datasource implementation(Hikaricp, c3p0...) How it could be? io.seata.rm.datasource.sql.SQLVisitorFactory using AlibabaDruid SQLStatement, We can use JSqlParser or custom SqlUtils. refer: https://github.com/seata/seata/pull/1703 after 1.0 Thanks! @funcfoo Seat just uses druid internally as sqlparser, and the database connection pool used by the business can be arbitrary. Yes, I know that. @slievrly Yes, I know that. but I can't agree with this design. We just use a little library, Why using a Heavyweight library.
gharchive/issue
2019-12-12T06:29:47
2025-04-01T06:45:45.405356
{ "authors": [ "funcfoo", "slievrly", "zjinlei" ], "repo": "seata/seata", "url": "https://github.com/seata/seata/issues/2025", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1082074566
optimize:simplify the interface of console [ ] I have registered the PR changes. Ⅰ. Describe what this PR did 简化控制台接口数量,提高拓展性: 1.控制台接口合并,使用对象作为入参; 2.防止查询条件变更时,要增加接口提供拓展; 老接口: PageResult<GlobalLockVO> queryByTable(String tableName); PageResult<GlobalLockVO> queryByXid(String xid); 新接口: PageResult<GlobalLockVO> query(GlobalLockParam param); public class GlobalLockParam implements Serializable { private static final long serialVersionUID = 615412528070131284L; private String xid; private String tableName; } Ⅱ. Does this pull request fix one issue? Ⅲ. Why don't you add test cases (unit test/integration test)? Ⅳ. Describe how to verify it Ⅴ. Special notes for reviews Codecov Report Merging #4212 (d3b7568) into develop (caa8443) will decrease coverage by 0.04%. The diff coverage is 0.00%. :exclamation: Current head d3b7568 differs from pull request most recent head 9d7a43b. Consider uploading reports for the commit 9d7a43b to get more accurate results @@ Coverage Diff @@ ## develop #4212 +/- ## ============================================= - Coverage 49.10% 49.06% -0.05% + Complexity 3774 3771 -3 ============================================= Files 719 722 +3 Lines 23948 23959 +11 Branches 2926 2926 ============================================= - Hits 11760 11755 -5 - Misses 10987 11000 +13 - Partials 1201 1204 +3 Impacted Files Coverage Δ ...o/seata/core/console/param/BranchSessionParam.java 0.00% <0.00%> (ø) ...a/io/seata/core/console/param/GlobalLockParam.java 0.00% <0.00%> (ø) ...o/seata/core/console/param/GlobalSessionParam.java 0.00% <0.00%> (ø) ...java/io/seata/core/console/vo/BranchSessionVO.java 0.00% <ø> (ø) ...in/java/io/seata/core/console/vo/GlobalLockVO.java 0.00% <ø> (ø) ...java/io/seata/core/console/vo/GlobalSessionVO.java 0.00% <ø> (ø) ...er/console/controller/BranchSessionController.java 50.00% <ø> (ø) ...erver/console/controller/GlobalLockController.java 50.00% <0.00%> (ø) ...er/console/controller/GlobalSessionController.java 33.33% <0.00%> (+8.33%) :arrow_up: ...er/console/impl/db/BranchSessionDBServiceImpl.java 0.00% <ø> (ø) ... and 11 more 改动点: 1.vo添加tostring; 2.部分类添加序列化; 3.接口合并; 4.result类移动位置; 5.分页结果pageResult部分无用方法删除;
gharchive/pull-request
2021-12-16T11:21:39
2025-04-01T06:45:45.422566
{ "authors": [ "codecov-commenter", "lightClouds917" ], "repo": "seata/seata", "url": "https://github.com/seata/seata/pull/4212", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2159152510
Update ChartRendererConfigOptions interface colors Update ChartRendererConfigOptions interface colors accordingly https://docs.seats.io/docs/renderer/colors Thanks for creating this pull request, we missed the type update for this one.
gharchive/pull-request
2024-02-28T14:50:39
2025-04-01T06:45:45.437581
{ "authors": [ "RafaelAlkmimDias", "mortendevold" ], "repo": "seatsio/seatsio-types", "url": "https://github.com/seatsio/seatsio-types/pull/8", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1126522786
Don't include screens without questions in total block count To determine if an application is ok to submit or requires more answers from the applicant, we count the number of active blocks and compare that with the number of active and completed blocks. This caused a bug where blocks that have no questions would count towards the total of active but incomplete blocks, making it impossible to complete the application. This change alters the logic for tallying completed blocks by not including blocks that have no questions in the tally. closing in favor of an alternative approach: https://github.com/seattle-uat/civiform/pull/1894
gharchive/pull-request
2022-02-07T21:33:38
2025-04-01T06:45:45.439412
{ "authors": [ "bion" ], "repo": "seattle-uat/civiform", "url": "https://github.com/seattle-uat/civiform/pull/1893", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
806870904
The ability to bring up a full dev stack which will automatically tear itself down. Description There is one TODO - the actual ARN for the role that needs to bring up the stack needs to change, or else the perms for that role need to be changed. Waiting on John for that. So I figured I'd send this for review and then edit that before submitting. Let's see if that testing github action works! (Will add reviewers once I see whether that needs changes) Checklist [ ] Created tests which fail without the change (if possible) [ ] Extended the README / documentation, if necessary Cool - I think that's right.
gharchive/pull-request
2021-02-12T01:10:40
2025-04-01T06:45:45.441382
{ "authors": [ "ndmckinley" ], "repo": "seattle-uat/universal-application-tool", "url": "https://github.com/seattle-uat/universal-application-tool/pull/144", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2502228010
updated modal related files to make sure the double scroll bar will n… fix: modal double scroll bar issue fixed Continued in #1572
gharchive/pull-request
2024-09-03T08:20:13
2025-04-01T06:45:45.452156
{ "authors": [ "GauriKudtarkar", "splashdust" ], "repo": "seb-oss/green", "url": "https://github.com/seb-oss/green/pull/1553", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2095106296
Make templates able to access instVars of the object sent as rendering context Make the syntax of the template to be cleaner. For example, given the following model: model := STTDummyJobModel new firstName: 'Questioner'; answer: 42; yourself. Instead of <div><st= self firstName, ' ', self answer asString></div> It should work if you don't use the accessors too, like in: <div><st= firstName, ' ', answer asString></div> PR here https://github.com/sebastianconcept/STTemplate/pull/10 Merged
gharchive/issue
2024-01-23T02:12:44
2025-04-01T06:45:45.468761
{ "authors": [ "sebastianconcept" ], "repo": "sebastianconcept/STTemplate", "url": "https://github.com/sebastianconcept/STTemplate/issues/9", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
874353209
Error loading module 'tabline' Hey seblj I install this plugin throw vim-plug After sourcing the lines in my init.vim require('tabline').setup{ no_name = '[No Name]', -- Name for buffers with no name modified_icon = '', -- Icon for showing modified buffer close_icon = '', -- Icon for closing tab with mouse separator = "▌", -- Separator icon on the left side space = 3, -- Prefix and suffix space color_all_icons = false, -- Color devicons in active and inactive tabs always_show_tabs = false, -- Always show tabline right_separator = false, -- Show right separator on the last tab -- highlights = { -- Default colors for highlights. -- fill = { -- guifg = M.inactive_background, -- guibg = M.inactive_background -- }, -- filename_active = { -- guifg = M.active_text, -- guibg = M.active_background, -- gui = 'bold,italic' -- }, -- filename_inactive = { -- guifg = M.inactive_text, -- guibg = M.inactive_background -- }, -- padding_active = { -- guifg = M.active_background, -- guibg = M.active_background -- }, -- padding_inactive = { -- guif = M.inactive_background, -- guibg = M.inactive_background -- }, -- separator_active = { -- guifg = M.active_separator, -- guibg = M.active_background -- }, -- separator_inactive = { -- guifg = M.inactive_text, -- guibg = M.inactive_background -- }, -- modified_active = { -- guifg = M.active_text, -- guibg = M.active_background -- }, -- modified_inactive = { -- guifg = M.inactive_text, -- guibg = M.inactive_background -- }, -- close_active = { -- guifg = M.active_text, -- guibg = M.active_background -- }, -- close_inactive = { -- guifg = M.inactive_text, -- guibg = M.inactive_background -- } -- } } I think the problem is with how you require things. Do you have a file called tabline.lua where you setup the plugin? If so, lua gets confused on which file to require since your config file, and my plugin is called the same thing. You should try to add another folder inside the lua folder (for example named config). Then you can do require('config.tabline') from your init.vim. Your path to your config for the plugin will then be: /home/va/.config/nvim/lua/config/tabline.lua Ok that fixed the ISSUE 👍🏻
gharchive/issue
2021-05-03T08:26:31
2025-04-01T06:45:45.483762
{ "authors": [ "seblj", "terminal-guy" ], "repo": "seblj/nvim-tabline", "url": "https://github.com/seblj/nvim-tabline/issues/4", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
394861824
Unsupported "Accept" header I have run the app in env test. I have attached env file too. I get the below error env.txt You forgot to specify Accept in your request headers. It's supposed to equal "application/json" You'll also need to pass Content-Type with the same value how do I use Accep in header
gharchive/issue
2018-12-30T10:31:51
2025-04-01T06:45:45.517994
{ "authors": [ "hlogeon", "thomsonimjd" ], "repo": "secret-tech/backend-ico-dashboard", "url": "https://github.com/secret-tech/backend-ico-dashboard/issues/193", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2243298845
【文档】真是部署中,数据是怎么传输的 Hi 我在看文档使用SPU进行逻辑回归。 我理解,该文档对应的是单机演示环境的。那我有如下几个问题, 假设两方Alice(192.168.1.2 有x_1) 和Bob(192.168.1.3 有x_2 和y) 该文档的代码,是在Alice 执行还是在Bob执行 假设是在Alice执行,那Bob的 x_2 和y 如何传递给Alice(或Alice的代码应该做怎样的调整?)即要在Alice的机器上执行什么操作,在Bob的机器上执行什么操作? Reproduction code to reproduce the issue. 1 请查看这个文档
gharchive/issue
2024-04-15T10:29:22
2025-04-01T06:45:45.520636
{ "authors": [ "anakinxc", "gxcuit" ], "repo": "secretflow/secretflow", "url": "https://github.com/secretflow/secretflow/issues/1251", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1732274507
Update awesome-pets.md Added my github id as a contributor to the zkp branch. Thank you for your approval ! Hi, would you mind to paste your merged PR here? Just to make sure everything is on track. Thanks @zlyber @Jamie-Cui https://github.com/secretflow/secretflow/pull/597 Thanks @xfap , I'm merge this PR
gharchive/pull-request
2023-05-30T13:24:26
2025-04-01T06:45:45.522667
{ "authors": [ "Jamie-Cui", "xfap", "zlyber" ], "repo": "secretflow/secretflow", "url": "https://github.com/secretflow/secretflow/pull/602", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
832400094
Angular 11 - JWT Authentication Example & Tutorial With PHP Article checklist NOTE: (Please ensure that you have only one open issue + linked pull request at a time. This will ensure that we complete the article in a timely manner from inception to publishing.) Preliminary Checks - Formatting and Structure [x] Does your article follow any of the suggested structure formats? - see example formats [x] Is your article properly formatted in Markdown? - see Github Markdown guide [x] Have you used the correct folder and file structure? - see Contribution Guidelines [x] Is your article over 750 words? - Tool: Wordcounter.net [x] Have you used our preferred heading structure? - H3 (###) and up [x] Does your article provide enough value and detail about your topic? - Articles should be clear, accurate, and fully explained. [x] Can your article be understood by beginners? - Assume the audience is smart but has no prior exposure to the common terminology in your article. [x] Have you included a hero image that is 600x400 pixels and have the copyright to use it? - Tools: Pixlr Image Editor for resizing and Unsplash for Creative Commons images [x] In your PR - add a comment that states "This fixes #(enter your Idea Content Form #)" to link and close your Idea Content (issue) when the PR is merged. Grammar & plagiarism checks [x] Have you spellchecked and grammar checked your article? - Tools: Write&Improve - ProWritingAid - Grammarly [x] Have you placed your article through a 3rd party plagiarism checker? We suggest using Quetext, this tool is free and has a daily limit. We typically accept articles with 10% or less. [x] Have you checked your article for readability? - Tool: Hemmingway [x] Have you added sources for quotes and images that aren't yours? Technical checks [x] Are your code snippets properly formatted for syntax highlighting - see Syntax guide Contribution guidelines For first-time contributors and for more details, see Contributing Guidelines This fixes #1691 Hello @jumamiller. The title of this article conflicts with that of another already published article. See https://www.section.io/engineering-education/getting-started-with-jwt-using-angular8-and-nodejs/. The article talks about jwt in angular 11 using a node.js server. Looking at your article, seems like there's some PHP used. Just a suggestion, maybe you could update the title to include PHP, to differentiate it from the other article. @paulodhiambo, done! @jumamiller an amazing article on Angular 11 authentication. @hectorkambow this article is ready for final review. @jumamiller 👋 Good afternoon - Upon running your article through our 3rd party plagiarism checker it seemed to raise a few flags and the % was higher than we typically accept. Please see attached PDF - and revisit the article to ensure we are contributing wholly unique and original content. Be sure to see our resources page to see more info on plagiarism and what is considered as such. 🤔 Angular 11 - JWT Authentication Example & Tutorial With PHP #1699.pdf @hectorkambow, I've updated most of the boilerplate codes, you can recheck! cc @paulodhiambo
gharchive/pull-request
2021-03-16T04:32:44
2025-04-01T06:45:45.537000
{ "authors": [ "geoffrey45", "hectorkambow", "jumamiller", "paulodhiambo" ], "repo": "section-engineering-education/engineering-education", "url": "https://github.com/section-engineering-education/engineering-education/pull/1699", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1070455479
Creating Responsive Line and Bar charts with React and Semantic-UI Article checklist Pre-submission checks Please ensure that you have only one open topic suggestion (issue) + in-review article (linked pull request) at a time. This will ensure that we complete the review process in a timely manner from inception to publishing If you have not already, please go over our Resources Page for more tips and tricks on how to improve your overall technical writing so reviews are swifter, increase the chance of a payout and provide more value to our readers. For first-time contributors and for more details on our submission guidelines, see our Contributing Guidelines. Formatting and structure checks [x] Have you used the correct folder and file structure? - see Contribution Guidelines [x] Is your article properly formatted in Markdown? - see Github Markdown guide [x] Have you used our preferred heading structure? - H3 (###) and up [x] Does your article follow any of the suggested structure formats? - see example formats [x] Is your article over 750 words? - Tool: Wordcounter.net [x] Is the article you are submitting an in-depth and unique article? Does it go beyond what is in the official docs and what is covered in other blog sites. See these articles as examples. [x] Does your article provide enough value and detail about your topic? - Articles should be clear, accurate, and fully explained. [x] Can your article be understood by beginners? - Assume the audience is smart but has no prior exposure to the common terminology in your article. [x] Have you included a hero image that is 600x400 pixels and have the copyright to use it? - Tools: Pixlr Image Editor for resizing and Unsplash for Creative Commons images Grammar & plagiarism checks [x] Have you spellchecked and grammar checked your article? - Tools: Write&Improve - ProWritingAid - Grammarly [x] Please place your article through a 3rd party plagiarism checker. We suggest using Quetext, this tool is free and has a daily limit. We typically accept articles with 10% or less. [x] Have you checked your article for readability? - Tool: Hemmingway [x] Have you added sources for quotes and images that aren't yours? Technical checks [x] Are your code snippets properly formatted for syntax highlighting - see Syntax guide [x] Have you checked your code runs correctly and you've highlighted all necessary dependencies for installation? [x] Are the software programs and packages you're highlighting in your article up to date, using current versions and not deprecated? Topic suggestion this closes Remove the backticks and add the issue number below to link and close your Topic Suggestion (issue) when your article has been published (PR has been merged). See this video for more details. This closes #4724 Finally, delete the article checklist notes in blockquotes and submit your PR. We look forward to reviewing your article. hello @lalith1403, @jerimkaura I have made the required changes to the previous PR #4944. please proceed with the final review thank you cc. @hectorkambow @jerimkaura, @lalith1403 I'm yet to receive any response from you. #4944 was one commit away from being merged. Thank you @srishilesh @jerimkaura is this article ready for a final review? Take in mind that (generally) any duplicate article should NOT require any additional revisions - & should be fully ready to publish. If any more (major) changes, revisions, or edits are required we will be closing the PR indefinitely. cc @FredBenson Thank you @srishilesh @jerimkaura is this article ready for a final review? Take in mind that (generally) any duplicate article should NOT require any additional revisions - & should be fully ready to publish. If any more (major) changes, revisions, or edits are required we will be closing the PR indefinitely. cc @FredBenson Yes the article is ready for final review. @FredBenson Please reach out via Slack DM 👍
gharchive/pull-request
2021-12-03T10:43:48
2025-04-01T06:45:45.554156
{ "authors": [ "FredBenson", "hectorkambow", "jerimkaura" ], "repo": "section-engineering-education/engineering-education", "url": "https://github.com/section-engineering-education/engineering-education/pull/5366", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1394806607
Update readme with About + links to the main technologies used. Make it more searchable. Done
gharchive/issue
2022-10-03T14:09:54
2025-04-01T06:45:45.607004
{ "authors": [ "UberMeatShield" ], "repo": "secureworks/azure_auth", "url": "https://github.com/secureworks/azure_auth/issues/1", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
211930910
How to make it work with Fish? I've tried eval (dircolors ~/.dir_colors), replacing backsticks with parentheses but it didn't help. Fish outputs error. I've also tried to manually set and export LS_COLORS environment variable, but also without results. The string is incredibly long. Please help with fish shell, the standard command for exporting variable is set -x, and there's no equal (=) character when setting the value (set var value is enough) unlike var=value;export $var of zsh and bash. I could finally solve it. set -x LS_COLORS "no=00;38;5;244:rs=0:di=00;38;5;33:ln=00;38;5;37:mh=00:pi=48;5;230;38;5;136;01:so=48;5;230;38;5;136;01:do=48;5;230;38;5;136;01:bd=48;5;230;38;5;244;01:cd=48;5;230;38;5;244;01:or=48;5;235;38;5;160:su=48;5;160;38;5;230:sg=48;5;136;38;5;230:ca=30;41:tw=48;5;64;38;5;230:ow=48;5;235;38;5;33:st=48;5;33;38;5;230:ex=00;38;5;64:*.tar=00;38;5;61:*.tgz=00;38;5;61:*.arj=00;38;5;61:*.taz=00;38;5;61:*.lzh=00;38;5;61:*.lzma=00;38;5;61:*.tlz=00;38;5;61:*.txz=00;38;5;61:*.zip=00;38;5;61:*.z=00;38;5;61:*.Z=00;38;5;61:*.dz=00;38;5;61:*.gz=00;38;5;61:*.lz=00;38;5;61:*.xz=00;38;5;61:*.bz2=00;38;5;61:*.bz=00;38;5;61:*.tbz=00;38;5;61:*.tbz2=00;38;5;61:*.tz=00;38;5;61:*.deb=00;38;5;61:*.rpm=00;38;5;61:*.jar=00;38;5;61:*.rar=00;38;5;61:*.ace=00;38;5;61:*.zoo=00;38;5;61:*.cpio=00;38;5;61:*.7z=00;38;5;61:*.rz=00;38;5;61:*.apk=00;38;5;61:*.gem=00;38;5;61:*.jpg=00;38;5;136:*.JPG=00;38;5;136:*.jpeg=00;38;5;136:*.gif=00;38;5;136:*.bmp=00;38;5;136:*.pbm=00;38;5;136:*.pgm=00;38;5;136:*.ppm=00;38;5;136:*.tga=00;38;5;136:*.xbm=00;38;5;136:*.xpm=00;38;5;136:*.tif=00;38;5;136:*.tiff=00;38;5;136:*.png=00;38;5;136:*.PNG=00;38;5;136:*.svg=00;38;5;136:*.svgz=00;38;5;136:*.mng=00;38;5;136:*.pcx=00;38;5;136:*.dl=00;38;5;136:*.xcf=00;38;5;136:*.xwd=00;38;5;136:*.yuv=00;38;5;136:*.cgm=00;38;5;136:*.emf=00;38;5;136:*.eps=00;38;5;136:*.CR2=00;38;5;136:*.ico=00;38;5;136:*.tex=00;38;5;245:*.rdf=00;38;5;245:*.owl=00;38;5;245:*.n3=00;38;5;245:*.ttl=00;38;5;245:*.nt=00;38;5;245:*.torrent=00;38;5;245:*.xml=00;38;5;245:*Makefile=00;38;5;245:*Rakefile=00;38;5;245:*Dockerfile=00;38;5;245:*build.xml=00;38;5;245:*rc=00;38;5;245:*1=00;38;5;245:*.nfo=00;38;5;245:*README=00;38;5;245:*README.txt=00;38;5;245:*readme.txt=00;38;5;245:*.md=00;38;5;245:*README.markdown=00;38;5;245:*.ini=00;38;5;245:*.yml=00;38;5;245:*.cfg=00;38;5;245:*.conf=00;38;5;245:*.h=00;38;5;245:*.hpp=00;38;5;245:*.c=00;38;5;245:*.cpp=00;38;5;245:*.cxx=00;38;5;245:*.cc=00;38;5;245:*.objc=00;38;5;245:*.sqlite=00;38;5;245:*.go=00;38;5;245:*.sql=00;38;5;245:*.csv=00;38;5;245:*.log=00;38;5;240:*.bak=00;38;5;240:*.aux=00;38;5;240:*.lof=00;38;5;240:*.lol=00;38;5;240:*.lot=00;38;5;240:*.out=00;38;5;240:*.toc=00;38;5;240:*.bbl=00;38;5;240:*.blg=00;38;5;240:*~=00;38;5;240:*#=00;38;5;240:*.part=00;38;5;240:*.incomplete=00;38;5;240:*.swp=00;38;5;240:*.tmp=00;38;5;240:*.temp=00;38;5;240:*.o=00;38;5;240:*.pyc=00;38;5;240:*.class=00;38;5;240:*.cache=00;38;5;240:*.aac=00;38;5;166:*.au=00;38;5;166:*.flac=00;38;5;166:*.mid=00;38;5;166:*.midi=00;38;5;166:*.mka=00;38;5;166:*.mp3=00;38;5;166:*.mpc=00;38;5;166:*.ogg=00;38;5;166:*.opus=00;38;5;166:*.ra=00;38;5;166:*.wav=00;38;5;166:*.m4a=00;38;5;166:*.axa=00;38;5;166:*.oga=00;38;5;166:*.spx=00;38;5;166:*.xspf=00;38;5;166:*.mov=00;38;5;166:*.MOV=00;38;5;166:*.mpg=00;38;5;166:*.mpeg=00;38;5;166:*.m2v=00;38;5;166:*.mkv=00;38;5;166:*.ogm=00;38;5;166:*.mp4=00;38;5;166:*.m4v=00;38;5;166:*.mp4v=00;38;5;166:*.vob=00;38;5;166:*.qt=00;38;5;166:*.nuv=00;38;5;166:*.wmv=00;38;5;166:*.asf=00;38;5;166:*.rm=00;38;5;166:*.rmvb=00;38;5;166:*.flc=00;38;5;166:*.avi=00;38;5;166:*.fli=00;38;5;166:*.flv=00;38;5;166:*.gl=00;38;5;166:*.m2ts=00;38;5;166:*.divx=00;38;5;166:*.webm=00;38;5;166:*.axv=00;38;5;166:*.anx=00;38;5;166:*.ogv=00;38;5;166:*.ogx=00;38;5;166:" This long string should be put inside ~/.config/config.fish. Just to make sure: dircolors command, based on dircolors.256dark, generate value for LS_COLORS environment variable which contains color codes for ls command. Thanks for the great colorscheme, the issue is solved. This works as of the current version of FISH: eval (dircolors -c LS_COLORS) In fish 3.1.2, I neede to use this: eval (dircolors -c ~/.dircolors)
gharchive/issue
2017-03-05T05:11:13
2025-04-01T06:45:45.641178
{ "authors": [ "Chinggis6", "blinry", "duganchen" ], "repo": "seebi/dircolors-solarized", "url": "https://github.com/seebi/dircolors-solarized/issues/76", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2209457030
Move table name and description creation into project-id-metadata-create Right now the table names and descriptions are created in the project-id-view.html as a dialog box. I think it makes more sense to have that creation in the project id metadata create. It would simplify things, like not needing to pass around the table metadata object and it would make it so the create URL is more consistent with our naming scheme. Right now, the URL is project id metadata id create which feels a bit inconsistent, since in order to have an id, we need to create it first, but in this case we have an id as we're creating it... This is on hold until the team has agreed on the design, see issue #397 We decided to start the web app from scratch after the major design/architecture decisions and changes we made. No longer relevant.
gharchive/issue
2024-03-26T22:29:54
2025-04-01T06:45:45.646284
{ "authors": [ "lwjohnst86", "signekb" ], "repo": "seedcase-project/seedcase-sprout", "url": "https://github.com/seedcase-project/seedcase-sprout/issues/341", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
138823853
PhoneNumberOfflineGeocoder ? Hello, i'm trying to get the geo info for a number but cant seen to find the PhoneMetadataCollection function/values ?? i even searched the source but not luck :fearful: Thanks Unfortunately, PhoneNumberOfflineGeocoder has not been ported to javascript so don't expect to see this functionality on this wrapper package any time soon. You might want to submit an issue on the official repository. :disappointed: ... will just write my own geo location function i guess Thanks You could submit a PR with the port to their repository :) @Ahmed-90 did you find/write anything? I can't even seem to find an NPM package that maps USA numbers to the state. I scraped wikipedia for North America ... I saved the code here: https://gist.github.com/avimar/bcc7a8159d214c51b5496be5f5e14362 I asked for it here, apparently it's a leet request... https://github.com/googlei18n/libphonenumber/issues/1337 @avimar yeah.. well kindof.. anyway you can find the geo data in the resources https://github.com/googlei18n/libphonenumber/tree/master/resources/geocoding/en sorted by the geo code also the nanpa has public listings of the codes -> https://www.nationalnanpa.com/reports/area_code_relief_planning.html @avimar i just used my own lookup without the lib so .. simply converted the data to json object (keyBy areacode) and just did a lookup when i can't identify the number not the cleanest way :P but worked fine Is there any update on this?? Can i use geo-location and timezone mapper facility with Javascript/Node?? If so please reply... @sahilkanani1812, see original comment https://github.com/seegno/google-libphonenumber/issues/46#issuecomment-192973464. It's still valid.
gharchive/issue
2016-03-06T19:03:42
2025-04-01T06:45:45.652970
{ "authors": [ "Ahmed-90", "avimar", "ruimarinho", "sahilkanani1812" ], "repo": "seegno/google-libphonenumber", "url": "https://github.com/seegno/google-libphonenumber/issues/46", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
334315951
feat(disableAsideGap): Adding new props for Aside ** Please provide as much detail as possible, but feel free to remove irrelevant sections ** In the split mode it do not need the space in between. At the same time doing come clean up in showmore component as well BREAKING CHANGE: adding a new props RFC URL: none MIGRATION GUIDE: After: <AsidedLayout disableGapInBetween renderAside={() => {...}} size="30%" > EXAMPLE USAGE: <AsidedLayout disableGapInBetween renderAside={() => {...}} size="30%" > <Card> <Section> <Text shouting> Main Content </Text> <Text> This card is provided as children. </Text> </Section> </Card> <Card> <Section> <Text shouting> Another Card </Text> <Text> Here's another card for good measure. </Text> </Section> </Card> </AsidedLayout> Try not to merge with only one approving review
gharchive/pull-request
2018-06-21T02:34:55
2025-04-01T06:45:45.657687
{ "authors": [ "ChanJit", "PepeFranco" ], "repo": "seekinternational/seek-asia-style-guide", "url": "https://github.com/seekinternational/seek-asia-style-guide/pull/139", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2300093385
线条插槽不能使用-- 并且使用了ovUseLineSlot属性 relation-graph@2.1.42 vue@2.6.10 使用 线条插槽 组件里复制的dome <template #line="{ line, link, lineIndex }"> 线条不生效 报错 Cannot read properties of undefined (reading 'getArrow') 说明 relationGraph 为undefined relation-graph@2.1.x中在自定义节点、线条插槽中通过依赖注入获取graphInstance的方式发生了变化: 以前是:通过依赖的对象graph获取:graph.instance 现在是:通过依赖的方法graphInstance来获取:graphInstance() 你可以将你的代码以下内容按照下图修改一下就可以了: 相关的示例: https://www.relation-graph.com/#/demo/vue2?id=adv-line-slot2
gharchive/issue
2024-05-16T11:02:45
2025-04-01T06:45:45.661071
{ "authors": [ "ZhouJing0", "seeksdream" ], "repo": "seeksdream/relation-graph", "url": "https://github.com/seeksdream/relation-graph/issues/309", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2704448650
Dependency lib can upgrade: System.Text.Json Here is the detail. https://github.com/segmentio/Serialization.NET/issues/16 hi @arlyxiao, thanks for reporting this. we will get it updated. in the meantime, you should be able to override the dependency version by directly referencing it. see here
gharchive/issue
2024-11-29T09:08:19
2025-04-01T06:45:45.666881
{ "authors": [ "arlyxiao", "wenxi-zeng" ], "repo": "segmentio/Analytics-CSharp", "url": "https://github.com/segmentio/Analytics-CSharp/issues/118", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
541753549
Using credential_process when multiple IAM roles available I'm trying to get aws-okta configured as the credential_process for a particular profile. The Okta application configured at the SAML URL has multiple AWS IAM roles assigned, requiring the particular role to be chosen interactively. This interactive prompt causes aws-okta to hang indefinitely. I attempted to use the role_arn option to specify the desired role, but this seems to cause the CLI to bypass the credential_process option entirely, throwing an error about partial credentials. Is it possible to configure profiles to assume specific roles directly without an interactive prompt? been struggling with that as well, this is due to AWS behaviour, when role_arn is specified AWS sdk will either look for source_profile or direct credentials for this profile in credentials file, it does not support credential_process on main account i use credential_process and am not specifing role_arn and for roles i create profiles with source_profile pointing to main profile. This will AssumeRole in chain and work fine except the assumed roles will hit AWS hard limit for 1hr sts token when assumerolechain, which will be ok as okta will extend when needed. BUT if some shit dont understand credential_process in assumedchain roles like terrafrom s3 backend bucket then you need aws-okta env to inject session token to env and that one expires within 1hr no option for change and have to remove it manually I found a relevant issue in the botocore library: https://github.com/boto/botocore/issues/1329. It looks like there's still some work to be done in the AWS SDKs to make this work smoothly.
gharchive/issue
2019-12-23T13:23:48
2025-04-01T06:45:45.686588
{ "authors": [ "bgshacklett", "gacopl" ], "repo": "segmentio/aws-okta", "url": "https://github.com/segmentio/aws-okta/issues/266", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
129546671
adds validation for context.app.name This will return an error when context.app.name isn't passed through a server-side .screen() call. GA requires that an : Application Name be passed through .screen() calls. @sperand-io :eyes: @f2prateek updated tests
gharchive/pull-request
2016-01-28T19:06:53
2025-04-01T06:45:45.701685
{ "authors": [ "ladanazita" ], "repo": "segmentio/integration-google-analytics", "url": "https://github.com/segmentio/integration-google-analytics/pull/28", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
160463778
Allow for async processing This builds on #5 and adds options.async to requests, which internally uses null for the request id. This just means that the JSON-RPC server should not block the client while it waits for an answer, it will simply complete the request asynchronously. This could be useful where we don't necessarily care about waiting for the response, such as during sign up. Subsequent gets will just silently fail until the background work is done. This is just an idea, we'll have to see if it works in practice. /cc @stephenmathieson @stevenmiller888 Good idea! actually can we do this on the per-request basis? we generally use the same rpc client to make requests we do care about AND requests we don't @stephenmathieson: yeah, this is per-request (you add { async: true } after your params) @stephenmathieson: yeah, this is per-request (you add { async: true } after your params, this is not part of the constructor) haha gotcha whoops test added! lgtm
gharchive/pull-request
2016-06-15T16:17:27
2025-04-01T06:45:45.704719
{ "authors": [ "dominicbarnes", "stephenmathieson", "stevenmiller888" ], "repo": "segmentio/jsonrpc2.js", "url": "https://github.com/segmentio/jsonrpc2.js/pull/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
119697909
Add json options docs Highlight which options metalsmith.json recognises. Given these are detailed in the API section, I've just referred on to it. Fixes #106 :+1: I guess I have the ability to merge this myself now. So... poof! :smile:
gharchive/pull-request
2015-12-01T11:31:08
2025-04-01T06:45:45.706042
{ "authors": [ "Ajedi32", "iwootten" ], "repo": "segmentio/metalsmith", "url": "https://github.com/segmentio/metalsmith/pull/189", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
542121406
support pidfd related to #295 Job object maintain pidfds instead of pids for race-free process management use epoll for process waiting need linux kernel 5.4 or later no longer need it, since now call waitpid(-1)
gharchive/issue
2019-12-24T12:57:11
2025-04-01T06:45:45.724387
{ "authors": [ "sekiguchi-nagisa" ], "repo": "sekiguchi-nagisa/ydsh", "url": "https://github.com/sekiguchi-nagisa/ydsh/issues/392", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
80949710
how do i assign data attribute of select option on ajax request. Hello there, i have html select and i am binding data dynamically, now i have scenario i want to set three value of each select option so i want to assign data attribute of select option using ajaxrequest like: HELLO WORLD $(".select2").select2({ ajax: { url: '@Url.Action("GetProjectList","Test")', dataType: 'json', delay: 250, data: function (params) { return { search: params.term, // search term page: params.page }; }, processResults: function (data, page) { // response from server // how can i create the option data attribute return { results: data }; }, cache: true }, escapeMarkup: function (markup) { return markup; }, // let our custom formatter work minimumInputLength: 1, //templateResult: formatRepo, // omitted for brevity, see the source of this page //templateSelection: formatRepoSelection // omitted for brevity, see the source of this page }); This appears to be more of a usage question than a bug report or common feature request. As we continue to keep the Select2 issue tracker geared towards these two areas, bug reports and common feature requests, we are redirecting usage questions to other communities such as the mailing list, IRC, or Stack Overflow.
gharchive/issue
2015-05-26T12:53:42
2025-04-01T06:45:45.730607
{ "authors": [ "Mubusher87", "kevin-brown" ], "repo": "select2/select2", "url": "https://github.com/select2/select2/issues/3408", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
103528380
Select2 Onchange event triggered while applying data manually(how to suppress this) Previously I am using select2 3.3.2, in that while applying data through this syntax, jQuery('#elementId').select2('data', {"id":1,"text": "some value"}, true); It is not triggering that elements onchange event. I upgraded to 3.5.1, now its triggering the onchange event. If the change is intentional, is there any way to suppress that onchange event by passing a parameter to select2? Thanks in advance. Just for reference, 3.3.2 was released on May 9, 2013 and is pretty far out of the support cycle. 3.5.1 also isn't the latest release in the 3.5.x line, I'd recommend 3.5.2. The select2('data') method supports a triggerChange argument as the third parameter, and you're passing in true so it does trigger the change event. Try passing in false.
gharchive/issue
2015-08-27T14:53:42
2025-04-01T06:45:45.733594
{ "authors": [ "jaggu3017", "kevin-brown" ], "repo": "select2/select2", "url": "https://github.com/select2/select2/issues/3714", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
173218182
cannot display the placeholder with the numbers I use this sccript: <select class = "form-control select2" placeholder data = "1" style = "width: 100%;"> I have a problem displaying a placeholder number. give me a solution. thank you Do you mean when it initial by the following snippet , and it's not working ? $("#test").select2({ placeholder:134 }); if it's the situation you mention , you can initial it by add double quote like the following $("#test").select2({ placeholder:"134" }); its work. Thanks.
gharchive/issue
2016-08-25T14:26:32
2025-04-01T06:45:45.735807
{ "authors": [ "fannysispriadi", "holmes2136" ], "repo": "select2/select2", "url": "https://github.com/select2/select2/issues/4551", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
206498173
Tags case sensitivity Well the problem is again with tags and case sensitivity. Select2 allows inserting tags with different cases, where (at least in my opinion) it should not allow it. If a list has a "purple" tag the user should not be able to add "PURPLE", "PuRPLe" and so on. Granted I now have to manually choose to input the duplicate tag by pressing UP on the keyboard, which makes it a little bit harder but not impossible, which is what I'd like to achieve. This is the old bug report from 2015 july: https://github.com/select2/select2/issues/3566 In any case it's still possible to input duplicate tags: Prerequisites [ ] I have searched for similar issues in both open and closed tickets and cannot find a duplicate [x] The issue still exists against the latest master branch of Select2 [x] This is not a usage question (Those should be directed to the community) [x] I have attempted to find the simplest possible steps to reproduce the issue [ ] I have included a failing test as a pull request (Optional) Steps to reproduce the issue Go to https://select2.github.io/examples.html#tags Input PURPLE Press UP and ENTER Input PuRPLe Press UP and ENTER Etc. Expected behavior and actual behavior When I follow those steps, I see that the duplicate tags are being added to the list. I was expecting Select2 to limit me from inputing duplicate tags. Environment Browsers [x] Google Chrome [x] Mozilla Firefox [x] Internet Explorer Operating System [x] Windows [ ] Mac OS X [ ] Linux [ ] Mobile Libraries jQuery version: The one on the examples page Select2 version: The one on the examples page Isolating the problem [x] This bug happens on the examples page [x] The bug happens consistently across all tested browsers [x] This bug happens when using Select2 without other pluigns [ ] I can reproduce this bug in a jsbin It looks like this issue was actually fixed in https://github.com/select2/select2/commit/cb9a90457867ffb14c7b1550bb67e872e0a5c2dd, but it just hasn't made it into a release yet. I'm trying to get a new release out ASAP (and tighten the overall release cycle), so I would say that this issue is resolved.
gharchive/issue
2017-02-09T13:06:44
2025-04-01T06:45:45.745597
{ "authors": [ "alexweissman", "loshose" ], "repo": "select2/select2", "url": "https://github.com/select2/select2/issues/4784", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
900027640
Word wrap / Text Wrap within a Hidden Input Multiple Dropdown I am currently trying to word-wrap / text-wrap within the options (or the select) portion of the drop-down menu that has the hidden input with multiple selections with a size of 4, this way the text can be fully read and not be cut off. And instead of being cut off, the text goes down to the second or third line, aka a wrap Trying to work on the CSS and the Select2 jquery but unable to solve this issue. Below is my HTML. <html> <form> <select id="select2" class="select2 narrow wrap" multiple size="4"> <option value="01">01 Really long text description that normarly</option> <option value="02">02 Really long text description that normarly</option> <option value="03">03 Really long text description</option> <option value="04">04 Really long text description that normarly</option> <option value="05">05 Really long text description</option> <option value="06">06 Really long text description that normarly</option> </select> </form> </html> Here is my CSS: .select2.narrow { width: 200px; } .wrap.select2-selection--single { height: 100%; } .select2-container .wrap.select2-selection--multiple .select2-selection__rendered { word-wrap: break-word !important; text-overflow: inherit !important; white-space: normal !important; } .wrap { word-wrap: break-word; text-overflow: inherit; white-space: normal; } And here is my javascript: var $select2 = $('.select2').select2({ containerCssClass: "wrap" }) var $select2 = $('#select2')#select2({ minimumResultsForSearch: -4 }); Any help to see if this can be done is appreciated. This appears to be more of a usage question than a bug report or common feature request. As we continue to keep the Select2 issue tracker geared towards these two areas, bug reports and common feature requests, we are redirecting usage questions to other communities such as the mailing list, IRC, or Stack Overflow.
gharchive/issue
2021-05-24T21:26:38
2025-04-01T06:45:45.749624
{ "authors": [ "camanicnack", "kevin-brown" ], "repo": "select2/select2", "url": "https://github.com/select2/select2/issues/6038", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1164887002
On open second time modal - dropdown list duplicates Hi all, i run into problem, on second time when i open my modal popup, select2 dropdown gets duplicated.. 1st, 3rd, 4th ..etc.. works normally... Don't blame for Russian flag.. :D Tried using 4.0.7 / 4.0.12 / 4.0.13 / 4.1.0rc Using with transition, keepalive, router-view, tried without its, directly anyway got this problem... ;( Browser: Brave Version 1.36.105 Chromium: 99.0.4844.45 (Official Build) unknown (64-bit) / Opera OS: Linux Manjaro XFCE "jquery": "^3.6.0", "vue": "^3.0.0", temporary fix .select2:nth-child(even) { display: none; } This appears to be more of a usage question than a bug report or common feature request. As we continue to keep the Select2 issue tracker geared towards these two areas, bug reports and common feature requests, we are redirecting usage questions to other communities such as the mailing list, IRC, or Stack Overflow. If it's appears more of a usage question, than you know the answer. If you know the answer and you do not tell me, this appears to be be a you dont give f*** about that question... Sorry <3 Thanks! Neither of those inferences is logically valid. i know i just try to understand, why i cant find that problem "or question or whatever this kind of entity is" at www... This has been classified as a usage question instead of a bug because it appears to be caused by the specific usage that you describe. Please take this usage question to one of the recommended forums if you're looking to get your specific configuration working. Thanks! You didn't follow the required issue template for bugs and feature requests, so that's also a huge sign that this is actually a usage question.
gharchive/issue
2022-03-10T08:03:23
2025-04-01T06:45:45.756794
{ "authors": [ "iBrabus", "jonahgreenthal", "kevin-brown" ], "repo": "select2/select2", "url": "https://github.com/select2/select2/issues/6148", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
292498728
Actualize readme fix typo "alrady" => elaborate "around_filter/around_action" instead b/c it actually has been battle tested against 5.0 Fixed it in master, thanks for reporting!
gharchive/pull-request
2018-01-29T17:44:46
2025-04-01T06:45:45.807626
{ "authors": [ "AlexKVal", "semaperepelitsa" ], "repo": "semaperepelitsa/subdomain_locale", "url": "https://github.com/semaperepelitsa/subdomain_locale/pull/10", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2282552283
Playbook error I have a playbook with this list of tasks: ` - name: Fetching available collections command: ansible-galaxy collection list register: result - name: Available collections debug: var: result.stdout_lines - name: Fetching available roles command: ansible-galaxy list register: result - name: Available roles debug: var: result.stdout_lines ` and I get this error: TASK [Fetching available collections] ****************************************** fatal: [localhost]: FAILED! => { "changed":true, "cmd":[ "ansible-galaxy", "collection", "list" ], "delta":"0:00:02.423948", "end":"2024-04-30 13:26:21.383339", "msg":"non-zero return code", "rc":5, "start":"2024-04-30 13:26:18.959391", "stderr":"\u001b[0;31mERROR! - None of the provided paths were usable. Please specify a valid path with --collections-path\u001b[0m", "stderr_lines":[ "\u001b[0;31mERROR! - None of the provided paths were usable. Please specify a valid path with --collections-path\u001b[0m" ], "stdout":"usage: ansible-galaxy [-h] [--version] [-v] TYPE ...\n\nPerform various Role and Collection related operations.\n\npositional arguments:\n TYPE\n collection Manage an Ansible Galaxy collection.\n role Manage an Ansible Galaxy role.\n\noptions:\n --version show program's version number, config file location,\n configured module search path, module location, executable\n location and exit\n -h, --help show this help message and exit\n -v, --verbose Causes Ansible to print more debug messages. Adding multiple\n -v will increase the verbosity, the builtin plugins currently\n evaluate up to -vvvvvv. A reasonable level to start is -vvv,\n connection debugging might require -vvvv.", "stdout_lines":[ "usage: ansible-galaxy [-h] [--version] [-v] TYPE ...", "", "Perform various Role and Collection related operations.", "", "positional arguments:", " TYPE", " collection Manage an Ansible Galaxy collection.", " role Manage an Ansible Galaxy role.", "", "options:", " --version show program's version number, config file location,", " configured module search path, module location, executable", " location and exit", " -h, --help show this help message and exit", " -v, --verbose Causes Ansible to print more debug messages. Adding multiple", " -v will increase the verbosity, the builtin plugins currently", " evaluate up to -vvvvvv. A reasonable level to start is -vvv,", " connection debugging might require -vvvv." ] } I have a ansible.cfg that is specifying the COLLECTIONS_PATH and DEFAULT_ROLES_PATH to custom paths. This tasks are working fine on my local machine when I run the playbook but in in Ansible Semaphore UI. I guess this could be related to #1984 Not possible to set working directory for a template task In this version (v2.10.2-cc91501-1718257299) it seems to be working: Now the problem is that the following is still not wrking: name: Fetching available roles command: ansible-galaxy list register: result I have read som documentation to specify the paths for changing paths for collections and roles and Semaphore UI does not seem to respect them if I add them as CLI arguments for a task. [ "--connection=local", "-e", "ANSIBLE_CONFIG=folder/ansible/ansible.cfg" ] or [ "--connection=local", "-e", "ANSIBLE_COLLECTIONS_PATH=folder/ansible/.collections" ]
gharchive/issue
2024-05-07T07:53:34
2025-04-01T06:45:45.814596
{ "authors": [ "mikaelparkefelt" ], "repo": "semaphoreui/semaphore", "url": "https://github.com/semaphoreui/semaphore/issues/1998", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
363521472
SC-2853 LA config improvements By default, US region is set and it can be changed to EU on container start to update receivers endpoints. This PR solves the issue when custom LOG_URL is set and invalidates region settings. Also fixes logging message with receivers that are in use. Docs need to be changed before merging this PR. Docs need to be changed before merging this PR. Is there a PR for that @komljen ? Not yet, they will be added to the same PR.
gharchive/pull-request
2018-09-25T11:00:16
2025-04-01T06:45:45.816513
{ "authors": [ "komljen", "otisg" ], "repo": "sematext/logagent-js", "url": "https://github.com/sematext/logagent-js/pull/139", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
701728716
GraphQL Get group filter missing The GraphQL query with group filter for entity merging is missing. Queries like this are impossible with the client now: { Get { Things { Publication( group:{ type: merge, force:0.05 } ) { name } } } } Should be fixed/implemented in v1.0.0-rc4. Keeping this issues open until I have feedback that it works as desired. Confirmed working by @laura-ham. Closing.
gharchive/issue
2020-09-15T08:34:25
2025-04-01T06:45:45.818180
{ "authors": [ "etiennedi", "laura-ham" ], "repo": "semi-technologies/weaviate-javascript-client", "url": "https://github.com/semi-technologies/weaviate-javascript-client/issues/7", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2084756789
Test stream blocks I've added a single test for the stream_blocks function. Since this function is async, I've had to use tokio_test::block_on to test it, added it as a dev dependency. One thing I noticed is that stream_blocks just inputs everything into the writer but there isn't a way to read blocks from this writer later. extract_blocks could be adapted to do this for the stream of blocks, however. Should we be concerned about this or this unnecessary? I've added a single test for the stream_blocks function. Since this function is async, I've had to use tokio_test::block_on to test it, added it as a dev dependency. One thing I noticed is that stream_blocks just inputs everything into the writer but there isn't a way to read blocks from this writer later. extract_blocks could be adapted to do this for the stream of blocks, however. Should we be concerned about this or this unnecessary? Regarding this, its a nice to have but not urgent, since most components are tested individually
gharchive/pull-request
2024-01-16T19:58:39
2025-04-01T06:45:45.820484
{ "authors": [ "pedrohba1", "srosati" ], "repo": "semiotic-ai/flat-files-decoder", "url": "https://github.com/semiotic-ai/flat-files-decoder/pull/6", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
304192043
Failed to execute 'attachShader' on 'WebGLRenderingContext': parameter 1 is not of type 'WebGLProgram' After rendering the background for a while, the example throws this error 1/second and stops rendering: Uncaught TypeError: Failed to execute 'attachShader' on 'WebGLRenderingContext': parameter 1 is not of type 'WebGLProgram'. at WebGLSpriteRenderer.createProgram (VM253 three.js:4648) at WebGLSpriteRenderer.init (VM253 three.js:4451) at WebGLSpriteRenderer.render (VM253 three.js:4488) at WebGLRenderer.render (VM253 three.js:13947) at THREE.RenderPass.render (VM513 RenderPass.js:51) at THREE.EffectComposer.render (VM512 EffectComposer.js:88) at render (webgl_materials_cubemap_dynamic.html:981) at animate (webgl_materials_cubemap_dynamic.html:866) https://htmlpreview.github.io/?https://github.com/semleti/three.ts/blob/master/examples/webgl_materials_cubemap_dynamic.html same for sprites: https://htmlpreview.github.io/?https://github.com/semleti/three.ts/blob/master/examples/webgl_sprites.html same for misc ubiquity test2: https://htmlpreview.github.io/?https://github.com/semleti/three.ts/blob/master/examples/misc_ubiquity_test2.html Solved. Added a 'this' instead of using the local let in WebGLSpiteRenderer.createProgram .
gharchive/issue
2018-03-11T20:42:48
2025-04-01T06:45:45.823866
{ "authors": [ "semleti" ], "repo": "semleti/three.ts", "url": "https://github.com/semleti/three.ts/issues/19", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1955020470
Error handling Closes #5 Hey @semolex, any thoughts here? Sorry for delay. Let's merge it
gharchive/pull-request
2023-10-20T21:14:25
2025-04-01T06:45:45.824933
{ "authors": [ "krasnoukhov", "semolex" ], "repo": "semolex/novaposhta-python-client", "url": "https://github.com/semolex/novaposhta-python-client/pull/6", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2010883715
Model on online demo Hi. It's not clear to me what model is being used in the online demo, but it appears to be LLaMA 2, correct me if I am wrong. If this is so, may I suggest considering switching to something like mistral-openorca, zephyr, dolphin2.2.1-mistral, openhermes2.5-mistral, openchat or neural-chat? Unlike LLaMA (2), which has a proprietary license by Meta, these models are all under an open source license and are based on Mistral, which is under the Apache 2.0 license and some of them may even outperform LLaMA 2. Hi! The demo is using OpenChat 3.5 by Alignment Lab, which is Mistral based. Thank you, @kasumi-1, my bad. I will close the issue. Is this documented anywhere? The reason I thought it was LLaMA 2 is that it appears as the suggested default model if you open the Ollama settings: I don't think it is documented, just from launch announcement. I'll add to the faq on heyamica.com - thank you Thank you a lot! Please, consider updating Ollama default settings for consistency.
gharchive/issue
2023-11-26T04:10:05
2025-04-01T06:45:45.828628
{ "authors": [ "Aspie96", "kasumi-1" ], "repo": "semperai/amica", "url": "https://github.com/semperai/amica/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
374747926
Update readme.md This change is optional but most important to get updated npm Description of the change: Reason for the change: Link to original source: Closes # Hi, I don't have merge permission, it's just 1 line in this commit, so can you please try resolving that ? Hello @vk-sourcefuse,Thanks again for the PR! It's HACKTOBERFEST! We want to show our appreciation by sending you some special Hacktoberfest swag. If you have not already, could you please fill out this form so we can send it to you? Thanks! Team SendGrid DX @vk-sourcefuse - I accidentially overwrote your PR when I was fixing the conflict. Thanks for adding this - I made sure you got the hacktoberfest credit with us and with digital ocean!
gharchive/pull-request
2018-10-28T12:20:28
2025-04-01T06:45:45.833147
{ "authors": [ "ksigler7", "thinkingserious", "vk-sourcefuse" ], "repo": "sendgrid/docs", "url": "https://github.com/sendgrid/docs/pull/4553", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
455948736
Adding link to Pipedream pipeline Description of the change: This Pipedream pipeline ingests Sendgrid webhook data and lets you run SQL on event-level engagement data so you can ask interesting questions. It's all free, thought it would be useful to other devs. Reason for the change: Just thought it would be a cool, useful resource for other devs. Thank you for your submission, we really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.You have signed the CLA already but the status is still pending? Let us recheck it. Hello @dylburger,Thanks again for the PR! We want to show our appreciation by sending you some swag. Could you please fill out this form so we can send it to you? Thanks! Team SendGrid DX
gharchive/pull-request
2019-06-13T20:55:40
2025-04-01T06:45:45.837188
{ "authors": [ "SendGridDX", "dylburger", "thinkingserious" ], "repo": "sendgrid/docs", "url": "https://github.com/sendgrid/docs/pull/5303", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
279418139
< is not interpreted as less than sign Hi, I am trying to add text that contains a less than sign and has special styling, so I am using html converter to apply this and when I convert the less than character to < it doesn't get rendered in the generated file although the greater than sign is rendered the same way by replcing it to > in the HTML text. Is there a way to apply this Thanks @ahmedelmahalawey so after doing some digging the reason we lose HTML entities is that we pass the provided HTML string though Nokogiri::HTML.fragment here: https://github.com/senny/sablon/blob/2d2399e7343618f82a63f2335ddd144023e85dc9/lib/sablon/html/converter.rb#L18 and then through Nokogiri::XML.fragment here: https://github.com/senny/sablon/blob/2d2399e7343618f82a63f2335ddd144023e85dc9/lib/sablon/content.rb#L75 The HTML entities defined initially get converted by Nokogiri during the HTML parsing step and then get lost in the translation to XML. This stack overflow answer might be a possible solution but I'll need to investigate/test more. I ran into this issue as well but since I was rendering the HTML from Rails partials I figured it was a Rails produced problem and never looked into it. I think I have a fix in #80, would you mind testing it in your work flow @ahmedelmahalawey ? Yeah it worked perfectly, thanks.
gharchive/issue
2017-12-05T15:31:52
2025-04-01T06:45:45.860035
{ "authors": [ "ahmedelmahalawey", "stadelmanma" ], "repo": "senny/sablon", "url": "https://github.com/senny/sablon/issues/79", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1253175013
🛑 Blockly is down In 4526eb2, Blockly (https://blockly.sensebox.de/ardublockly/?board=sensebox-mcu) was down: HTTP code: 0 Response time: 0 ms Resolved: Blockly is back up in 7fe21af.
gharchive/issue
2022-05-30T23:05:21
2025-04-01T06:45:45.864644
{ "authors": [ "mpfeil" ], "repo": "sensebox/upptime", "url": "https://github.com/sensebox/upptime/issues/95", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }