Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
#### Overview

JupyterGIS provides a powerful suite of vector geospatial processing toolssuch as buffer creation, centroid calculation, and convex hull generation—directly in the browser. These capabilities are enabled by a GDAL WebAssembly (WASM) build, eliminating the need for server-side processing or local installations.
JupyterGIS provides a powerful suite of vector geospatial processing tools - such as buffer creation, centroid calculation, and convex hull generation—directly in the browser. These capabilities are enabled by a GDAL WebAssembly (WASM) build, eliminating the need for server-side processing or local installations.

Building on this foundation, we are now expanding support to raster processing tools, unlocking even more geospatial workflows in JupyterLite and notebook.link. This will enable users to perform complex raster operations entirely client-side, making geospatial analysis more accessible, collaborative, and scalable.

Expand Down
6 changes: 3 additions & 3 deletions src/components/fundable/descriptions/SVE2SupportInXsimd.md
Original file line number Diff line number Diff line change
@@ -1,11 +1,11 @@
#### Overview
xsimd is a C++ header-only library that abstracts SIMD (vectorization) intrinsics behind a single, generic API. The same code - `xsimd::batch<float>` - compiles to optimal machine code on x86 SSE/AVX, ARM NEON, RISC-V, and WebAssembly, with no runtime overhead. When an intrinsic is missing on a given target, xsimd falls back gracefully rather than failing or leaving the developer to write platform-specific branches. This is why projects like Mozilla Firefox, Apache Arrow, Meta Velox, KDE Krita, and Pythran have adopted it as their vectorization layer.

Writing SIMD intrinsics by hand is notoriously painful. Each architecture exposes different types, different naming conventions, and different gotchas alignment rules, predicate registers, lane-width surprises, compiler bugs. Targeting two architectures means two separate, opaque, hard-to-test code paths. xsimd eliminates that cost entirely.
Writing SIMD intrinsics by hand is notoriously painful. Each architecture exposes different types, different naming conventions, and different gotchas - alignment rules, predicate registers, lane-width surprises, compiler bugs. Targeting two architectures means two separate, opaque, hard-to-test code paths. xsimd eliminates that cost entirely.

ARM is now a first-class server architecture. AWS Graviton4 broadly deployed today across EC2 instance families ships with SVE2, ARM's latest vector ISA. Google and Microsoft have both announced their own ARM-based server chips (Axion and Cobalt 200 respectively), signalling an industry-wide shift. SVE2 brings a materially richer instruction set than NEON, and xsimd currently cannot exploit it falling back to SVE and NEON on hardware that is capable of significantly more.
ARM is now a first-class server architecture. AWS Graviton4 - broadly deployed today across EC2 instance families - ships with SVE2, ARM's latest vector ISA. Google and Microsoft have both announced their own ARM-based server chips (Axion and Cobalt 200 respectively), signalling an industry-wide shift. SVE2 brings a materially richer instruction set than NEON, and xsimd currently cannot exploit it - falling back to SVE and NEON on hardware that is capable of significantly more.

We propose to implement a complete xsimd::sve2 backend all arithmetic, mathematical, and reduction operations . The work will be fully tested, documented, and merged upstream into the public repository under the existing BSD 3-Clause license.
We propose to implement a complete xsimd::sve2 backend - all arithmetic, mathematical, and reduction operations . The work will be fully tested, documented, and merged upstream into the public repository under the existing BSD 3-Clause license.

Funding this development will directly enhance the cutting-edge data processing workflow using Arrow and Velox, paving the way for the future of ARM computing!

Expand Down
2 changes: 1 addition & 1 deletion src/components/fundable/projectsDetails.ts
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ export const fundableProjectsDetails = {
category: "Scientific Computing",
title: "Implementing Kazushige Goto Algorithms for Matrix Operations in xtensor",
pageName: "MatrixOperationsInXtensor",
shortDescription: "This project aims to integrate Kazushige Goto’s highly optimized matrix multiplication algorithms into the [xtensor](https://github.com/xtensor-stack/xtensor/) framework, leveraging the [xsimd](https://github.com/xtensor-stack/xsimd/) library for SIMD acceleration.",
shortDescription: "This project aims to integrate Kazushige Goto’s highly optimized matrix multiplication algorithms into the xtensor framework, leveraging the xsimd library for SIMD acceleration.",
description: MatrixOperationsInXtensor,
price: "45 000 €",
maxNbOfFunders: 1,
Expand Down
Loading