llama.cpp/.github
Reese Levine 21c021745d
ggml: Add initial WebGPU backend (#14521)
* Minimal setup of webgpu backend with dawn. Just prints out the adapter and segfaults

* Initialize webgpu device

* Making progress on setting up the backend

* Finish more boilerplate/utility functions

* Organize file and work on alloc buffer

* Add webgpu_context to prepare for actually running some shaders

* Work on memset and add shader loading

* Work on memset polyfill

* Implement set_tensor as webgpu WriteBuffer, remove host_buffer stubs since webgpu doesn't support it

* Implement get_tensor and buffer_clear

* Finish rest of setup

* Start work on compute graph

* Basic mat mul working

* Work on emscripten build

* Basic WebGPU backend instructions

* Use EMSCRIPTEN flag

* Work on passing ci, implement 4d tensor multiplication

* Pass thread safety test

* Implement permuting for mul_mat and cpy

* minor cleanups

* Address feedback

* Remove division by type size in cpy op

* Fix formatting and add github action workflows for vulkan and metal (m-series) webgpu backends

* Fix name

* Fix macos dawn prefix path
2025-07-16 18:18:51 +03:00
..
ISSUE_TEMPLATE ggml : remove kompute backend (#14501) 2025-07-03 07:48:32 +03:00
actions releases : use arm version of curl for arm releases (#13592) 2025-05-16 19:36:51 +02:00
workflows ggml: Add initial WebGPU backend (#14521) 2025-07-16 18:18:51 +03:00
labeler.yml ggml : remove kompute backend (#14501) 2025-07-03 07:48:32 +03:00
pull_request_template.md repo : update links to new url (#11886) 2025-02-15 16:40:57 +02:00