Lists (18)
Sort Name ascending (A-Z)
Stars
- All languages
- Assembly
- AutoHotkey
- Awk
- Batchfile
- C
- C#
- C++
- CSS
- CoffeeScript
- Crystal
- Cuda
- Dart
- Dockerfile
- Elixir
- Emacs Lisp
- Erlang
- F#
- Fluent
- Go
- HTML
- Haskell
- Inno Setup
- Java
- JavaScript
- Julia
- Jupyter Notebook
- Kotlin
- Lua
- MDX
- Makefile
- Markdown
- Max
- Nim
- Nix
- OCaml
- Objective-C
- Objective-C++
- PHP
- POV-Ray SDL
- PowerShell
- Python
- Rich Text Format
- Roff
- Ruby
- Rust
- SCSS
- Scheme
- Shell
- Swift
- TeX
- Tree-sitter Query
- TypeScript
- Typst
- Vim Script
- Vue
- YARA
- Zig
在大陆地区使用普通银联信用卡通过 Apple Store 订阅 ChatGPT Plus 等产品的一些实践
Rust macOS automation toolkit — screenshot, OCR, accessibility CLI, break reminder, and more, built on Apple native frameworks via objc2
Open-source CUDA compiler targeting multiple GPU architectures. Compiles .cu to AMD and Tenstorrent GPU's
Local voice input with screen-aware context. Push-to-talk → Whisper → LLM refinement, all on your own GPU.
Pure C inference of Mistral Voxtral Realtime 4B speech to text model
A set of ready to use Agent Skills for research, science, engineering, analysis, finance and writing.
Slide decks, coding exercises, and quick references for learning the JAX AI Stack
A collection of research papers on low-precision training methods
Simple PyTorch Tutorials Zero to ALL!
Row-wise block scaling for fp8 quantization matrix multiplication. Solution to GPU mode AMD challenge.
We want to compare how good Qwen3-1.7B-Base using B200 to continue pretraining on Malaysian multi-lingual corpus on different mixed precision training with proper truncated multi-packing.
how to optimize some algorithm in cuda.
implement GPT-OSS 20B & 120B C++ inference from scratch on AMD GPUs
trholding / llama2.c
Forked from karpathy/llama2.cLlama 2 Everywhere (L2E)
Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step




