Upload README.md with huggingface_hub

#2
by sayakpaul HF Staff - opened
Files changed (1) hide show
  1. README.md +50 -5
README.md CHANGED
@@ -1,8 +1,53 @@
1
  ---
2
- tags:
3
- - kernels
4
- - cuda
5
  ---
6
- # SageAttention
7
 
8
- This is a build of the [SageAttention](https://github.com/thu-ml/SageAttention) compatible with kernels library
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ library_name: kernels
3
+ license: apache-2.0
 
4
  ---
 
5
 
6
+ <!-- This model card has automatically been generated. You
7
+ should probably proofread and complete it, then remove this comment. -->
8
+
9
+
10
+ This is the repository card of {repo_id} that has been pushed on the Hub. It was built to be used with the [`kernels` library](https://github.com/huggingface/kernels). This card was automatically generated.
11
+
12
+
13
+ ## How to use
14
+
15
+ ```python
16
+ # make sure `kernels` is installed: `pip install -U kernels`
17
+ from kernels import get_kernel
18
+
19
+ kernel_module = get_kernel("kernels-community/sage_attention") # <- change the ID if needed
20
+ per_block_int8 = kernel_module.per_block_int8
21
+
22
+ per_block_int8(...)
23
+ ```
24
+
25
+ ## Available functions
26
+
27
+ - `per_block_int8`
28
+ - `per_warp_int8`
29
+ - `sub_mean`
30
+ - `per_channel_fp8`
31
+ - `sageattn`
32
+
33
+ ## Supported backends
34
+
35
+ - cuda
36
+
37
+ ## CUDA Capabilities
38
+
39
+ - 9.0a
40
+ - 8.9
41
+ - 8.0
42
+
43
+ ## Benchmarks
44
+
45
+ [TODO: provide benchmarks if available]
46
+
47
+ ## Source code
48
+
49
+ [TODO: provide original source code and other relevant citations if available]
50
+
51
+ ## Notes
52
+
53
+ [TODO: provide additional notes about this kernel if needed]