jdp8 commited on
Commit
baabeb7
·
verified ·
1 Parent(s): 9847c7c

Initial WASM files commit

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. web-llm-models/v0_2_30/Llama-2-13b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  2. web-llm-models/v0_2_30/Llama-2-70b-chat-hf-q3f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  3. web-llm-models/v0_2_30/Llama-2-70b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  4. web-llm-models/v0_2_30/Llama-2-7b-chat-hf-q4f16_1-ctx1k-webgpu.wasm +3 -0
  5. web-llm-models/v0_2_30/Llama-2-7b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  6. web-llm-models/v0_2_30/Llama-2-7b-chat-hf-q4f32_1-ctx4k_cs1k-webgpu.wasm +3 -0
  7. web-llm-models/v0_2_30/Llama-3-70B-Instruct-q3f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  8. web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f16_1-ctx1k_cs1k-webgpu.wasm +3 -0
  9. web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  10. web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f32_1-ctx1k_cs1k-webgpu.wasm +3 -0
  11. web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm +3 -0
  12. web-llm-models/v0_2_30/Mistral-7B-Instruct-v0.2-q4f16_1-sw4k_cs1k-webgpu.wasm +3 -0
  13. web-llm-models/v0_2_30/Mistral-7B-Instruct-v0.2-q4f32_1-sw4k_cs1k-webgpu.wasm +3 -0
  14. web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f16_1-ctx1k-webgpu.wasm +3 -0
  15. web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f16_1-ctx2k-webgpu.wasm +3 -0
  16. web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f32_1-ctx1k-webgpu.wasm +3 -0
  17. web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f32_1-ctx2k-webgpu.wasm +3 -0
  18. web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q0f16-ctx2k-webgpu.wasm +3 -0
  19. web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q0f32-ctx2k-webgpu.wasm +3 -0
  20. web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q4f16_1-ctx1k-webgpu.wasm +3 -0
  21. web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q4f32_1-ctx1k-webgpu.wasm +3 -0
  22. web-llm-models/v0_2_30/gemma-2b-it-q4f16_1-ctx1k_cs1k-webgpu.wasm +3 -0
  23. web-llm-models/v0_2_30/gemma-2b-it-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  24. web-llm-models/v0_2_30/gemma-2b-it-q4f32_1-ctx1k_cs1k-webgpu.wasm +3 -0
  25. web-llm-models/v0_2_30/gemma-2b-it-q4f32_1-ctx4k_cs1k-webgpu.wasm +3 -0
  26. web-llm-models/v0_2_30/gpt2-medium-q0f16-ctx1k-webgpu.wasm +3 -0
  27. web-llm-models/v0_2_30/gpt2-q0f16-ctx1k-webgpu.wasm +3 -0
  28. web-llm-models/v0_2_30/phi-1_5-q0f16-ctx2k-webgpu.wasm +3 -0
  29. web-llm-models/v0_2_30/phi-1_5-q0f32-ctx2k-webgpu.wasm +3 -0
  30. web-llm-models/v0_2_30/phi-1_5-q4f16_1-ctx1k-webgpu.wasm +3 -0
  31. web-llm-models/v0_2_30/phi-1_5-q4f16_1-ctx2k-webgpu.wasm +3 -0
  32. web-llm-models/v0_2_30/phi-1_5-q4f32_1-ctx1k-webgpu.wasm +3 -0
  33. web-llm-models/v0_2_30/phi-1_5-q4f32_1-ctx2k-webgpu.wasm +3 -0
  34. web-llm-models/v0_2_30/phi-2-q0f16-ctx2k-webgpu.wasm +3 -0
  35. web-llm-models/v0_2_30/phi-2-q0f32-ctx2k-webgpu.wasm +3 -0
  36. web-llm-models/v0_2_30/phi-2-q4f16_1-ctx1k-webgpu.wasm +3 -0
  37. web-llm-models/v0_2_30/phi-2-q4f16_1-ctx2k-webgpu.wasm +3 -0
  38. web-llm-models/v0_2_30/phi-2-q4f32_1-ctx1k-webgpu.wasm +3 -0
  39. web-llm-models/v0_2_30/phi-2-q4f32_1-ctx2k-webgpu.wasm +3 -0
  40. web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f16_1-ctx1k_cs1k-webgpu.wasm +3 -0
  41. web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  42. web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f32_1-ctx1k_cs1k-webgpu.wasm +3 -0
  43. web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f32_1-ctx4k_cs1k-webgpu.wasm +3 -0
  44. web-llm-models/v0_2_34/Hermes-2-Pro-Mistral-7B-q4f16_1-sw4k_cs1k-webgpu.wasm +3 -0
  45. web-llm-models/v0_2_34/Llama-2-13b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  46. web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f16_1-ctx1k-webgpu.wasm +3 -0
  47. web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
  48. web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f32_1-ctx1k-webgpu.wasm +3 -0
  49. web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f32_1-ctx4k_cs1k-webgpu.wasm +3 -0
  50. web-llm-models/v0_2_34/Llama-3-70B-Instruct-q3f16_1-ctx4k_cs1k-webgpu.wasm +3 -0
web-llm-models/v0_2_30/Llama-2-13b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a78f2c00a22d6e1618222402be2bf0dd6525722b31f394fdb1a46c9816160a44
3
+ size 4000074
web-llm-models/v0_2_30/Llama-2-70b-chat-hf-q3f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31803d87eca0913d7043e832c8f93c54f94ea8a7e285350f2d9bbf9a51a03066
3
+ size 5088677
web-llm-models/v0_2_30/Llama-2-70b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa6bd27a64a17c8e50ba491b67386281712b1965b7ddb40dcf64273c58888c46
3
+ size 5050091
web-llm-models/v0_2_30/Llama-2-7b-chat-hf-q4f16_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7d973a204e9da3f4dfcd224825440f2f7e2d79c8f145091a147c3fae6bed024
3
+ size 3755714
web-llm-models/v0_2_30/Llama-2-7b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b5c586d0d2cc8aafb540d7dab758a1880d6f411f41eb47ba8f0cda7d1d7f235
3
+ size 3756119
web-llm-models/v0_2_30/Llama-2-7b-chat-hf-q4f32_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85975ec6bde81dff5a7506dc2d11db21fbd8e401c90abf06a7cc500277711a72
3
+ size 3646307
web-llm-models/v0_2_30/Llama-3-70B-Instruct-q3f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a88a180704505a432209c3f76ea1b2f1c74029e8744ebb00e323af4693cb240
3
+ size 7167312
web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f16_1-ctx1k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6823b0f6aa474236435ca0dd0f7c67f2cca0e6a9e5467f850ab650f4d076049
3
+ size 4718018
web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9445821d84e2094957e160239e1fe89e019c3bceec88e41213e35f3df020e7c
3
+ size 4723547
web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f32_1-ctx1k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:223957f419543c75b3a1b865e3173eb2d32c774a9772b064199f8b9aa0ecc524
3
+ size 4520045
web-llm-models/v0_2_30/Llama-3-8B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2d377dabfb84751b61e96cc44ca1c29c51f370acd5c6e04d431d29376ceb271
3
+ size 4525793
web-llm-models/v0_2_30/Mistral-7B-Instruct-v0.2-q4f16_1-sw4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:38c244730831d5bd1f05fb51af382239728c5c08e67740893845e4fa3621955d
3
+ size 3783202
web-llm-models/v0_2_30/Mistral-7B-Instruct-v0.2-q4f32_1-sw4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d8f7f06db2bd041505908205de6b364e02e30c789da6ccfab6296c8db71e231
3
+ size 3676961
web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f16_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a74f278bd2e22a9823b7ff7206ec7fc1f6d9df6ec7aad5a4d4116ed7ed45df29
3
+ size 4045134
web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f16_1-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b58c74e7f60701a267666cc0fb1a285bbfe2dc1ebc180ac40c96f1ac82bcaae4
3
+ size 4043659
web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f32_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8c120c6305e1efe9ba2376e413ba848f802501797792f70e0fb69e6051cd37d
3
+ size 4025155
web-llm-models/v0_2_30/RedPajama-INCITE-Chat-3B-v1-q4f32_1-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0877749fefacebf0999d895bbd05c5f25f7b891921a44c74b11fc22da27814b2
3
+ size 4023882
web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q0f16-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b6698ac220384f2b6112dc9e76d9f1c55b202c659c6abe78e0bce86d9c9f08a
3
+ size 3268536
web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q0f32-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58cc54315bf8f98e6851f9c810377d86e43a91c4eb75b8f4b5fed75279ce1813
3
+ size 3248462
web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q4f16_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47e6f746cbd34486e7d20d540643bd1bbd41d48b4de19e589c747caa57fca9ff
3
+ size 3510691
web-llm-models/v0_2_30/TinyLlama-1.1B-Chat-v0.4-q4f32_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3382168f379740bf73ebbf5b55a19554bb356f06c696e00a91386bbdcee525fb
3
+ size 3439397
web-llm-models/v0_2_30/gemma-2b-it-q4f16_1-ctx1k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ac65d002c3add9f2c727fc1050c0b0f5dbeebea9434a9c5d9454217e8102fd1
3
+ size 3426696
web-llm-models/v0_2_30/gemma-2b-it-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc98b1682d804c4ca1f1ac9c0a16bf313a5f006184c37f887e5fab8491947125
3
+ size 3418409
web-llm-models/v0_2_30/gemma-2b-it-q4f32_1-ctx1k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95f2bea36f5b9d2c2ad4837313e5ff90132c5580e6396e119d1ab5b351a92ca9
3
+ size 3357441
web-llm-models/v0_2_30/gemma-2b-it-q4f32_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:177c736725df0f91eb6421491baed13817bc45ffa87b3ee3d37b90f832031d2b
3
+ size 3359729
web-llm-models/v0_2_30/gpt2-medium-q0f16-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3056254097dc3f5483fdef51516d5b1d847eb520aab3e7e3446ea5a30a81b3c
3
+ size 3685171
web-llm-models/v0_2_30/gpt2-q0f16-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:584cd0b1a56d7b46bd3283a5965a9c2478b161bba2ca0d213d51773289b4f92e
3
+ size 3321904
web-llm-models/v0_2_30/phi-1_5-q0f16-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f52492b33a05317fcbda6110c7bc901458f35ef3be985370a428e08d344d1235
3
+ size 3615730
web-llm-models/v0_2_30/phi-1_5-q0f32-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8dd57a9c011e1a7d45635d3a13eabb7e51ad32492035304f90669347bb9478c
3
+ size 3597640
web-llm-models/v0_2_30/phi-1_5-q4f16_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4d5fda186db4187605bd0786f0f1279d61e98e3c2d73d64070eef8c16d227df
3
+ size 3722907
web-llm-models/v0_2_30/phi-1_5-q4f16_1-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc09ffd12c91c8f4ed033b47d71860f0ecbf463d3e04e04eb7ab25be80cb34ff
3
+ size 3716842
web-llm-models/v0_2_30/phi-1_5-q4f32_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff07062905553d550449280169414db220f8177ad7e0e4a47bd789d225c9af42
3
+ size 3705968
web-llm-models/v0_2_30/phi-1_5-q4f32_1-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:044a910df9ebdaaa6c4bf3faea0c32b15fd7c47ec47c5875922d65be6802498b
3
+ size 3706135
web-llm-models/v0_2_30/phi-2-q0f16-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b1e78f7430328bf8817ea385ae2e56f74fff0844a59bb9d2fdde97c08315d37
3
+ size 3706591
web-llm-models/v0_2_30/phi-2-q0f32-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d45483c6812679950b2c2e802f333c20a12a741399149c91e770f8b875c0eca2
3
+ size 3704831
web-llm-models/v0_2_30/phi-2-q4f16_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ad6097925c35134d42aa74dcc1fa98f0228e1eeeb97ff84245bb07bde88f94d
3
+ size 3874153
web-llm-models/v0_2_30/phi-2-q4f16_1-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cb52cb8607bc4ba5bea2e41e3f0017fa8a2356e819cf80e8b21ffea7664732c
3
+ size 3873139
web-llm-models/v0_2_30/phi-2-q4f32_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33a547dddb660429b029e5c1d56f8ca8fdb51f1bb5c0b1b5800b29f24d5ab2a4
3
+ size 3865938
web-llm-models/v0_2_30/phi-2-q4f32_1-ctx2k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41e1175d97e3bfd031c4586f2f407ddb17de2a7ff65c4fd7d00771d899a46420
3
+ size 3868789
web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f16_1-ctx1k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4b6aaee15ec3993a77c62b381b9d1d81ffedca2a901098ec0121fbfa9de3146
3
+ size 3692600
web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9682df1920b1523a5ea52603b7d5609b70415a3a61cf402d74b5473d2d6e6bef
3
+ size 3700368
web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f32_1-ctx1k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:002494fb0d3612965b878a55b7b97fbd01f7bb71313711b6f412c462473edafe
3
+ size 3673424
web-llm-models/v0_2_30/stablelm-2-zephyr-1_6b-q4f32_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c8f7b4c3eb20540d2ab40d0a7fa9e725fd4730d2dad9cfba2f1e966b641cb7f
3
+ size 3679797
web-llm-models/v0_2_34/Hermes-2-Pro-Mistral-7B-q4f16_1-sw4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc85bfb1bb25e18fe436bb4dd1239acc1817a7019c9211cb7e0bf8e735869470
3
+ size 3888798
web-llm-models/v0_2_34/Llama-2-13b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42f11f0db35c2301adfb4ce77e4f08e430fce0491799467ecebe431d940982b2
3
+ size 5217380
web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f16_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e09883113764ba1fdaaa6b30dd8d1f56585b1d8273f072a6443157205c559f20
3
+ size 4776918
web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad27f14797266027df84580647e1b4c43544423e605afeb9da444ec44e29d491
3
+ size 4776915
web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f32_1-ctx1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2376cb97eab019bd58cf8f6b216aefca98871d324bfd04414c6bb3e22cab7216
3
+ size 4577106
web-llm-models/v0_2_34/Llama-2-7b-chat-hf-q4f32_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69807e2898b26324e0c0a1478d98430c78aa1f1273c794d845630d6fcab0bcb9
3
+ size 4577123
web-llm-models/v0_2_34/Llama-3-70B-Instruct-q3f16_1-ctx4k_cs1k-webgpu.wasm ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb18875bb93ac8300fe4b720fbdb7f3ff96b0ade0f0713817d9b363d90a71881
3
+ size 7266794