File tree 9 files changed +13
-10
lines changed
9 files changed +13
-10
lines changed Original file line number Diff line number Diff line change @@ -8,7 +8,7 @@ brings language model chats directly onto web browsers with hardware acceleratio
8
8
** Everything runs inside the browser with no server support and accelerated with WebGPU.**
9
9
We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration.
10
10
11
- ** [ Check out our demo webpage to try out!] ( https://mlc.ai/web-llm / ) **
11
+ ** [ Check out our demo webpage to try out!] ( https://webllm. mlc.ai/ ) **
12
12
This project is a companion project of [ MLC LLM] ( https://github.com/mlc-ai/mlc-llm ) ,
13
13
our companion project that runs LLMs natively on iPhone and other native local environments.
14
14
@@ -213,7 +213,7 @@ WebLLM package is a web runtime designed for [MLC LLM](https://github.com/mlc-ai
213
213
214
214
## Links
215
215
216
- - [Demo page](https://mlc.ai/web-llm /)
216
+ - [Demo page](https://webllm. mlc.ai/)
217
217
- If you want to run LLM on native runtime, check out [MLC-LLM](https://github.com/mlc-ai/mlc-llm)
218
218
- You might also be interested in [Web Stable Diffusion](https://github.com/mlc-ai/web-stable-diffusion/).
219
219
Original file line number Diff line number Diff line change 5
5
"scripts" : {
6
6
"start" : " cp src/gh-config.js src/app-config.js && parcel src/llm_chat.html --port 8888" ,
7
7
"mlc-local" : " cp src/mlc-local-config.js src/app-config.js && parcel src/llm_chat.html --port 8888" ,
8
- "build" : " cp src/gh-config.js src/app-config.js && parcel build src/llm_chat.html --dist-dir lib --no-content-hash --public-url /web-llm "
8
+ "build" : " cp src/gh-config.js src/app-config.js && parcel build src/llm_chat.html --dist-dir lib --no-content-hash"
9
9
},
10
10
"devDependencies" : {
11
11
"buffer" : " ^5.7.1" ,
Original file line number Diff line number Diff line change @@ -11,7 +11,7 @@ export default {
11
11
"local_id" : "vicuna-v1-7b-q4f32_0"
12
12
} ,
13
13
// fp16 options are enabled through chrome canary flags
14
- // chrome --enale -dawn-features=enable_unsafe_apis
14
+ // chrome --enable -dawn-features=enable_unsafe_apis
15
15
{
16
16
"model_url" : "http://localhost:8000/RedPajama-INCITE-Chat-3B-v1-q4f16_0/params/" ,
17
17
"local_id" : "RedPajama-INCITE-Chat-3B-v1-q4f16_0" ,
Original file line number Diff line number Diff line change 17
17
"url" : " git+https://github.com/mlc-ai/web-llm"
18
18
},
19
19
"keywords" : [
20
- " llm" , " large language model" , " machine learning"
20
+ " llm" ,
21
+ " large language model" ,
22
+ " machine learning"
21
23
],
22
24
"license" : " Apache-2.0" ,
23
25
"homepage" : " https://github.com/mlc-ai/web-llm" ,
Original file line number Diff line number Diff line change @@ -15,6 +15,7 @@ rm -rf docs .gitignore
15
15
mkdir -p docs
16
16
cp -rf site/_site/* docs
17
17
touch docs/.nojekyll
18
+ echo " webllm.mlc.ai" >> docs/CNAME
18
19
19
20
DATE=` date`
20
21
git add docs && git commit -am " Build at ${DATE} "
Original file line number Diff line number Diff line change 8
8
9
9
cp examples/simple-chat/lib/* site
10
10
11
- cd site && jekyll serve --host localhost --baseurl /web-llm -- port 8888
11
+ cd site && jekyll serve --host localhost --port 8888
Original file line number Diff line number Diff line change 2
2
name : " Web LLM"
3
3
short_name : " WebLLM"
4
4
5
- url : https://mlc.ai/web-llm
5
+ url : https://webllm. mlc.ai
6
6
7
7
exclude : [README.md, serve_local.sh]
8
8
@@ -24,7 +24,7 @@ permalink: /blog/:year/:month/:day/:title.html
24
24
front_page_news : 8
25
25
26
26
# Base pathname for links.
27
- base : ' /web-llm '
27
+ base : ' '
28
28
29
29
# make pages for the _projects folder
30
30
collections :
Original file line number Diff line number Diff line change @@ -53,7 +53,7 @@ The chat demo is based on [vicuna-7b-v1.1](https://huggingface.co/lmsys/vicuna-7
53
53
## Links
54
54
55
55
- [ Web LLM Github] ( https://github.com/mlc-ai/web-llm )
56
- - You might also be interested in [ Web Stable Diffusion] ( https://mlc.ai/web-stable-diffusion / ) .
56
+ - You might also be interested in [ Web Stable Diffusion] ( https://websd. mlc.ai/ ) .
57
57
58
58
## Disclaimer
59
59
Original file line number Diff line number Diff line change @@ -109,7 +109,7 @@ export class ChatModule implements ChatInterface {
109
109
throw Error (
110
110
"This model requires WebGPU extension shader-f16, " +
111
111
"which is not enabled in this browser. " +
112
- "You can try Chrome Canary with flag --enable-dawn-features=allow_unsafe_api "
112
+ "You can try Chrome Canary with flag --enable-dawn-features=allow_unsafe_apis "
113
113
) ;
114
114
}
115
115
throw Error (
You can’t perform that action at this time.
0 commit comments