Skip to content

Conversation

@CharlieFRuan
Copy link
Member

@CharlieFRuan CharlieFRuan commented May 29, 2024

Compiled at:

Main changes include:

@CharlieFRuan CharlieFRuan marked this pull request as ready for review May 29, 2024 21:32
@CharlieFRuan CharlieFRuan merged commit d87aa9e into mlc-ai:main May 29, 2024
CharlieFRuan added a commit to mlc-ai/web-llm that referenced this pull request May 29, 2024
…433) This PR updates models to v0.2.39 compiled with mlc-ai/binary-mlc-llm-libs#123 The main change is the new MLC-LLM runtime, which supports grammar (i.e. json mode) for Llama3. - Hence we now read in field `tokenizer_info` (or deprecated `token_table_postproc_method`) from `mlc-chat-config.json` when post processing token table for Grammar - If neither is available, we use the default `byte_fallback` New prebuilt models introduced: - Phi3-mini-4k - Hermes-2-Pro-Llama-3-8B - Qwen1.5-1.8B - StableLM-2-zephyr_1.6B Updates on examples: - json-mode and json-schema now use Llama3 to demonstrate - Function calling inside json-schema now uses `Hermes-2-Pro-Llama-3-8B` instead of `Hermes-2-Pro-Mistral`
CharlieFRuan added a commit to mlc-ai/web-llm that referenced this pull request May 30, 2024
### Changes Main changes include: - New prebuilt models: - Phi3-mini - StableLM-2-zephyr-1.6B - Qwen1.5-1.8B - Hermes2-Pro-Llama-3-8B to prebuilt models - Updates on `ModelRecord` fields - For detail see: #435 - Update all WASMs - For detail see: #433 - Update all WASMs to v0.2.39 - Support grammar for Llama3, hence update examples/json-mode to use `Llama3` and `Hermes2-pro-Llama3-8B` for function calling in `examples/json-schema` - Use `loglevel` package: - For details see #427 - Fix `index.js.map` issue for Vite - #420 - Enhance error handling and ServiceWorker ### TVMjs TVMjs compiled at apache/tvm@71f7af7 - Main changes include: - apache/tvm#17031 - apache/tvm#17028 - apache/tvm#17021 ### WASM version - All wasms updated to 0.2.39 via mlc-ai/binary-mlc-llm-libs#123 for new MLC-LLM runtime (mainly grammar)
jingyi-zhao-01 pushed a commit to jingyi-zhao-01/web-llm that referenced this pull request Dec 8, 2024
…lc-ai#433) This PR updates models to v0.2.39 compiled with mlc-ai/binary-mlc-llm-libs#123 The main change is the new MLC-LLM runtime, which supports grammar (i.e. json mode) for Llama3. - Hence we now read in field `tokenizer_info` (or deprecated `token_table_postproc_method`) from `mlc-chat-config.json` when post processing token table for Grammar - If neither is available, we use the default `byte_fallback` New prebuilt models introduced: - Phi3-mini-4k - Hermes-2-Pro-Llama-3-8B - Qwen1.5-1.8B - StableLM-2-zephyr_1.6B Updates on examples: - json-mode and json-schema now use Llama3 to demonstrate - Function calling inside json-schema now uses `Hermes-2-Pro-Llama-3-8B` instead of `Hermes-2-Pro-Mistral`
jingyi-zhao-01 pushed a commit to jingyi-zhao-01/web-llm that referenced this pull request Dec 8, 2024
### Changes Main changes include: - New prebuilt models: - Phi3-mini - StableLM-2-zephyr-1.6B - Qwen1.5-1.8B - Hermes2-Pro-Llama-3-8B to prebuilt models - Updates on `ModelRecord` fields - For detail see: mlc-ai#435 - Update all WASMs - For detail see: mlc-ai#433 - Update all WASMs to v0.2.39 - Support grammar for Llama3, hence update examples/json-mode to use `Llama3` and `Hermes2-pro-Llama3-8B` for function calling in `examples/json-schema` - Use `loglevel` package: - For details see mlc-ai#427 - Fix `index.js.map` issue for Vite - mlc-ai#420 - Enhance error handling and ServiceWorker ### TVMjs TVMjs compiled at apache/tvm@71f7af7 - Main changes include: - apache/tvm#17031 - apache/tvm#17028 - apache/tvm#17021 ### WASM version - All wasms updated to 0.2.39 via mlc-ai/binary-mlc-llm-libs#123 for new MLC-LLM runtime (mainly grammar)
atebites-hub pushed a commit to atebites-hub/web-llm that referenced this pull request Oct 4, 2025
…lc-ai#433) This PR updates models to v0.2.39 compiled with mlc-ai/binary-mlc-llm-libs#123 The main change is the new MLC-LLM runtime, which supports grammar (i.e. json mode) for Llama3. - Hence we now read in field `tokenizer_info` (or deprecated `token_table_postproc_method`) from `mlc-chat-config.json` when post processing token table for Grammar - If neither is available, we use the default `byte_fallback` New prebuilt models introduced: - Phi3-mini-4k - Hermes-2-Pro-Llama-3-8B - Qwen1.5-1.8B - StableLM-2-zephyr_1.6B Updates on examples: - json-mode and json-schema now use Llama3 to demonstrate - Function calling inside json-schema now uses `Hermes-2-Pro-Llama-3-8B` instead of `Hermes-2-Pro-Mistral`
atebites-hub pushed a commit to atebites-hub/web-llm that referenced this pull request Oct 4, 2025
### Changes Main changes include: - New prebuilt models: - Phi3-mini - StableLM-2-zephyr-1.6B - Qwen1.5-1.8B - Hermes2-Pro-Llama-3-8B to prebuilt models - Updates on `ModelRecord` fields - For detail see: mlc-ai#435 - Update all WASMs - For detail see: mlc-ai#433 - Update all WASMs to v0.2.39 - Support grammar for Llama3, hence update examples/json-mode to use `Llama3` and `Hermes2-pro-Llama3-8B` for function calling in `examples/json-schema` - Use `loglevel` package: - For details see mlc-ai#427 - Fix `index.js.map` issue for Vite - mlc-ai#420 - Enhance error handling and ServiceWorker ### TVMjs TVMjs compiled at apache/tvm@71f7af7 - Main changes include: - apache/tvm#17031 - apache/tvm#17028 - apache/tvm#17021 ### WASM version - All wasms updated to 0.2.39 via mlc-ai/binary-mlc-llm-libs#123 for new MLC-LLM runtime (mainly grammar)
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

1 participant