release v1.7.3
This commit is contained in:
parent
bb1a6191b0
commit
2604d3c47b
@ -2,16 +2,24 @@
|
|||||||
|
|
||||||
### Features
|
### Features
|
||||||
|
|
||||||
- allow setting tokenChunkSize of WebGPU mode
|
- add Docker support (#291) @LonghronShen
|
||||||
- expose global_penalty
|
|
||||||
|
### Fixes
|
||||||
|
|
||||||
|
- fix a generation exception caused by potentially dangerous regex being passed into the stop array
|
||||||
|
- fix max_tokens parameter of Chat page not being passed to backend
|
||||||
|
- fix the issue where penalty_decay and global_penalty are not being passed to the backend default config when running
|
||||||
|
the model through client
|
||||||
|
|
||||||
### Improvements
|
### Improvements
|
||||||
|
|
||||||
- improve parameters controllable range
|
- prevent 'torch' has no attribute 'cuda' error in torch_gc, so user can use CPU or WebGPU (#302)
|
||||||
|
|
||||||
### Chores
|
### Chores
|
||||||
|
|
||||||
- update defaultModelConfigs
|
- bump dependencies
|
||||||
|
- add pre-release workflow
|
||||||
|
- dep_check.py now ignores GPUtil
|
||||||
|
|
||||||
## Install
|
## Install
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user