Skip to content

Conversation

honghua
Copy link

@honghua honghua commented Sep 27, 2024

No description provided.

Harry Yang added 2 commits September 27, 2024 20:36
1. on my VM running inference with CPU is only using one core, instead of all 16. delete `torch.set_default_tensor_type(torch.BFloat16Tensor)` make it using all 16 cores 2. reduce default max_seq_len to 128
* `device` is undefined
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
1 participant