Skip to content

Conversation

honghua
Copy link

@honghua honghua commented Sep 27, 2024

No description provided.

Harry Yang added 2 commits September 27, 2024 20:36
1. on my VM running inference with CPU is only using one core, instead of all 16.
delete `torch.set_default_tensor_type(torch.BFloat16Tensor)` make it using
all 16 cores

2. reduce default max_seq_len to 128
* `device` is undefined
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant