[Fix] Fix clean_up_tokenization_spaces in tokenizer (#1510)
This commit is contained in:
@@ -129,6 +129,7 @@ def get_tokenizer(
|
||||
*args,
|
||||
trust_remote_code=trust_remote_code,
|
||||
tokenizer_revision=tokenizer_revision,
|
||||
clean_up_tokenization_spaces=False,
|
||||
**kwargs,
|
||||
)
|
||||
except TypeError as e:
|
||||
|
||||
Reference in New Issue
Block a user