Fix/4742 ollama num gpu option not consistent with allowed values (#4751)

This commit is contained in:
crazywoola 2024-05-29 13:33:35 +08:00 committed by GitHub
parent f4a240d225
commit 705a6e3a8e
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194

View File

@ -534,12 +534,14 @@ class OllamaLargeLanguageModel(LargeLanguageModel):
), ),
ParameterRule( ParameterRule(
name='num_gpu', name='num_gpu',
label=I18nObject(en_US="Num GPU"), label=I18nObject(en_US="GPU Layers"),
type=ParameterType.INT, type=ParameterType.INT,
help=I18nObject(en_US="The number of layers to send to the GPU(s). " help=I18nObject(en_US="The number of layers to offload to the GPU(s). "
"On macOS it defaults to 1 to enable metal support, 0 to disable."), "On macOS it defaults to 1 to enable metal support, 0 to disable."
min=0, "As long as a model fits into one gpu it stays in one. "
max=1 "It does not set the number of GPU(s). "),
min=-1,
default=1
), ),
ParameterRule( ParameterRule(
name='num_thread', name='num_thread',