We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4e3c319 commit 2f5feb4Copy full SHA for 2f5feb4
api/config.go
@@ -35,6 +35,7 @@ type Config struct {
35
NGPULayers int `yaml:"gpu_layers"`
36
MMap bool `yaml:"mmap"`
37
MMlock bool `yaml:"mmlock"`
38
+ LowVRAM bool `yaml:"low_vram"`
39
40
TensorSplit string `yaml:"tensor_split"`
41
MainGPU string `yaml:"main_gpu"`
api/prediction.go
@@ -48,6 +48,10 @@ func defaultLLamaOpts(c Config) []llama.ModelOption {
48
llamaOpts = append(llamaOpts, llama.SetNBatch(512))
49
}
50
51
+ if c.LowVRAM {
52
+ llamaOpts = append(llamaOpts, llama.EnabelLowVRAM)
53
+ }
54
+
55
return llamaOpts
56
57
0 commit comments