Class: LlamaCpp::LlamaModelQuantizeParams
- Inherits:
-
Object
- Object
- LlamaCpp::LlamaModelQuantizeParams
- Defined in:
- ext/llama_cpp/llama_cpp.c,
ext/llama_cpp/llama_cpp.c
Overview
“struct llama_model_quantize_params” wrapper class
Instance Method Summary collapse
- #allow_requantize ⇒ Boolean
- #allow_requantize=(allow_requantize) ⇒ Boolean
- #ftype ⇒ Integer
- #ftype=(ftype) ⇒ Integer
- #keep_split ⇒ Boolean
- #keep_split=(keep_split) ⇒ Boolean
- #nthread ⇒ Integer
- #nthread=(nthread) ⇒ Integer
- #only_copy ⇒ Boolean
- #only_copy=(only_copy) ⇒ Boolean
- #output_tensor_type ⇒ Integer
- #output_tensor_type=(output_tensor_type) ⇒ Integer
- #pure ⇒ Boolean
- #pure=(pure) ⇒ Boolean
- #quantize_output_tensor ⇒ Boolean
- #quantize_output_tensor=(quantize_output_tensor) ⇒ Boolean
- #token_embedding_type ⇒ Integer
- #token_embedding_type=(token_embedding_type) ⇒ Integer
Instance Method Details
#allow_requantize ⇒ Boolean
862 863 864 865 |
# File 'ext/llama_cpp/llama_cpp.c', line 862
static VALUE llama_model_quantize_params_get_allow_requantize(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->allow_requantize ? Qtrue : Qfalse;
}
|
#allow_requantize=(allow_requantize) ⇒ Boolean
867 868 869 870 871 |
# File 'ext/llama_cpp/llama_cpp.c', line 867
static VALUE llama_model_quantize_params_set_allow_requantize(VALUE self, VALUE allow_requantize) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->allow_requantize = RTEST(allow_requantize) ? true : false;
return allow_requantize;
}
|
#ftype ⇒ Integer
829 830 831 832 |
# File 'ext/llama_cpp/llama_cpp.c', line 829
static VALUE llama_model_quantize_params_get_ftype(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->ftype);
}
|
#ftype=(ftype) ⇒ Integer
834 835 836 837 838 |
# File 'ext/llama_cpp/llama_cpp.c', line 834
static VALUE llama_model_quantize_params_set_ftype(VALUE self, VALUE ftype) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->ftype = (enum llama_ftype)NUM2INT(ftype);
return ftype;
}
|
#keep_split ⇒ Boolean
906 907 908 909 |
# File 'ext/llama_cpp/llama_cpp.c', line 906
static VALUE llama_model_quantize_params_get_keep_split(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->keep_split ? Qtrue : Qfalse;
}
|
#keep_split=(keep_split) ⇒ Boolean
911 912 913 914 915 |
# File 'ext/llama_cpp/llama_cpp.c', line 911
static VALUE llama_model_quantize_params_set_keep_split(VALUE self, VALUE keep_split) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->keep_split = RTEST(keep_split) ? true : false;
return keep_split;
}
|
#nthread ⇒ Integer
818 819 820 821 |
# File 'ext/llama_cpp/llama_cpp.c', line 818
static VALUE llama_model_quantize_params_get_nthread(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->nthread);
}
|
#nthread=(nthread) ⇒ Integer
823 824 825 826 827 |
# File 'ext/llama_cpp/llama_cpp.c', line 823
static VALUE llama_model_quantize_params_set_nthread(VALUE self, VALUE nthread) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->nthread = NUM2INT(nthread);
return nthread;
}
|
#only_copy ⇒ Boolean
884 885 886 887 |
# File 'ext/llama_cpp/llama_cpp.c', line 884
static VALUE llama_model_quantize_params_get_only_copy(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->only_copy ? Qtrue : Qfalse;
}
|
#only_copy=(only_copy) ⇒ Boolean
889 890 891 892 893 |
# File 'ext/llama_cpp/llama_cpp.c', line 889
static VALUE llama_model_quantize_params_set_only_copy(VALUE self, VALUE only_copy) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->only_copy = RTEST(only_copy) ? true : false;
return only_copy;
}
|
#output_tensor_type ⇒ Integer
840 841 842 843 |
# File 'ext/llama_cpp/llama_cpp.c', line 840
static VALUE llama_model_quantize_params_get_output_tensor_type(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->output_tensor_type);
}
|
#output_tensor_type=(output_tensor_type) ⇒ Integer
845 846 847 848 849 |
# File 'ext/llama_cpp/llama_cpp.c', line 845
static VALUE llama_model_quantize_params_set_output_tensor_type(VALUE self, VALUE output_tensor_type) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->output_tensor_type = (enum ggml_type)NUM2INT(output_tensor_type);
return output_tensor_type;
}
|
#pure ⇒ Boolean
895 896 897 898 |
# File 'ext/llama_cpp/llama_cpp.c', line 895
static VALUE llama_model_quantize_params_get_pure(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->pure ? Qtrue : Qfalse;
}
|
#pure=(pure) ⇒ Boolean
900 901 902 903 904 |
# File 'ext/llama_cpp/llama_cpp.c', line 900
static VALUE llama_model_quantize_params_set_pure(VALUE self, VALUE pure) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->pure = RTEST(pure) ? true : false;
return pure;
}
|
#quantize_output_tensor ⇒ Boolean
873 874 875 876 |
# File 'ext/llama_cpp/llama_cpp.c', line 873
static VALUE llama_model_quantize_params_get_quantize_output_tensor(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->quantize_output_tensor ? Qtrue : Qfalse;
}
|
#quantize_output_tensor=(quantize_output_tensor) ⇒ Boolean
878 879 880 881 882 |
# File 'ext/llama_cpp/llama_cpp.c', line 878
static VALUE llama_model_quantize_params_set_quantize_output_tensor(VALUE self, VALUE quantize_output_tensor) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->quantize_output_tensor = RTEST(quantize_output_tensor) ? true : false;
return quantize_output_tensor;
}
|
#token_embedding_type ⇒ Integer
851 852 853 854 |
# File 'ext/llama_cpp/llama_cpp.c', line 851
static VALUE llama_model_quantize_params_get_token_embedding_type(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->token_embedding_type);
}
|
#token_embedding_type=(token_embedding_type) ⇒ Integer
856 857 858 859 860 |
# File 'ext/llama_cpp/llama_cpp.c', line 856
static VALUE llama_model_quantize_params_set_token_embedding_type(VALUE self, VALUE token_embedding_type) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->token_embedding_type = (enum ggml_type)NUM2INT(token_embedding_type);
return token_embedding_type;
}
|