Class: LlamaCpp::LlamaModelQuantizeParams
- Inherits:
-
Object
- Object
- LlamaCpp::LlamaModelQuantizeParams
- Defined in:
- ext/llama_cpp/llama_cpp.c,
ext/llama_cpp/llama_cpp.c
Overview
“struct llama_model_quantize_params” wrapper class
Instance Method Summary collapse
- #allow_requantize ⇒ Boolean
- #allow_requantize=(allow_requantize) ⇒ Boolean
- #ftype ⇒ Integer
- #ftype=(ftype) ⇒ Integer
- #keep_split ⇒ Boolean
- #keep_split=(keep_split) ⇒ Boolean
- #nthread ⇒ Integer
- #nthread=(nthread) ⇒ Integer
- #only_copy ⇒ Boolean
- #only_copy=(only_copy) ⇒ Boolean
- #output_tensor_type ⇒ Integer
- #output_tensor_type=(output_tensor_type) ⇒ Integer
- #pure ⇒ Boolean
- #pure=(pure) ⇒ Boolean
- #quantize_output_tensor ⇒ Boolean
- #quantize_output_tensor=(quantize_output_tensor) ⇒ Boolean
- #token_embedding_type ⇒ Integer
- #token_embedding_type=(token_embedding_type) ⇒ Integer
Instance Method Details
#allow_requantize ⇒ Boolean
896 897 898 899 |
# File 'ext/llama_cpp/llama_cpp.c', line 896
static VALUE llama_model_quantize_params_get_allow_requantize(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->allow_requantize ? Qtrue : Qfalse;
}
|
#allow_requantize=(allow_requantize) ⇒ Boolean
901 902 903 904 905 |
# File 'ext/llama_cpp/llama_cpp.c', line 901
static VALUE llama_model_quantize_params_set_allow_requantize(VALUE self, VALUE allow_requantize) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->allow_requantize = RTEST(allow_requantize) ? true : false;
return allow_requantize;
}
|
#ftype ⇒ Integer
863 864 865 866 |
# File 'ext/llama_cpp/llama_cpp.c', line 863
static VALUE llama_model_quantize_params_get_ftype(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->ftype);
}
|
#ftype=(ftype) ⇒ Integer
868 869 870 871 872 |
# File 'ext/llama_cpp/llama_cpp.c', line 868
static VALUE llama_model_quantize_params_set_ftype(VALUE self, VALUE ftype) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->ftype = (enum llama_ftype)NUM2INT(ftype);
return ftype;
}
|
#keep_split ⇒ Boolean
940 941 942 943 |
# File 'ext/llama_cpp/llama_cpp.c', line 940
static VALUE llama_model_quantize_params_get_keep_split(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->keep_split ? Qtrue : Qfalse;
}
|
#keep_split=(keep_split) ⇒ Boolean
945 946 947 948 949 |
# File 'ext/llama_cpp/llama_cpp.c', line 945
static VALUE llama_model_quantize_params_set_keep_split(VALUE self, VALUE keep_split) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->keep_split = RTEST(keep_split) ? true : false;
return keep_split;
}
|
#nthread ⇒ Integer
852 853 854 855 |
# File 'ext/llama_cpp/llama_cpp.c', line 852
static VALUE llama_model_quantize_params_get_nthread(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->nthread);
}
|
#nthread=(nthread) ⇒ Integer
857 858 859 860 861 |
# File 'ext/llama_cpp/llama_cpp.c', line 857
static VALUE llama_model_quantize_params_set_nthread(VALUE self, VALUE nthread) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->nthread = NUM2INT(nthread);
return nthread;
}
|
#only_copy ⇒ Boolean
918 919 920 921 |
# File 'ext/llama_cpp/llama_cpp.c', line 918
static VALUE llama_model_quantize_params_get_only_copy(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->only_copy ? Qtrue : Qfalse;
}
|
#only_copy=(only_copy) ⇒ Boolean
923 924 925 926 927 |
# File 'ext/llama_cpp/llama_cpp.c', line 923
static VALUE llama_model_quantize_params_set_only_copy(VALUE self, VALUE only_copy) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->only_copy = RTEST(only_copy) ? true : false;
return only_copy;
}
|
#output_tensor_type ⇒ Integer
874 875 876 877 |
# File 'ext/llama_cpp/llama_cpp.c', line 874
static VALUE llama_model_quantize_params_get_output_tensor_type(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->output_tensor_type);
}
|
#output_tensor_type=(output_tensor_type) ⇒ Integer
879 880 881 882 883 |
# File 'ext/llama_cpp/llama_cpp.c', line 879
static VALUE llama_model_quantize_params_set_output_tensor_type(VALUE self, VALUE output_tensor_type) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->output_tensor_type = (enum ggml_type)NUM2INT(output_tensor_type);
return output_tensor_type;
}
|
#pure ⇒ Boolean
929 930 931 932 |
# File 'ext/llama_cpp/llama_cpp.c', line 929
static VALUE llama_model_quantize_params_get_pure(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->pure ? Qtrue : Qfalse;
}
|
#pure=(pure) ⇒ Boolean
934 935 936 937 938 |
# File 'ext/llama_cpp/llama_cpp.c', line 934
static VALUE llama_model_quantize_params_set_pure(VALUE self, VALUE pure) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->pure = RTEST(pure) ? true : false;
return pure;
}
|
#quantize_output_tensor ⇒ Boolean
907 908 909 910 |
# File 'ext/llama_cpp/llama_cpp.c', line 907
static VALUE llama_model_quantize_params_get_quantize_output_tensor(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->quantize_output_tensor ? Qtrue : Qfalse;
}
|
#quantize_output_tensor=(quantize_output_tensor) ⇒ Boolean
912 913 914 915 916 |
# File 'ext/llama_cpp/llama_cpp.c', line 912
static VALUE llama_model_quantize_params_set_quantize_output_tensor(VALUE self, VALUE quantize_output_tensor) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->quantize_output_tensor = RTEST(quantize_output_tensor) ? true : false;
return quantize_output_tensor;
}
|
#token_embedding_type ⇒ Integer
885 886 887 888 |
# File 'ext/llama_cpp/llama_cpp.c', line 885
static VALUE llama_model_quantize_params_get_token_embedding_type(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->token_embedding_type);
}
|
#token_embedding_type=(token_embedding_type) ⇒ Integer
890 891 892 893 894 |
# File 'ext/llama_cpp/llama_cpp.c', line 890
static VALUE llama_model_quantize_params_set_token_embedding_type(VALUE self, VALUE token_embedding_type) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->token_embedding_type = (enum ggml_type)NUM2INT(token_embedding_type);
return token_embedding_type;
}
|