Class: LlamaCpp::LlamaModelQuantizeParams
- Inherits:
-
Object
- Object
- LlamaCpp::LlamaModelQuantizeParams
- Defined in:
- ext/llama_cpp/llama_cpp.c,
ext/llama_cpp/llama_cpp.c
Overview
“struct llama_model_quantize_params” wrapper class
Instance Method Summary collapse
- #allow_requantize ⇒ Boolean
- #allow_requantize=(allow_requantize) ⇒ Boolean
- #ftype ⇒ Integer
- #ftype=(ftype) ⇒ Integer
- #keep_split ⇒ Boolean
- #keep_split=(keep_split) ⇒ Boolean
- #nthread ⇒ Integer
- #nthread=(nthread) ⇒ Integer
- #only_copy ⇒ Boolean
- #only_copy=(only_copy) ⇒ Boolean
- #output_tensor_type ⇒ Integer
- #output_tensor_type=(output_tensor_type) ⇒ Integer
- #pure ⇒ Boolean
- #pure=(pure) ⇒ Boolean
- #quantize_output_tensor ⇒ Boolean
- #quantize_output_tensor=(quantize_output_tensor) ⇒ Boolean
- #token_embedding_type ⇒ Integer
- #token_embedding_type=(token_embedding_type) ⇒ Integer
Instance Method Details
#allow_requantize ⇒ Boolean
863 864 865 866 |
# File 'ext/llama_cpp/llama_cpp.c', line 863
static VALUE llama_model_quantize_params_get_allow_requantize(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->allow_requantize ? Qtrue : Qfalse;
}
|
#allow_requantize=(allow_requantize) ⇒ Boolean
868 869 870 871 872 |
# File 'ext/llama_cpp/llama_cpp.c', line 868
static VALUE llama_model_quantize_params_set_allow_requantize(VALUE self, VALUE allow_requantize) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->allow_requantize = RTEST(allow_requantize) ? true : false;
return allow_requantize;
}
|
#ftype ⇒ Integer
830 831 832 833 |
# File 'ext/llama_cpp/llama_cpp.c', line 830
static VALUE llama_model_quantize_params_get_ftype(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->ftype);
}
|
#ftype=(ftype) ⇒ Integer
835 836 837 838 839 |
# File 'ext/llama_cpp/llama_cpp.c', line 835
static VALUE llama_model_quantize_params_set_ftype(VALUE self, VALUE ftype) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->ftype = (enum llama_ftype)NUM2INT(ftype);
return ftype;
}
|
#keep_split ⇒ Boolean
907 908 909 910 |
# File 'ext/llama_cpp/llama_cpp.c', line 907
static VALUE llama_model_quantize_params_get_keep_split(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->keep_split ? Qtrue : Qfalse;
}
|
#keep_split=(keep_split) ⇒ Boolean
912 913 914 915 916 |
# File 'ext/llama_cpp/llama_cpp.c', line 912
static VALUE llama_model_quantize_params_set_keep_split(VALUE self, VALUE keep_split) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->keep_split = RTEST(keep_split) ? true : false;
return keep_split;
}
|
#nthread ⇒ Integer
819 820 821 822 |
# File 'ext/llama_cpp/llama_cpp.c', line 819
static VALUE llama_model_quantize_params_get_nthread(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->nthread);
}
|
#nthread=(nthread) ⇒ Integer
824 825 826 827 828 |
# File 'ext/llama_cpp/llama_cpp.c', line 824
static VALUE llama_model_quantize_params_set_nthread(VALUE self, VALUE nthread) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->nthread = NUM2INT(nthread);
return nthread;
}
|
#only_copy ⇒ Boolean
885 886 887 888 |
# File 'ext/llama_cpp/llama_cpp.c', line 885
static VALUE llama_model_quantize_params_get_only_copy(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->only_copy ? Qtrue : Qfalse;
}
|
#only_copy=(only_copy) ⇒ Boolean
890 891 892 893 894 |
# File 'ext/llama_cpp/llama_cpp.c', line 890
static VALUE llama_model_quantize_params_set_only_copy(VALUE self, VALUE only_copy) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->only_copy = RTEST(only_copy) ? true : false;
return only_copy;
}
|
#output_tensor_type ⇒ Integer
841 842 843 844 |
# File 'ext/llama_cpp/llama_cpp.c', line 841
static VALUE llama_model_quantize_params_get_output_tensor_type(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->output_tensor_type);
}
|
#output_tensor_type=(output_tensor_type) ⇒ Integer
846 847 848 849 850 |
# File 'ext/llama_cpp/llama_cpp.c', line 846
static VALUE llama_model_quantize_params_set_output_tensor_type(VALUE self, VALUE output_tensor_type) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->output_tensor_type = (enum ggml_type)NUM2INT(output_tensor_type);
return output_tensor_type;
}
|
#pure ⇒ Boolean
896 897 898 899 |
# File 'ext/llama_cpp/llama_cpp.c', line 896
static VALUE llama_model_quantize_params_get_pure(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->pure ? Qtrue : Qfalse;
}
|
#pure=(pure) ⇒ Boolean
901 902 903 904 905 |
# File 'ext/llama_cpp/llama_cpp.c', line 901
static VALUE llama_model_quantize_params_set_pure(VALUE self, VALUE pure) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->pure = RTEST(pure) ? true : false;
return pure;
}
|
#quantize_output_tensor ⇒ Boolean
874 875 876 877 |
# File 'ext/llama_cpp/llama_cpp.c', line 874
static VALUE llama_model_quantize_params_get_quantize_output_tensor(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return data->quantize_output_tensor ? Qtrue : Qfalse;
}
|
#quantize_output_tensor=(quantize_output_tensor) ⇒ Boolean
879 880 881 882 883 |
# File 'ext/llama_cpp/llama_cpp.c', line 879
static VALUE llama_model_quantize_params_set_quantize_output_tensor(VALUE self, VALUE quantize_output_tensor) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->quantize_output_tensor = RTEST(quantize_output_tensor) ? true : false;
return quantize_output_tensor;
}
|
#token_embedding_type ⇒ Integer
852 853 854 855 |
# File 'ext/llama_cpp/llama_cpp.c', line 852
static VALUE llama_model_quantize_params_get_token_embedding_type(VALUE self) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
return INT2NUM(data->token_embedding_type);
}
|
#token_embedding_type=(token_embedding_type) ⇒ Integer
857 858 859 860 861 |
# File 'ext/llama_cpp/llama_cpp.c', line 857
static VALUE llama_model_quantize_params_set_token_embedding_type(VALUE self, VALUE token_embedding_type) {
llama_model_quantize_params* data = get_llama_model_quantize_params(self);
data->token_embedding_type = (enum ggml_type)NUM2INT(token_embedding_type);
return token_embedding_type;
}
|