Class: LlamaCpp::LlamaModelQuantizeParams

Inherits:
Object
  • Object
show all
Defined in:
ext/llama_cpp/llama_cpp.c,
ext/llama_cpp/llama_cpp.c

Overview

“struct llama_model_quantize_params” wrapper class

Instance Method Summary collapse

Instance Method Details

#allow_requantizeBoolean

Returns:

  • (Boolean)


863
864
865
866
# File 'ext/llama_cpp/llama_cpp.c', line 863

static VALUE llama_model_quantize_params_get_allow_requantize(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return data->allow_requantize ? Qtrue : Qfalse;
}

#allow_requantize=(allow_requantize) ⇒ Boolean

Parameters:

  • allow_requantize (Boolean)

Returns:

  • (Boolean)


868
869
870
871
872
# File 'ext/llama_cpp/llama_cpp.c', line 868

static VALUE llama_model_quantize_params_set_allow_requantize(VALUE self, VALUE allow_requantize) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->allow_requantize = RTEST(allow_requantize) ? true : false;
  return allow_requantize;
}

#ftypeInteger

Returns:

  • (Integer)


830
831
832
833
# File 'ext/llama_cpp/llama_cpp.c', line 830

static VALUE llama_model_quantize_params_get_ftype(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return INT2NUM(data->ftype);
}

#ftype=(ftype) ⇒ Integer

Parameters:

  • ftype (Integer)

Returns:

  • (Integer)


835
836
837
838
839
# File 'ext/llama_cpp/llama_cpp.c', line 835

static VALUE llama_model_quantize_params_set_ftype(VALUE self, VALUE ftype) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->ftype = (enum llama_ftype)NUM2INT(ftype);
  return ftype;
}

#keep_splitBoolean

Returns:

  • (Boolean)


907
908
909
910
# File 'ext/llama_cpp/llama_cpp.c', line 907

static VALUE llama_model_quantize_params_get_keep_split(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return data->keep_split ? Qtrue : Qfalse;
}

#keep_split=(keep_split) ⇒ Boolean

Parameters:

  • keep_split (Boolean)

Returns:

  • (Boolean)


912
913
914
915
916
# File 'ext/llama_cpp/llama_cpp.c', line 912

static VALUE llama_model_quantize_params_set_keep_split(VALUE self, VALUE keep_split) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->keep_split = RTEST(keep_split) ? true : false;
  return keep_split;
}

#nthreadInteger

Returns:

  • (Integer)


819
820
821
822
# File 'ext/llama_cpp/llama_cpp.c', line 819

static VALUE llama_model_quantize_params_get_nthread(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return INT2NUM(data->nthread);
}

#nthread=(nthread) ⇒ Integer

Parameters:

  • nthread (Integer)

Returns:

  • (Integer)


824
825
826
827
828
# File 'ext/llama_cpp/llama_cpp.c', line 824

static VALUE llama_model_quantize_params_set_nthread(VALUE self, VALUE nthread) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->nthread = NUM2INT(nthread);
  return nthread;
}

#only_copyBoolean

Returns:

  • (Boolean)


885
886
887
888
# File 'ext/llama_cpp/llama_cpp.c', line 885

static VALUE llama_model_quantize_params_get_only_copy(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return data->only_copy ? Qtrue : Qfalse;
}

#only_copy=(only_copy) ⇒ Boolean

Parameters:

  • only_copy (Boolean)

Returns:

  • (Boolean)


890
891
892
893
894
# File 'ext/llama_cpp/llama_cpp.c', line 890

static VALUE llama_model_quantize_params_set_only_copy(VALUE self, VALUE only_copy) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->only_copy = RTEST(only_copy) ? true : false;
  return only_copy;
}

#output_tensor_typeInteger

Returns:

  • (Integer)


841
842
843
844
# File 'ext/llama_cpp/llama_cpp.c', line 841

static VALUE llama_model_quantize_params_get_output_tensor_type(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return INT2NUM(data->output_tensor_type);
}

#output_tensor_type=(output_tensor_type) ⇒ Integer

Parameters:

  • output_tensor_type (Integer)

Returns:

  • (Integer)


846
847
848
849
850
# File 'ext/llama_cpp/llama_cpp.c', line 846

static VALUE llama_model_quantize_params_set_output_tensor_type(VALUE self, VALUE output_tensor_type) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->output_tensor_type = (enum ggml_type)NUM2INT(output_tensor_type);
  return output_tensor_type;
}

#pureBoolean

Returns:

  • (Boolean)


896
897
898
899
# File 'ext/llama_cpp/llama_cpp.c', line 896

static VALUE llama_model_quantize_params_get_pure(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return data->pure ? Qtrue : Qfalse;
}

#pure=(pure) ⇒ Boolean

Parameters:

  • pure (Boolean)

Returns:

  • (Boolean)


901
902
903
904
905
# File 'ext/llama_cpp/llama_cpp.c', line 901

static VALUE llama_model_quantize_params_set_pure(VALUE self, VALUE pure) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->pure = RTEST(pure) ? true : false;
  return pure;
}

#quantize_output_tensorBoolean

Returns:

  • (Boolean)


874
875
876
877
# File 'ext/llama_cpp/llama_cpp.c', line 874

static VALUE llama_model_quantize_params_get_quantize_output_tensor(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return data->quantize_output_tensor ? Qtrue : Qfalse;
}

#quantize_output_tensor=(quantize_output_tensor) ⇒ Boolean

Parameters:

  • quantize_output_tensor (Boolean)

Returns:

  • (Boolean)


879
880
881
882
883
# File 'ext/llama_cpp/llama_cpp.c', line 879

static VALUE llama_model_quantize_params_set_quantize_output_tensor(VALUE self, VALUE quantize_output_tensor) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->quantize_output_tensor = RTEST(quantize_output_tensor) ? true : false;
  return quantize_output_tensor;
}

#token_embedding_typeInteger

Returns:

  • (Integer)


852
853
854
855
# File 'ext/llama_cpp/llama_cpp.c', line 852

static VALUE llama_model_quantize_params_get_token_embedding_type(VALUE self) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  return INT2NUM(data->token_embedding_type);
}

#token_embedding_type=(token_embedding_type) ⇒ Integer

Parameters:

  • token_embedding_type (Integer)

Returns:

  • (Integer)


857
858
859
860
861
# File 'ext/llama_cpp/llama_cpp.c', line 857

static VALUE llama_model_quantize_params_set_token_embedding_type(VALUE self, VALUE token_embedding_type) {
  llama_model_quantize_params* data = get_llama_model_quantize_params(self);
  data->token_embedding_type = (enum ggml_type)NUM2INT(token_embedding_type);
  return token_embedding_type;
}