| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331 |
- #pragma once
- #include "ggml.h"
- #include "clip.h"
- #include "clip-impl.h"
- #include <array>
- #include <vector>
- #include <unordered_set>
- #include <cstdint>
- #include <cmath>
- enum ffn_op_type {
- FFN_GELU,
- FFN_GELU_ERF,
- FFN_SILU,
- FFN_GELU_QUICK,
- };
- enum norm_type {
- NORM_TYPE_NORMAL,
- NORM_TYPE_RMS,
- };
- enum patch_merge_type {
- PATCH_MERGE_FLAT,
- PATCH_MERGE_SPATIAL_UNPAD,
- };
- struct clip_hparams {
- int32_t image_size = 0;
- int32_t patch_size = 0;
- int32_t n_embd = 0;
- int32_t n_ff = 0;
- int32_t projection_dim = 0;
- int32_t n_head = 0;
- int32_t n_layer = 0;
- // idefics3
- int32_t image_longest_edge = 0;
- int32_t image_min_pixels = -1;
- int32_t image_max_pixels = -1;
- int32_t n_merge = 0; // number of patch merges **per-side**
- float image_mean[3];
- float image_std[3];
- // for models using dynamic image size, we need to have a smaller image size to warmup
- // otherwise, user will get OOM everytime they load the model
- int32_t warmup_image_size = 0;
- int32_t warmup_audio_size = 3000;
- ffn_op_type ffn_op = FFN_GELU;
- patch_merge_type mm_patch_merge_type = PATCH_MERGE_FLAT;
- float eps = 1e-6;
- float rope_theta = 0.0;
- std::vector<clip_image_size> image_res_candidates; // for llava-uhd style models
- int32_t image_crop_resolution;
- std::unordered_set<int32_t> vision_feature_layer;
- int32_t attn_window_size = 0;
- int32_t n_wa_pattern = 0;
- // audio
- int32_t n_mel_bins = 0; // whisper preprocessor
- int32_t proj_stack_factor = 0; // ultravox
- // audio-to-mel preprocessor params
- int32_t audio_chunk_len = -1; // in seconds
- int32_t audio_sample_rate = -1;
- int32_t audio_n_fft = -1;
- int32_t audio_window_len = -1;
- int32_t audio_hop_len = -1;
- // legacy
- bool has_llava_projector = false;
- int minicpmv_version = 0;
- int32_t minicpmv_query_num = 0; // MiniCPM-V query number
- // custom value provided by user, can be undefined if not set
- int32_t custom_image_min_tokens = -1;
- int32_t custom_image_max_tokens = -1;
- void set_limit_image_tokens(int n_tokens_min, int n_tokens_max) {
- const int cur_merge = n_merge == 0 ? 1 : n_merge;
- const int patch_area = patch_size * patch_size * cur_merge * cur_merge;
- image_min_pixels = (custom_image_min_tokens > 0 ? custom_image_min_tokens : n_tokens_min) * patch_area;
- image_max_pixels = (custom_image_max_tokens > 0 ? custom_image_max_tokens : n_tokens_max) * patch_area;
- warmup_image_size = static_cast<int>(std::sqrt(image_max_pixels));
- }
- void set_warmup_n_tokens(int n_tokens) {
- int n_tok_per_side = static_cast<int>(std::sqrt(n_tokens));
- GGML_ASSERT(n_tok_per_side * n_tok_per_side == n_tokens && "n_tokens must be n*n");
- const int cur_merge = n_merge == 0 ? 1 : n_merge;
- warmup_image_size = n_tok_per_side * patch_size * cur_merge;
- // TODO: support warmup size for custom token numbers
- }
- };
- struct clip_layer {
- // attention
- ggml_tensor * k_w = nullptr;
- ggml_tensor * k_b = nullptr;
- ggml_tensor * q_w = nullptr;
- ggml_tensor * q_b = nullptr;
- ggml_tensor * v_w = nullptr;
- ggml_tensor * v_b = nullptr;
- ggml_tensor * qkv_w = nullptr;
- ggml_tensor * qkv_b = nullptr;
- ggml_tensor * o_w = nullptr;
- ggml_tensor * o_b = nullptr;
- ggml_tensor * k_norm = nullptr;
- ggml_tensor * q_norm = nullptr;
- // layernorm 1
- ggml_tensor * ln_1_w = nullptr;
- ggml_tensor * ln_1_b = nullptr;
- ggml_tensor * ff_up_w = nullptr;
- ggml_tensor * ff_up_b = nullptr;
- ggml_tensor * ff_gate_w = nullptr;
- ggml_tensor * ff_gate_b = nullptr;
- ggml_tensor * ff_down_w = nullptr;
- ggml_tensor * ff_down_b = nullptr;
- // layernorm 2
- ggml_tensor * ln_2_w = nullptr;
- ggml_tensor * ln_2_b = nullptr;
- // layer scale (no bias)
- ggml_tensor * ls_1_w = nullptr;
- ggml_tensor * ls_2_w = nullptr;
- // qwen3vl deepstack merger
- ggml_tensor * deepstack_norm_w = nullptr;
- ggml_tensor * deepstack_norm_b = nullptr;
- ggml_tensor * deepstack_fc1_w = nullptr;
- ggml_tensor * deepstack_fc1_b = nullptr;
- ggml_tensor * deepstack_fc2_w = nullptr;
- ggml_tensor * deepstack_fc2_b = nullptr;
- // lfm2
- ggml_tensor * ff_norm_w = nullptr;
- ggml_tensor * ff_norm_b = nullptr;
- ggml_tensor * ff_norm_1_w = nullptr;
- ggml_tensor * ff_norm_1_b = nullptr;
- ggml_tensor * ff_up_1_w = nullptr;
- ggml_tensor * ff_up_1_b = nullptr;
- ggml_tensor * ff_down_1_w = nullptr;
- ggml_tensor * ff_down_1_b = nullptr;
- ggml_tensor * pos_bias_u = nullptr;
- ggml_tensor * pos_bias_v = nullptr;
- ggml_tensor * norm_conv_w = nullptr;
- ggml_tensor * norm_conv_b = nullptr;
- ggml_tensor * linear_pos_w = nullptr;
- ggml_tensor * conv_norm_w = nullptr;
- ggml_tensor * conv_norm_b = nullptr;
- ggml_tensor * conv_dw_w = nullptr;
- ggml_tensor * conv_dw_b = nullptr;
- ggml_tensor * conv_pw1_w = nullptr;
- ggml_tensor * conv_pw1_b = nullptr;
- ggml_tensor * conv_pw2_w = nullptr;
- ggml_tensor * conv_pw2_b = nullptr;
- bool has_deepstack() const {
- return deepstack_fc1_w != nullptr;
- }
- };
- struct clip_model {
- clip_modality modality = CLIP_MODALITY_VISION;
- projector_type proj_type = PROJECTOR_TYPE_MLP;
- clip_hparams hparams;
- // embeddings
- ggml_tensor * class_embedding = nullptr;
- ggml_tensor * patch_embeddings_0 = nullptr;
- ggml_tensor * patch_embeddings_1 = nullptr; // second Conv2D kernel when we decouple Conv3D along temproal dimension (Qwen2VL)
- ggml_tensor * patch_bias = nullptr;
- ggml_tensor * position_embeddings = nullptr;
- ggml_tensor * norm_embd_w = nullptr;
- ggml_tensor * norm_embd_b = nullptr;
- ggml_tensor * pre_ln_w = nullptr;
- ggml_tensor * pre_ln_b = nullptr;
- std::vector<clip_layer> layers;
- int32_t n_deepstack_layers = 0; // used by Qwen3-VL, calculated from clip_layer
- ggml_tensor * post_ln_w;
- ggml_tensor * post_ln_b;
- ggml_tensor * projection; // TODO: rename it to fc (fully connected layer)
- ggml_tensor * mm_fc_w;
- ggml_tensor * mm_fc_b;
- ggml_tensor * mm_ffn_up_w = nullptr;
- ggml_tensor * mm_ffn_up_b = nullptr;
- ggml_tensor * mm_ffn_gate_w = nullptr;
- ggml_tensor * mm_ffn_gate_b = nullptr;
- ggml_tensor * mm_ffn_down_w = nullptr;
- ggml_tensor * mm_ffn_down_b = nullptr;
- ggml_tensor * mm_post_norm_w = nullptr;
- ggml_tensor * mm_post_norm_b = nullptr;
- // LLaVA projection
- ggml_tensor * mm_input_norm_w = nullptr;
- ggml_tensor * mm_input_norm_b = nullptr;
- ggml_tensor * mm_0_w = nullptr;
- ggml_tensor * mm_0_b = nullptr;
- ggml_tensor * mm_2_w = nullptr;
- ggml_tensor * mm_2_b = nullptr;
- ggml_tensor * image_newline = nullptr;
- // Yi type models with mlp+normalization projection
- ggml_tensor * mm_1_w = nullptr; // Yi type models have 0, 1, 3, 4
- ggml_tensor * mm_1_b = nullptr;
- ggml_tensor * mm_3_w = nullptr;
- ggml_tensor * mm_3_b = nullptr;
- ggml_tensor * mm_4_w = nullptr;
- ggml_tensor * mm_4_b = nullptr;
- // GLMV-Edge projection
- ggml_tensor * mm_model_adapter_conv_w = nullptr;
- ggml_tensor * mm_model_adapter_conv_b = nullptr;
- // MobileVLM projection
- ggml_tensor * mm_model_mlp_1_w = nullptr;
- ggml_tensor * mm_model_mlp_1_b = nullptr;
- ggml_tensor * mm_model_mlp_3_w = nullptr;
- ggml_tensor * mm_model_mlp_3_b = nullptr;
- ggml_tensor * mm_model_block_1_block_0_0_w = nullptr;
- ggml_tensor * mm_model_block_1_block_0_1_w = nullptr;
- ggml_tensor * mm_model_block_1_block_0_1_b = nullptr;
- ggml_tensor * mm_model_block_1_block_1_fc1_w = nullptr;
- ggml_tensor * mm_model_block_1_block_1_fc1_b = nullptr;
- ggml_tensor * mm_model_block_1_block_1_fc2_w = nullptr;
- ggml_tensor * mm_model_block_1_block_1_fc2_b = nullptr;
- ggml_tensor * mm_model_block_1_block_2_0_w = nullptr;
- ggml_tensor * mm_model_block_1_block_2_1_w = nullptr;
- ggml_tensor * mm_model_block_1_block_2_1_b = nullptr;
- ggml_tensor * mm_model_block_2_block_0_0_w = nullptr;
- ggml_tensor * mm_model_block_2_block_0_1_w = nullptr;
- ggml_tensor * mm_model_block_2_block_0_1_b = nullptr;
- ggml_tensor * mm_model_block_2_block_1_fc1_w = nullptr;
- ggml_tensor * mm_model_block_2_block_1_fc1_b = nullptr;
- ggml_tensor * mm_model_block_2_block_1_fc2_w = nullptr;
- ggml_tensor * mm_model_block_2_block_1_fc2_b = nullptr;
- ggml_tensor * mm_model_block_2_block_2_0_w = nullptr;
- ggml_tensor * mm_model_block_2_block_2_1_w = nullptr;
- ggml_tensor * mm_model_block_2_block_2_1_b = nullptr;
- // MobileVLM_V2 projection
- ggml_tensor * mm_model_mlp_0_w = nullptr;
- ggml_tensor * mm_model_mlp_0_b = nullptr;
- ggml_tensor * mm_model_mlp_2_w = nullptr;
- ggml_tensor * mm_model_mlp_2_b = nullptr;
- ggml_tensor * mm_model_peg_0_w = nullptr;
- ggml_tensor * mm_model_peg_0_b = nullptr;
- // MINICPMV projection
- ggml_tensor * mm_model_pos_embed_k = nullptr;
- ggml_tensor * mm_model_query = nullptr;
- ggml_tensor * mm_model_proj = nullptr;
- ggml_tensor * mm_model_kv_proj = nullptr;
- ggml_tensor * mm_model_attn_q_w = nullptr;
- ggml_tensor * mm_model_attn_q_b = nullptr;
- ggml_tensor * mm_model_attn_k_w = nullptr;
- ggml_tensor * mm_model_attn_k_b = nullptr;
- ggml_tensor * mm_model_attn_v_w = nullptr;
- ggml_tensor * mm_model_attn_v_b = nullptr;
- ggml_tensor * mm_model_attn_o_w = nullptr;
- ggml_tensor * mm_model_attn_o_b = nullptr;
- ggml_tensor * mm_model_ln_q_w = nullptr;
- ggml_tensor * mm_model_ln_q_b = nullptr;
- ggml_tensor * mm_model_ln_kv_w = nullptr;
- ggml_tensor * mm_model_ln_kv_b = nullptr;
- ggml_tensor * mm_model_ln_post_w = nullptr;
- ggml_tensor * mm_model_ln_post_b = nullptr;
- // gemma3
- ggml_tensor * mm_input_proj_w = nullptr;
- ggml_tensor * mm_soft_emb_norm_w = nullptr;
- // pixtral, glm4v
- ggml_tensor * token_embd_img_break = nullptr;
- ggml_tensor * mm_patch_merger_w = nullptr;
- ggml_tensor * mm_patch_merger_b = nullptr;
- // ultravox / whisper encoder
- ggml_tensor * conv1d_1_w = nullptr;
- ggml_tensor * conv1d_1_b = nullptr;
- ggml_tensor * conv1d_2_w = nullptr;
- ggml_tensor * conv1d_2_b = nullptr;
- ggml_tensor * mm_norm_pre_w = nullptr;
- ggml_tensor * mm_norm_pre_b = nullptr;
- ggml_tensor * mm_norm_mid_w = nullptr;
- // cogvlm
- ggml_tensor * mm_post_fc_norm_w = nullptr;
- ggml_tensor * mm_post_fc_norm_b = nullptr;
- ggml_tensor * mm_h_to_4h_w = nullptr;
- ggml_tensor * mm_gate_w = nullptr;
- ggml_tensor * mm_4h_to_h_w = nullptr;
- ggml_tensor * mm_boi = nullptr;
- ggml_tensor * mm_eoi = nullptr;
- // lfm2 audio
- std::array<ggml_tensor *, 7> pre_encode_conv_X_w = {nullptr};
- std::array<ggml_tensor *, 7> pre_encode_conv_X_b = {nullptr};
- ggml_tensor * pre_encode_out_w = nullptr;
- ggml_tensor * pre_encode_out_b = nullptr;
- bool audio_has_avgpool() const {
- return proj_type == PROJECTOR_TYPE_QWEN2A
- || proj_type == PROJECTOR_TYPE_VOXTRAL;
- }
- bool audio_has_stack_frames() const {
- return proj_type == PROJECTOR_TYPE_ULTRAVOX
- || proj_type == PROJECTOR_TYPE_VOXTRAL;
- }
- };
- const clip_hparams * clip_get_hparams(const struct clip_ctx * ctx);
|