EADST

QWEN7B to LLAMA GPTQ model structure

Here is the markdown format for the GPTQ model structure, detailing each layer and component:


GPTQ Model Structure

The GPTQ model consists of the following layers and components:

Embedding Layer

  • model.embed_tokens.weight: torch.Size([151851, 4096])

Layers

Each layer in the model has the following components:

Layer 0 to Layer 31

Each layer (model.layers.[0-31]) includes:

  • input_layernorm.weight: torch.Size([4096])

  • Self-Attention Sublayer:

    • k_proj:

      • qweight: torch.Size([512, 4096])

      • qzeros: torch.Size([32, 512])

      • scales: torch.Size([32, 4096])

      • g_idx: torch.Size([4096])

      • bias: torch.Size([4096])

    • o_proj:

      • qweight: torch.Size([512, 4096])

      • qzeros: torch.Size([32, 512])

      • scales: torch.Size([32, 4096])

      • g_idx: torch.Size([4096])

      • bias: torch.Size([4096])

    • q_proj:

      • qweight: torch.Size([512, 4096])

      • qzeros: torch.Size([32, 512])

      • scales: torch.Size([32, 4096])

      • g_idx: torch.Size([4096])

      • bias: torch.Size([4096])

    • v_proj:

      • qweight: torch.Size([512, 4096])

      • qzeros: torch.Size([32, 512])

      • scales: torch.Size([32, 4096])

      • g_idx: torch.Size([4096])

      • bias: torch.Size([4096])

  • MLP (Multi-Layer Perceptron) Sublayer:

    • down_proj:

      • qweight: torch.Size([1376, 4096])

      • qzeros: torch.Size([86, 512])

      • scales: torch.Size([86, 4096])

      • g_idx: torch.Size([11008])

      • bias: torch.Size([4096])

    • gate_proj:

      • qweight: torch.Size([512, 11008])

      • qzeros: torch.Size([32, 1376])

      • scales: torch.Size([32, 11008])

      • g_idx: torch.Size([4096])

      • bias: torch.Size([11008])

    • up_proj:

      • qweight: torch.Size([512, 11008])

      • qzeros: torch.Size([32, 1376])

      • scales: torch.Size([32, 11008])

      • g_idx: torch.Size([4096])

      • bias: torch.Size([11008])

  • post_attention_layernorm.weight: torch.Size([4096])

Final Layer Normalization and Output

  • model.norm.weight: torch.Size([4096])
  • lm_head.weight: torch.Size([151851, 4096])
相关标签
About Me
XD
Goals determine what you are going to be.
Category
标签云
Breakpoint Firewall Linux YOLO WebCrawler Git NLTK Conda RAR Jetson tar 搞笑 scipy Clash Tensor Permission Math tqdm Input Gemma Distillation Crawler Card Knowledge Bitcoin CC uWSGI 签证 Miniforge git-lfs PIP Statistics DeepStream Template Ptyhon Logo mmap NLP Password Python Hilton OpenCV FP32 Streamlit BF16 XGBoost Magnet FP64 TensorFlow Disk Hungarian 多线程 ChatGPT VPN CAM 继承 Github 域名 Jupyter Food 证件照 VGG-16 报税 NameSilo Numpy BeautifulSoup LaTeX Mixtral v2ray Bin 阿里云 Zip Pillow DeepSeek EXCEL Pandas VSCode FlashAttention AI Safetensors SAM CLAP Plate InvalidArgumentError 版权 GoogLeNet 飞书 Bert 腾讯云 OCR Attention Tracking Proxy GPTQ Nginx Color uwsgi Excel JSON FP8 QWEN ONNX Translation Video CTC ModelScope 财报 Random SPIE 算法题 CEIR Augmentation ResNet-50 PDF Anaconda TensorRT Django Freesound GGML CUDA printf API Markdown TSV Domain UNIX Data FP16 COCO Qwen Diagram transformers Use Claude Docker Paper Llama Transformers Ubuntu MD5 LeetCode Google Hotel CSV logger LLM Animate Bipartite Image2Text PDB Heatmap Plotly PyCharm Datetime Interview UI Shortcut OpenAI GIT SQLite TTS Algorithm Quantization llama.cpp Pickle RGB Qwen2 SVR Web Review CV Sklearn v0.dev WAN BTC IndexTTS2 Base64 关于博主 Qwen2.5 多进程 GPT4 torchinfo HaggingFace Cloudreve HuggingFace LLAMA git Pytorch Vmess Michelin Quantize Windows Tiktoken Paddle hf LoRA Vim diffusers Website C++ SQL Baidu PyTorch Land 净利润 XML 公式 FastAPI Dataset 音频
站点统计

本站现有博文311篇,共被浏览741912

本站已经建立2381天!

热门文章
文章归档
回到顶部