{"id":4867,"date":"2024-12-05T17:52:05","date_gmt":"2024-12-05T09:52:05","guid":{"rendered":"https:\/\/www.aqwu.net\/wp\/?p=4867"},"modified":"2024-12-05T18:05:41","modified_gmt":"2024-12-05T10:05:41","slug":"%e4%bd%bf%e7%94%a8gpu%e5%86%85%e5%ad%98%e5%92%8ccpu%e5%86%85%e5%ad%98%e8%a3%85%e8%bd%bd%e5%a4%a7%e6%a8%a1%e5%9e%8b","status":"publish","type":"post","link":"https:\/\/www.aqwu.net\/wp\/?p=4867","title":{"rendered":"\u4f7f\u7528GPU\u5185\u5b58\u548cCPU\u5185\u5b58\u88c5\u8f7d\u5927\u6a21\u578b"},"content":{"rendered":"\n<p>\u8fd9\u91cc\u6d4b\u8bd5\u7684\u662f\u5f00\u6e90\u6a21\u578b\u662f <a href=\"https:\/\/huggingface.co\/meta-llama\/Llama-3.1-405B-Instruct\">meta-llama\/Llama-3.1-405B-Instruct<\/a>\uff0c\u5c5e\u4e8e\u5f00\u6e90\u6a21\u578b\u91cc\u9762\u6700\u5927\u7684\u3002<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">1. \u9ed8\u8ba4\u65b9\u6cd5<\/h2>\n\n\n\n<p>\u4ee3\u7801\u5982\u4e0b\uff0c\u8fd9\u4e2a\u65b9\u6cd5\u53ef\u4ee5\u52a0\u8f7d\u8d77\u6765\u6a21\u578b\uff0c\u4f46\u662f\u5728\u63a8\u7406\u7684\u65f6\u5019\u4f1a\u975e\u5e38\u6162\u3002<\/p>\n\n\n\n<div class=\"wp-block-urvanov-syntax-highlighter-code-block\"><pre class=\"lang:python decode:true \" >import torch\nfrom transformers import AutoModelForCausalLM\n\nMODEL_ID = \"meta-llama\/Llama-3.1-405B-Instruct\"\n\nprint(f\"Load Model {MODEL_ID} ... \")\nmodel = AutoModelForCausalLM.from_pretrained(\n    MODEL_ID, \n    device_map=\"auto\", \n    trust_remote_code=True,\n    torch_dtype=torch.bfloat16\n)\n<\/pre><\/div>\n\n\n\n<h3 class=\"wp-block-heading\">2. \u4f7f\u7528 BitsAndBytesConfig<\/h3>\n\n\n\n<p>\u8fd9\u4e2a\u65b9\u6cd5\u53ef\u4ee5\u628a\u66f4\u591a\u7684\u6743\u91cd\u52a0\u8f7d\u5230GPU\u91cc\u9762\uff0c\u8fd9\u6837\u901f\u5ea6\u4f1a\u5feb\u5f88\u591a\uff0c\u4e0b\u9762\u662f8\u4e2a24G\u7684GPU\u5185\u5b58\u7684\u5185\u5b58\u6620\u5c04\u4ee3\u7801\u3002<\/p>\n\n\n\n<div class=\"wp-block-urvanov-syntax-highlighter-code-block\"><pre class=\"lang:python decode:true \" >import torch\nfrom transformers import AutoModelForCausalLM, BitsAndBytesConfig\n\n# \u68c0\u6d4b\u53ef\u7528\u7684GPU\u6570\u91cf\nNUM_GPUS = torch.cuda.device_count()\nprint(f\"NUM_GPUS: {NUM_GPUS}\")\n\n# \u5b9a\u4e49Transformer\u6a21\u578b\u7684\u5c42\u6570\nNUM_TRANS_LAYERS = 126\nMODEL_ID = \"meta-llama\/Llama-3.1-405B-Instruct\"\n\ndef create_device_map():\n    device_map = {\n        'model.embed_tokens': 0,\n        'model.norm': 0,\n        'model.rotary_emb': 0,\n        'lm_head': 0\n    }\n    # \u6839\u636e GPU \u6570\u91cf\u914d\u7f6e\u8bbe\u5907\u6620\u5c04\n    if NUM_GPUS &gt; 0:\n        for start, end, gpu_id in [(0, 6, 0), (6, 18, 1), (18, 30, 2), (30, 42, 3), (42, 54, 4), (54, 66, 5), (66, 78, 6), (78, 90, 7)]:\n            for i in range(start, end):\n                device_map[f'model.layers.{i}'] = gpu_id\n    for i in range(90, NUM_TRANS_LAYERS):\n        device_map[f'model.layers.{i}'] = \"cpu\"  # \u4f7f\u7528 CPU \u5904\u7406\u5269\u4f59\u5c42\n    return device_map\n\ndevice_map = create_device_map() if NUM_GPUS &gt; 0 else None\n\n# \u6a21\u578b\u52a0\u8f7d\nprint(f\"Loading Model {MODEL_ID} ...\")\nmodel = AutoModelForCausalLM.from_pretrained(\n    MODEL_ID,\n    device_map=device_map,\n    trust_remote_code=True,\n    torch_dtype=torch.bfloat16,\n    quantization_config=BitsAndBytesConfig(\n        load_in_4bit=True, \n        bnb_4bit_compute_dtype=torch.float16,\n        llm_int8_enable_fp32_cpu_offload=True  # \u542f\u7528 CPU \u8f85\u52a9\n    )\n)\n<\/pre><\/div>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>\u8fd9\u91cc\u6d4b\u8bd5\u7684\u662f\u5f00\u6e90\u6a21\u578b\u662f meta-llama\/Llama-3.1-405B-Instruct\uff0c\u5c5e\u4e8e\u5f00\u6e90\u6a21\u578b\u91cc\u9762 [&hellip;]<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"default","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"set","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"_jetpack_memberships_contains_paid_content":false,"footnotes":""},"categories":[444,445,443,442],"tags":[571],"class_list":["post-4867","post","type-post","status-publish","format-standard","hentry","category-ai","category-ainews","category-llm","category-llms","tag-met-llama"],"views":2784,"jetpack_sharing_enabled":true,"jetpack_featured_media_url":"","_links":{"self":[{"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=\/wp\/v2\/posts\/4867","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=4867"}],"version-history":[{"count":4,"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=\/wp\/v2\/posts\/4867\/revisions"}],"predecessor-version":[{"id":4871,"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=\/wp\/v2\/posts\/4867\/revisions\/4871"}],"wp:attachment":[{"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=4867"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=4867"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.aqwu.net\/wp\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=4867"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}