

{"id":446,"date":"2025-11-25T09:17:26","date_gmt":"2025-11-25T01:17:26","guid":{"rendered":"https:\/\/high-flyer.in.suopu.cc\/?p=446"},"modified":"2025-11-25T09:17:26","modified_gmt":"2025-11-25T01:17:26","slug":"haiscale-%e5%b9%bb%e6%96%b9%e8%90%a4%e7%81%ab%e9%ab%98%e6%80%a7%e8%83%bd%e5%b9%b6%e8%a1%8c%e8%ae%ad%e7%bb%83%e5%b7%a5%e5%85%b7%e5%ba%93","status":"publish","type":"post","link":"https:\/\/high-flyer.in.suopu.cc\/en\/blog\/446\/","title":{"rendered":"haiscale | \u5e7b\u65b9\u8424\u706b\u9ad8\u6027\u80fd\u5e76\u884c\u8bad\u7ec3\u5de5\u5177\u5e93"},"content":{"rendered":"<p>haiscale (Highflyer AI Scale) \u662f\u4e00\u4e2a\u8f7b\u91cf\u7ea7\u7684\u9ad8\u6027\u80fd\u5e76\u884c\u8bad\u7ec3\u5de5\u5177\u5e93\uff0c\u5176\u6574\u5408\u4e86\u5e7b\u65b9 AI \u591a\u5e74\u7684\u5e76\u884c\u8bad\u7ec3\u7814\u53d1\u4f18\u5316\u7ecf\u9a8c\uff0c\u80fd\u591f\u5e2e\u52a9 PyTorch \u7528\u6237\u66f4\u52a0\u9ad8\u6548\u3001\u4fbf\u6377\u5730\u5728\u5927\u89c4\u6a21\u96c6\u7fa4\u4e0a\u8bad\u7ec3\u6a21\u578b\u3002<\/p>\n<p>haiscale \u4e2d\u5305\u542b\u4e86\u4ee5\u4e0b\u51e0\u79cd\u5de5\u5177\uff1a<\/p>\n<ol>\n<li><code class=\"language-text\">haiscale.ddp<\/code>: \u5206\u5e03\u5f0f\u6570\u636e\u5e76\u884c\u5de5\u5177\uff0c\u4ee5\u5e7b\u65b9 AI \u81ea\u7814\u7684\u00a0<a href=\"https:\/\/www.high-flyer.cn\/blog\/hf-reduce\/\">hfreduce<\/a>\u00a0\u901a\u4fe1\u4e3a\u540e\u7aef\uff0c\u76f8\u6bd4\u4e8e NCCL \u80fd\u591f\u83b7\u5f97\u66f4\u597d\u7684\u591a\u5361\u62d3\u5c55\u6027\u80fd\uff1b<\/li>\n<li><code class=\"language-text\">haiscale.fsdp<\/code>: \u6781\u81f4\u4f18\u5316\u00a0<code class=\"language-text\">Fully Sharded Data Parallel (FSDP)<\/code>\u00a0\u7b97\u6cd5\u7684\u5b9e\u73b0\uff0c\u76f8\u6bd4\u4e8e PyTorch FSDP \u901f\u5ea6\u66f4\u5feb\u3001\u5360\u7528\u663e\u5b58\u66f4\u5c11\uff1b<\/li>\n<li><code class=\"language-text\">haiscale.pipeline<\/code>: \u5206\u5e03\u5f0f\u6d41\u6c34\u7ebf\u5e76\u884c\uff08\u6216\u79f0\u6a21\u578b\u5e76\u884c\uff09\u5de5\u5177\u5305\uff0c\u5305\u542b GPipe, PipeDream \u7b49\u7b97\u6cd5\uff0c\u652f\u6301\u591a\u673a\u591a\u5361\u8bad\u7ec3\uff1b<\/li>\n<li><code class=\"language-text\">haiscale.cpu_offload<\/code>: \u795e\u7ecf\u7f51\u7edc\u6a21\u578b Offload \u5de5\u5177\uff0c\u8282\u7701\u8bad\u7ec3\u5360\u7528\u7684\u663e\u5b58\u3002<\/li>\n<\/ol>\n<p>\u4e0b\u56fe\u5c55\u793a\u4e86 haiscale \u4e09\u79cd\u5e76\u884c\u65b9\u5f0f\u7684\u6027\u80fd\uff0c\u5176\u76f8\u6bd4 PyTorch \u5b98\u65b9\u81ea\u5e26\u5de5\u5177\u90fd\u6709\u663e\u8457\u7684\u6027\u80fd\u63d0\u5347\uff1a<\/p>\n<p><span class=\"gatsby-resp-image-wrapper\"><a class=\"gatsby-resp-image-link\" href=\"https:\/\/hfai-static.high-flyer.cn\/static\/824a9800a50b56571ffc37e53ff3d98d\/2cefc\/bench.png\" target=\"_blank\" rel=\"noopener\"><img fetchpriority=\"high\" decoding=\"async\" class=\"alignnone size-full wp-image-448\" title=\"bench\" src=\"http:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/bench.png\" alt=\"bench\" width=\"650\" height=\"186\" srcset=\"https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/bench.png 650w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/bench-300x86.png 300w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/bench-150x43.png 150w\" sizes=\"(max-width: 650px) 100vw, 650px\" \/><\/a><\/span><\/p>\n<p>\u7528\u4e8e\u6d4b\u8bd5\u7684\u6a21\u578b\u662f\u00a0<strong>GPT-2 Medium<\/strong>\uff0c\u76f8\u5173\u4ee3\u7801\u5df2\u5f00\u6e90\u81f3 hfai \u6a21\u578b\u4ed3\u5e93\u3002\u4e0b\u9762\u5c06\u4e3a\u5927\u5bb6\u7b80\u8981\u4ecb\u7ecd\u3002<\/p>\n<p><strong>API\u6587\u6863<\/strong>\uff1a<a href=\"https:\/\/doc.hfai.high-flyer.cn\/api\/haiscale_ddp.html\">https:\/\/doc.hfai.high-flyer.cn\/api\/haiscale_ddp.html<\/a><\/p>\n<p><strong>\u793a\u4f8b\u6a21\u578b<\/strong>\uff1a<a href=\"https:\/\/github.com\/HFAiLab\/hfai-models\/tree\/main\/gpt\">https:\/\/github.com\/HFAiLab\/hfai-models\/tree\/main\/gpt<\/a><\/p>\n<h2 id=\"\u5b89\u88c5\">\u5b89\u88c5<\/h2>\n<p>haiscale \u63d0\u4f9b Python \u63a5\u53e3\uff0c\u901a\u8fc7\u5982\u4e0b\u65b9\u5f0f\u5b89\u88c5\uff1a<\/p>\n<ol>\n<li>\u5982\u679c\u8981\u4f7f\u7528 haiscale DDP\uff0c\u9996\u5148\u9700\u8981\u5148\u5b89\u88c5 hfreduce \uff08\u5982\u679c\u4e0d\u9700\u8981\u4f7f\u7528 DDP \u53ef\u8df3\u8fc7\u8fd9\u6b65\uff09\uff1a\n<div class=\"gatsby-highlight\" data-language=\"text\">\n<pre class=\"language-text\"><code class=\"language-text\">sudo apt install libnuma-dev\r\nsudo apt install libibverbs-dev\r\npip install hfreduce --extra-index-url https:\/\/pypi.hfai.high-flyer.cn\/simple --trusted-host pypi.hfai.high-flyer.cn<\/code><\/pre>\n<\/div>\n<\/li>\n<li>\u5b89\u88c5 haiscale\uff1a\n<div class=\"gatsby-highlight\" data-language=\"text\">\n<pre class=\"language-text\"><code class=\"language-text\">pip install haiscale --extra-index-url https:\/\/pypi.hfai.high-flyer.cn\/simple --trusted-host pypi.hfai.high-flyer.cn<\/code><\/pre>\n<\/div>\n<\/li>\n<\/ol>\n<h2 id=\"haiscaleddp\">haiscale.ddp<\/h2>\n<p><code class=\"language-text\">haiscale.ddp.DistributedDataParallel<\/code>\u00a0(haiscale DDP) \u662f\u4e00\u4e2a\u5206\u5e03\u5f0f\u6570\u636e\u5e76\u884c\u8bad\u7ec3\u5de5\u5177\uff0c\u4f7f\u7528 hfreduce \u4f5c\u4e3a\u901a\u8baf\u540e\u7aef\uff0c\u53cd\u5411\u4f20\u64ad\u7684\u540c\u65f6\u4f1a\u5f02\u6b65\u5730\u5bf9\u8ba1\u7b97\u597d\u7684\u68af\u5ea6\u505a allreduce\u3002<\/p>\n<p><span class=\"gatsby-resp-image-wrapper\"><a class=\"gatsby-resp-image-link\" href=\"https:\/\/hfai-static.high-flyer.cn\/static\/66f02e5b23fbe59d0e77df6fe423a535\/bca35\/DDP.png\" target=\"_blank\" rel=\"noopener\"><img decoding=\"async\" class=\"alignnone size-full wp-image-450\" title=\"ddp\" src=\"http:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/ddp.png\" alt=\"ddp\" width=\"650\" height=\"379\" srcset=\"https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/ddp.png 650w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/ddp-300x175.png 300w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/ddp-150x87.png 150w\" sizes=\"(max-width: 650px) 100vw, 650px\" \/><\/a><\/span><\/p>\n<p>haiscale DDP \u7684\u4f7f\u7528\u65b9\u5f0f\u548c pytorch DDP \u51e0\u4e4e\u76f8\u540c\uff0c\u4ee5\u4e0b\u662f\u4f7f\u7528\u793a\u4f8b\uff1a<\/p>\n<div class=\"gatsby-highlight\" data-language=\"python\">\n<pre class=\"language-python\"><code class=\"language-python\"><span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>ddp <span class=\"token keyword\">import<\/span> DistributedDataParallel\r\n\r\nmodel <span class=\"token operator\">=<\/span> MyModel<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\nmodel <span class=\"token operator\">=<\/span> DistributedDataParallel<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">)<\/span>\r\noptimizer <span class=\"token operator\">=<\/span> torch<span class=\"token punctuation\">.<\/span>optim<span class=\"token punctuation\">.<\/span>Adam<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">.<\/span>parameters<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> lr<span class=\"token operator\">=<\/span><span class=\"token number\">0.01<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token comment\"># training ...<\/span>\r\n<span class=\"token keyword\">for<\/span> step<span class=\"token punctuation\">,<\/span> <span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">,<\/span> y<span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">in<\/span> <span class=\"token builtin\">enumerate<\/span><span class=\"token punctuation\">(<\/span>dataloader<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    optimizer<span class=\"token punctuation\">.<\/span>zero_grad<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n    output <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">)<\/span>\r\n    loss_fn<span class=\"token punctuation\">(<\/span>y<span class=\"token punctuation\">,<\/span> output<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>backward<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n    optimizer<span class=\"token punctuation\">.<\/span>step<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token comment\"># stop hfreduce<\/span>\r\nmodel<span class=\"token punctuation\">.<\/span>reducer<span class=\"token punctuation\">.<\/span>stop<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><\/code><\/pre>\n<\/div>\n<p>\u5982\u679c\u9700\u8981\u505a\u68af\u5ea6\u7d2f\u52a0\uff0c\u53ef\u4ee5\u4f7f\u7528\u00a0<code class=\"language-text\">model.no_sync()<\/code>\u00a0\u6765\u51cf\u5c11\u901a\u8baf\u7684\u5f00\u9500\u3002\u6ce8\u610f\u53ea\u9700\u8981\u6700\u540e\u4e00\u6b21\u53cd\u5411\u4f20\u64ad\u65f6\u505a allreduce\uff1a<\/p>\n<div class=\"gatsby-highlight\" data-language=\"python\">\n<pre class=\"language-python\"><code class=\"language-python\"><span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>ddp <span class=\"token keyword\">import<\/span> DistributedDataParallel\r\n\r\nddp <span class=\"token operator\">=<\/span> DistributedDataParallel<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">,<\/span> <span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">)<\/span>\r\n<span class=\"token keyword\">with<\/span> ddp<span class=\"token punctuation\">.<\/span>no_sync<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    <span class=\"token keyword\">for<\/span> <span class=\"token builtin\">input<\/span> <span class=\"token keyword\">in<\/span> inputs<span class=\"token punctuation\">:<\/span>\r\n        ddp<span class=\"token punctuation\">(<\/span><span class=\"token builtin\">input<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>backward<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>  <span class=\"token comment\"># no synchronization, accumulate grads<\/span>\r\nddp<span class=\"token punctuation\">(<\/span>another_input<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>backward<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>  <span class=\"token comment\"># synchronize grads<\/span><\/code><\/pre>\n<\/div>\n<h2 id=\"haiscalefsdp\">haiscale.fsdp<\/h2>\n<p>Fully Sharded Data Parallel (FSDP) \u662f META \u5728 ZERO-3 \u7684\u57fa\u7840\u4e0a\u63d0\u51fa\u7684\u5206\u5e03\u5f0f\u6570\u636e\u5e76\u884c\u5de5\u5177\uff0c\u5b83\u628a\u6a21\u578b\u7684\u53c2\u6570\u8fdb\u884c\u5207\u5206\u5e76\u5206\u6563\u5230\u4e0d\u540c\u7684 GPU \u4e0a\uff0c\u6bcf\u5757 GPU \u4e0a\u53ea\u6709\u00a0<code class=\"language-text\">1\/ngpus<\/code>\u00a0\u7684\u53c2\u6570\u3002\u5728\u505a\u524d\u5411\u548c\u53cd\u5411\u4f20\u64ad\u65f6\uff0cFSDP \u4f1a\u5148\u505a allgather \u83b7\u5f97\u5b8c\u6574\u7684\u53c2\u6570\uff0c\u7136\u540e\u5728\u524d\u5411\u548c\u53cd\u5411\u4f20\u64ad\u7ed3\u675f\u540e\u91ca\u653e\u6389\uff0c\u53ea\u4fdd\u7559\u00a0<code class=\"language-text\">1\/ngpus<\/code>\u00a0\u7684\u53c2\u6570\u548c\u68af\u5ea6\u3002FSDP \u901a\u8fc7<strong>\u53c2\u6570\u5206\u7247<\/strong>\u7684\u65b9\u5f0f\uff0c\u80fd\u591f\u51cf\u5c11\u6a21\u578b\u53c2\u6570\u3001\u68af\u5ea6\u3001\u4f18\u5316\u5668\u72b6\u6001\u7684\u663e\u5b58\u5360\u7528\uff0c\u5e2e\u52a9\u6211\u4eec\u8bad\u7ec3\u66f4\u5927\u89c4\u6a21\u7684\u6a21\u578b\u3002<\/p>\n<p><span class=\"gatsby-resp-image-wrapper\"><a class=\"gatsby-resp-image-link\" href=\"https:\/\/hfai-static.high-flyer.cn\/static\/a3e18d9cef6129e91bc72e0de3edde9e\/83e77\/FSDP.png\" target=\"_blank\" rel=\"noopener\"><img decoding=\"async\" class=\"alignnone size-full wp-image-449\" title=\"fsdp\" src=\"http:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/fsdp.png\" alt=\"fsdp\" width=\"650\" height=\"304\" srcset=\"https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/fsdp.png 650w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/fsdp-300x140.png 300w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/fsdp-150x70.png 150w\" sizes=\"(max-width: 650px) 100vw, 650px\" \/><\/a><\/span><\/p>\n<p><code class=\"language-text\">haiscale.fsdp.FullyShardedDataParallel<\/code>\u00a0\u7684\u4f7f\u7528\u65b9\u6cd5\u548c DDP \u7c7b\u4f3c\uff0c\u4f46<strong>\u4f18\u5316\u5668\u5fc5\u987b\u5728 FSDP \u4e4b\u540e\u521b\u5efa<\/strong>\uff0c\u5e76\u4e14\u4fdd\u5b58\u6a21\u578b\u53c2\u6570\u7684\u65f6\u5019\u9700\u8981\u5148\u8c03\u7528\u00a0<code class=\"language-text\">summon_full_params<\/code>\u3002\u4ee5\u4e0b\u662f\u4f7f\u7528\u793a\u4f8b\uff1a<\/p>\n<div class=\"gatsby-highlight\" data-language=\"python\">\n<pre class=\"language-python\"><code class=\"language-python\"><span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>fsdp <span class=\"token keyword\">import<\/span> FullyShardedDataParallel\r\n\r\nmodel <span class=\"token operator\">=<\/span> MyModel<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\nmodel <span class=\"token operator\">=<\/span> FullyShardedDataParallel<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">)<\/span>\r\noptimizer <span class=\"token operator\">=<\/span> torch<span class=\"token punctuation\">.<\/span>optim<span class=\"token punctuation\">.<\/span>AdamW<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">.<\/span>parameters<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> lr<span class=\"token operator\">=<\/span><span class=\"token number\">0.01<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token comment\"># training ...<\/span>\r\n<span class=\"token keyword\">for<\/span> step<span class=\"token punctuation\">,<\/span> <span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">,<\/span> y<span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">in<\/span> <span class=\"token builtin\">enumerate<\/span><span class=\"token punctuation\">(<\/span>dataloader<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    optimizer<span class=\"token punctuation\">.<\/span>zero_grad<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n    output <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">)<\/span>\r\n    loss_fn<span class=\"token punctuation\">(<\/span>y<span class=\"token punctuation\">,<\/span> output<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>backward<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n    optimizer<span class=\"token punctuation\">.<\/span>step<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token comment\"># save checkpoint<\/span>\r\n<span class=\"token keyword\">with<\/span> model<span class=\"token punctuation\">.<\/span>summon_full_params<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    <span class=\"token keyword\">if<\/span> rank <span class=\"token operator\">==<\/span> <span class=\"token number\">0<\/span><span class=\"token punctuation\">:<\/span>\r\n        state <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">.<\/span>state_dict<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n        torch<span class=\"token punctuation\">.<\/span>save<span class=\"token punctuation\">(<\/span>state<span class=\"token punctuation\">,<\/span> <span class=\"token string\">'model.pt'<\/span><span class=\"token punctuation\">)<\/span><\/code><\/pre>\n<\/div>\n<p>haiscale FSDP \u8fd8\u652f\u6301\u4f20\u5165\u00a0<code class=\"language-text\">auto_wrap_policy<\/code>\u00a0\u53c2\u6570\uff0c\u5177\u4f53\u4f5c\u7528\u53ef\u4ee5\u53c2\u8003\u00a0<a href=\"https:\/\/pytorch.org\/docs\/1.12\/fsdp.html\">PyTorch FSDP \u7684\u6587\u6863<\/a>\u4ee5\u53ca\u6211\u4eec\u63d0\u4f9b\u7684\u00a0<a href=\"https:\/\/github.com\/HFAiLab\/hfai-models\/tree\/main\/gpt\">GPT-2 \u793a\u4f8b<\/a>\u3002<\/p>\n<h2 id=\"haiscalepipeline\">haiscale.pipeline<\/h2>\n<p><span class=\"gatsby-resp-image-wrapper\"><a class=\"gatsby-resp-image-link\" href=\"https:\/\/hfai-static.high-flyer.cn\/static\/8367cb02031b4ce2a576d98cd78a00aa\/f7616\/pipeline.png\" target=\"_blank\" rel=\"noopener\"><img loading=\"lazy\" decoding=\"async\" class=\"alignnone size-full wp-image-451\" title=\"pipeline\" src=\"http:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/pipeline.png\" alt=\"pipeline\" width=\"650\" height=\"394\" srcset=\"https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/pipeline.png 650w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/pipeline-300x182.png 300w, https:\/\/high-flyer.in.suopu.cc\/wp-content\/uploads\/2025\/11\/pipeline-150x91.png 150w\" sizes=\"(max-width: 650px) 100vw, 650px\" \/><\/a><\/span><\/p>\n<p>\u5982\u4e0a\u56fe\u6240\u793a\uff0c<code class=\"language-text\">haiscale.pipeline<\/code>\u00a0\u5de5\u5177\u5305\u4e2d\u63d0\u4f9b\u4e86\u4e09\u79cd\u6d41\u6c34\u7ebf\u5e76\u884c\u7684\u7b97\u6cd5\uff1a<\/p>\n<ol>\n<li><code class=\"language-text\">GPipe<\/code>: \u628a\u6a21\u578b\u5207\u5206\u6210\u00a0<code class=\"language-text\">ngpus<\/code>\u00a0\u4efd\uff0c\u6240\u6709 microbatch \u7684\u524d\u5411\u4f20\u64ad\u7ed3\u675f\u4e4b\u540e\u518d\u505a\u53cd\u5411\u4f20\u64ad\uff1b<\/li>\n<li><code class=\"language-text\">PipeDream<\/code>: \u628a\u6a21\u578b\u5207\u5206\u6210\u00a0<code class=\"language-text\">ngpus<\/code>\u00a0\u4efd\uff0c\u524d\u5411\u548c\u53cd\u5411\u4f20\u64ad\u4ea4\u66ff\u6267\u884c\uff08non-interleaved 1F1B\uff09\uff1b<\/li>\n<li><code class=\"language-text\">Interleaved1F1B<\/code>: \u628a\u6a21\u578b\u5207\u5206\u6210\u00a0<code class=\"language-text\">ngpus * num_model_chunks<\/code>\u00a0\u4efd\uff0c\u524d\u5411\u548c\u53cd\u5411\u4f20\u64ad\u4ea4\u66ff\u6267\u884c\u3002<\/li>\n<\/ol>\n<p>\u5bf9\u4e8e\u4e2d\u7b49\u89c4\u6a21\u7684\u6a21\u578b\uff08\u6bd4\u5982 GPT-2 Medium\uff09\uff0c\u6211\u4eec\u63a8\u8350\u4f18\u5148\u4f7f\u7528 PipeDream\uff0c\u76f8\u6bd4\u4e8e GPipe \u548c Interleaved1F1B \u5360\u7528\u663e\u5b58\u66f4\u5c11\uff0c\u901f\u5ea6\u66f4\u5feb\u3002 \u5bf9\u4e8e\u8d85\u5927\u89c4\u6a21\u7684\u6a21\u578b\uff08\u6bd4\u5982 GPT-3\uff09\uff0c\u6211\u4eec\u53ef\u4ee5\u66f4\u52a0\u5747\u5300\u3001\u7ec6\u7c92\u5ea6\u7684\u5207\u5206\u5b83\uff0c\u8fd9\u65f6\u5019\u63a8\u8350\u4f7f\u7528 Interleaved1F1B\u3002<\/p>\n<p><code class=\"language-text\">haiscale.pipeline<\/code>\u00a0\u63d0\u4f9b\u4e86\u4e00\u4e2a\u7edf\u4e00\u7684\u00a0<code class=\"language-text\">forward_backward<\/code>\u00a0\u63a5\u53e3\uff0c\u6211\u4eec\u9700\u8981\u4f20\u5165\u635f\u5931\u51fd\u6570\u00a0<code class=\"language-text\">criterion<\/code>\u00a0\u548c\u6807\u7b7e\u6570\u636e\u00a0<code class=\"language-text\">labels<\/code>\uff0c\u635f\u5931\u51fd\u6570\u4f1a\u901a\u8fc7\u00a0<code class=\"language-text\">loss = criterion(*outputs, *labels)<\/code>\u00a0\u7684\u65b9\u5f0f\u8c03\u7528\uff0c<code class=\"language-text\">forward_backward<\/code>\u00a0\u63a5\u53e3\u4f1a\u8fd4\u56de\u4e00\u4e2a\u5143\u7ec4\u00a0<code class=\"language-text\">(losses, outputs)<\/code>\uff0c\u5176\u4e2d<code class=\"language-text\">losses<\/code>\u00a0\u4ee3\u8868\u6bcf\u4e2a microbatch \u7684 loss \u503c\uff0c<code class=\"language-text\">outputs<\/code>\u00a0\u4ee3\u8868\u6a21\u578b\u7684\u8f93\u51fa\u3002\u53ea\u6709\u6700\u540e\u4e00\u4e2a rank \u7684\u8fdb\u7a0b\u80fd\u591f\u83b7\u5f97 loss \u548c\u8f93\u51fa\uff0c\u5176\u4ed6\u8fdb\u7a0b\u5f97\u5230\u7684\u662f\u00a0<code class=\"language-text\">(None, None)<\/code>\u3002<\/p>\n<p>\u4e0b\u9762\u901a\u8fc7\u793a\u4f8b\u5c55\u793a GPipe \u548c PipeDream \u7684\u7528\u6cd5\uff1a<\/p>\n<div class=\"gatsby-highlight\" data-language=\"python\">\n<pre class=\"language-python\"><code class=\"language-python\"><span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>pipeline <span class=\"token keyword\">import<\/span> GPipe<span class=\"token punctuation\">,<\/span> PipeDream<span class=\"token punctuation\">,<\/span> partition\r\n\r\ndist<span class=\"token punctuation\">.<\/span>init_process_group<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">)<\/span>\r\ntorch<span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">.<\/span>set_device<span class=\"token punctuation\">(<\/span>local_rank<span class=\"token punctuation\">)<\/span>\r\nrank<span class=\"token punctuation\">,<\/span> world_size <span class=\"token operator\">=<\/span> dist<span class=\"token punctuation\">.<\/span>get_rank<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> dist<span class=\"token punctuation\">.<\/span>get_world_size<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\ntorch<span class=\"token punctuation\">.<\/span>manual_seed<span class=\"token punctuation\">(<\/span><span class=\"token number\">12345<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token keyword\">def<\/span> <span class=\"token function\">loss_fn<\/span><span class=\"token punctuation\">(<\/span>out<span class=\"token punctuation\">,<\/span> y<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    <span class=\"token keyword\">return<\/span> <span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">(<\/span>out <span class=\"token operator\">-<\/span> y<span class=\"token punctuation\">)<\/span><span class=\"token operator\">**<\/span><span class=\"token number\">2<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span><span class=\"token builtin\">sum<\/span><span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\nmodel <span class=\"token operator\">=<\/span> nn<span class=\"token punctuation\">.<\/span>Sequential<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">)<\/span>\r\nmodel <span class=\"token operator\">=<\/span> partition<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">,<\/span> rank<span class=\"token punctuation\">,<\/span> world_size<span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token comment\"># chunks: number of microbatches<\/span>\r\nmodel <span class=\"token operator\">=<\/span> PipeDream<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> chunks<span class=\"token operator\">=<\/span><span class=\"token number\">32<\/span><span class=\"token punctuation\">)<\/span>\r\n<span class=\"token comment\"># or model = GPipe(model.cuda(), chunks=32)<\/span>\r\n\r\n<span class=\"token keyword\">for<\/span> x<span class=\"token punctuation\">,<\/span> y <span class=\"token keyword\">in<\/span> dataloader<span class=\"token punctuation\">:<\/span>\r\n    losses<span class=\"token punctuation\">,<\/span> outputs <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">.<\/span>forward_backward<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">,<\/span> criterion<span class=\"token operator\">=<\/span>loss_fn<span class=\"token punctuation\">,<\/span> labels<span class=\"token operator\">=<\/span><span class=\"token punctuation\">(<\/span>y<span class=\"token punctuation\">,<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> return_outputs<span class=\"token operator\">=<\/span><span class=\"token boolean\">True<\/span><span class=\"token punctuation\">)<\/span>\r\n    <span class=\"token keyword\">if<\/span> rank <span class=\"token operator\">==<\/span> world_size <span class=\"token operator\">-<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span>\r\n        loss <span class=\"token operator\">=<\/span> losses<span class=\"token punctuation\">.<\/span><span class=\"token builtin\">sum<\/span><span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>item<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>  <span class=\"token comment\"># losses: torch.Size([32])<\/span>\r\n\r\n<span class=\"token comment\"># eval<\/span>\r\n<span class=\"token keyword\">with<\/span> torch<span class=\"token punctuation\">.<\/span>no_grad<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    out <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">)<\/span>\r\n    <span class=\"token keyword\">if<\/span> rank <span class=\"token operator\">==<\/span> world_size <span class=\"token operator\">-<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span>\r\n        <span class=\"token comment\"># calculate metrics ...<\/span><\/code><\/pre>\n<\/div>\n<p>\u4ee5\u4e0b\u662f Interleaved1F1B \u7684\u4f7f\u7528\u793a\u4f8b\uff1a<\/p>\n<div class=\"gatsby-highlight\" data-language=\"python\">\n<pre class=\"language-python\"><code class=\"language-python\"><span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>pipeline <span class=\"token keyword\">import<\/span> Interleaved1F1B<span class=\"token punctuation\">,<\/span> partition\r\n\r\ndist<span class=\"token punctuation\">.<\/span>init_process_group<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">)<\/span>\r\ntorch<span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">.<\/span>set_device<span class=\"token punctuation\">(<\/span>local_rank<span class=\"token punctuation\">)<\/span>\r\nrank<span class=\"token punctuation\">,<\/span> world_size <span class=\"token operator\">=<\/span> dist<span class=\"token punctuation\">.<\/span>get_rank<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> dist<span class=\"token punctuation\">.<\/span>get_world_size<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\ntorch<span class=\"token punctuation\">.<\/span>manual_seed<span class=\"token punctuation\">(<\/span><span class=\"token number\">12345<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token keyword\">def<\/span> <span class=\"token function\">loss_fn<\/span><span class=\"token punctuation\">(<\/span>out<span class=\"token punctuation\">,<\/span> y<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    <span class=\"token keyword\">return<\/span> <span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">(<\/span>out <span class=\"token operator\">-<\/span> y<span class=\"token punctuation\">)<\/span><span class=\"token operator\">**<\/span><span class=\"token number\">2<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span><span class=\"token builtin\">sum<\/span><span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\nmodel <span class=\"token operator\">=<\/span> nn<span class=\"token punctuation\">.<\/span>Sequential<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">)<\/span>\r\nmodules <span class=\"token operator\">=<\/span> partition<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">,<\/span> rank<span class=\"token punctuation\">,<\/span> world_size<span class=\"token punctuation\">,<\/span> num_model_chunks<span class=\"token operator\">=<\/span><span class=\"token number\">2<\/span><span class=\"token punctuation\">)<\/span>  <span class=\"token comment\"># len(modules) = 2<\/span>\r\nmodules <span class=\"token operator\">=<\/span> <span class=\"token punctuation\">[<\/span>m<span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span> <span class=\"token keyword\">for<\/span> m <span class=\"token keyword\">in<\/span> modules<span class=\"token punctuation\">]<\/span>\r\nmodel <span class=\"token operator\">=<\/span> Interleaved1F1B<span class=\"token punctuation\">(<\/span>modules<span class=\"token punctuation\">,<\/span> chunks<span class=\"token operator\">=<\/span><span class=\"token number\">32<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token keyword\">for<\/span> x<span class=\"token punctuation\">,<\/span> y <span class=\"token keyword\">in<\/span> dataloader<span class=\"token punctuation\">:<\/span>\r\n    losses<span class=\"token punctuation\">,<\/span> outputs <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">.<\/span>forward_backward<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">,<\/span> criterion<span class=\"token operator\">=<\/span>loss_fn<span class=\"token punctuation\">,<\/span> labels<span class=\"token operator\">=<\/span><span class=\"token punctuation\">(<\/span>y<span class=\"token punctuation\">,<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> return_outputs<span class=\"token operator\">=<\/span><span class=\"token boolean\">True<\/span><span class=\"token punctuation\">)<\/span>\r\n    <span class=\"token keyword\">if<\/span> rank <span class=\"token operator\">==<\/span> world_size <span class=\"token operator\">-<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span>\r\n        loss <span class=\"token operator\">=<\/span> losses<span class=\"token punctuation\">.<\/span><span class=\"token builtin\">sum<\/span><span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>item<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>  <span class=\"token comment\"># losses: torch.Size([32])<\/span>\r\n\r\n<span class=\"token comment\"># eval<\/span>\r\n<span class=\"token keyword\">with<\/span> torch<span class=\"token punctuation\">.<\/span>no_grad<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n    out <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">)<\/span>\r\n    <span class=\"token keyword\">if<\/span> rank <span class=\"token operator\">==<\/span> world_size <span class=\"token operator\">-<\/span> <span class=\"token number\">1<\/span><span class=\"token punctuation\">:<\/span>\r\n        <span class=\"token comment\"># calculate metrics ...<\/span><\/code><\/pre>\n<\/div>\n<h2 id=\"\u6570\u636e\u5e76\u884c\u548c\u6d41\u6c34\u7ebf\u5e76\u884c\u7ec4\u5408\">\u6570\u636e\u5e76\u884c\u548c\u6d41\u6c34\u7ebf\u5e76\u884c\u7ec4\u5408<\/h2>\n<p>haiscale \u8fd8\u652f\u6301\u540c\u65f6\u4f7f\u7528 DDP \u548c\u6d41\u6c34\u7ebf\u5e76\u884c\u3002\u6bd4\u5982\u6211\u4eec\u6709 16 \u5757 GPU\uff0c\u6211\u4eec\u53ef\u4ee5\u628a\u8fd9 16 \u5757 GPU \u5212\u5206\u6210\u4e24\u4e2a\u7ec4\uff0c\u6bcf\u4e2a\u7ec4\u6709 8 \u5757 GPU\uff0c\u7136\u540e\u4e24\u4e2a\u7ec4\u4e4b\u95f4\u505a\u6570\u636e\u5e76\u884c\uff0c\u7ec4\u5185\u505a\u6d41\u6c34\u7ebf\u5e76\u884c\u3002<\/p>\n<p>\u770b\u5982\u4e0b\u4f7f\u7528\u793a\u4f8b\uff1a<\/p>\n<div class=\"gatsby-highlight\" data-language=\"python\">\n<pre class=\"language-python\"><code class=\"language-python\"><span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>ddp <span class=\"token keyword\">import<\/span> DistributedDataParallel <span class=\"token keyword\">as<\/span> DDP\r\n<span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>pipeline <span class=\"token keyword\">import<\/span> PipeDream<span class=\"token punctuation\">,<\/span> partition<span class=\"token punctuation\">,<\/span> make_subgroups\r\n\r\ndist<span class=\"token punctuation\">.<\/span>init_process_group<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">)<\/span>\r\ntorch<span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">.<\/span>set_device<span class=\"token punctuation\">(<\/span>local_rank<span class=\"token punctuation\">)<\/span>\r\nrank<span class=\"token punctuation\">,<\/span> world_size <span class=\"token operator\">=<\/span> dist<span class=\"token punctuation\">.<\/span>get_rank<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> dist<span class=\"token punctuation\">.<\/span>get_world_size<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\ndp_group<span class=\"token punctuation\">,<\/span> pp_group <span class=\"token operator\">=<\/span> make_subgroups<span class=\"token punctuation\">(<\/span>pp_size<span class=\"token operator\">=<\/span><span class=\"token number\">8<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\nmodel <span class=\"token operator\">=<\/span> nn<span class=\"token punctuation\">.<\/span>Sequential<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">.<\/span><span class=\"token punctuation\">)<\/span>\r\nmodel <span class=\"token operator\">=<\/span> partition<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">,<\/span> pp_group<span class=\"token punctuation\">.<\/span>rank<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> pp_group<span class=\"token punctuation\">.<\/span>size<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\nmodel <span class=\"token operator\">=<\/span> DDP<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> process_group<span class=\"token operator\">=<\/span>dp_group<span class=\"token punctuation\">)<\/span>\r\nmodel <span class=\"token operator\">=<\/span> PipeDream<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">,<\/span> chunks<span class=\"token operator\">=<\/span><span class=\"token number\">64<\/span><span class=\"token punctuation\">,<\/span> process_group<span class=\"token operator\">=<\/span>pp_group<span class=\"token punctuation\">)<\/span>\r\n\r\ncriterion <span class=\"token operator\">=<\/span> nn<span class=\"token punctuation\">.<\/span>MSELoss<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token keyword\">for<\/span> x<span class=\"token punctuation\">,<\/span> y <span class=\"token keyword\">in<\/span> dataloader<span class=\"token punctuation\">:<\/span>\r\n    model<span class=\"token punctuation\">.<\/span>forward_backward<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">,<\/span> criterion<span class=\"token operator\">=<\/span>criterion<span class=\"token punctuation\">,<\/span> labels<span class=\"token operator\">=<\/span><span class=\"token punctuation\">(<\/span>y<span class=\"token punctuation\">,<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">)<\/span><\/code><\/pre>\n<\/div>\n<h2 id=\"cpu-offload\">CPU Offload<\/h2>\n<p>\u9664\u4e86\u4ee5\u4e0a\u7684\u5e76\u884c\u7b56\u7565\uff0c\u5728\u6df1\u5ea6\u5b66\u4e60\u7684\u8bad\u7ec3\u8fc7\u7a0b\u4e2d\uff0c\u6211\u4eec\u5e38\u5e38\u4f1a\u9047\u5230\u663e\u5b58\u4e0d\u8db3\u7684\u95ee\u9898\u3002<\/p>\n<p><code class=\"language-text\">haiscale.cpu_offload.CPUOffload<\/code>\u00a0\u80fd\u591f\u5e2e\u52a9\u6211\u4eec\u5728\u8bad\u7ec3\u4e2d\u628a\u4e00\u90e8\u5206\u9700\u8981\u4fdd\u5b58\u7684\u4e2d\u95f4\u53d8\u91cf\u79fb\u52a8\u5230 CPU \u5185\u5b58\u4e0a\uff0c\u7136\u540e\u5728\u53cd\u5411\u4f20\u64ad\u65f6\u628a\u9700\u8981\u7528\u5230\u7684 tensor \u4f20\u8f93\u56de GPU \u663e\u5b58\u91cc\uff0c\u4ece\u800c\u8fbe\u5230\u8282\u7701\u663e\u5b58\u7684\u76ee\u7684\u3002<\/p>\n<p>haiscale \u91c7\u7528<strong>\u5f02\u6b65\u4f20\u8f93\u62f7\u8d1d<\/strong>\u7b56\u7565\uff0c\u80fd\u591f\u628a\u4e00\u90e8\u5206\u7684\u4f20\u8f93\u65f6\u95f4\u548c GPU \u7684\u8ba1\u7b97\u91cd\u53e0\u8d77\u6765\uff0c\u4ece\u800c\u51cf\u5c11\u62f7\u8d1d\u5e26\u6765\u7684\u5f00\u9500\uff0c\u63d0\u5347\u6574\u4f53\u8ba1\u7b97\u6548\u7387\u3002<\/p>\n<p>\u4f7f\u7528\u65f6\u9700\u8981\u6307\u5b9a\u00a0<code class=\"language-text\">offload_ratio<\/code>\u00a0\u53c2\u6570\uff0c\u5176\u4ee3\u8868\u9700\u8981 offload \u7684\u4e2d\u95f4\u53d8\u91cf\u7684\u6bd4\u4f8b\uff0c<code class=\"language-text\">offload_ratio=1<\/code>\u00a0\u4ee3\u8868\u6240\u6709\u4fdd\u5b58\u7684\u4e2d\u95f4\u53d8\u91cf\u90fd\u4f1a\u88ab\u79fb\u52a8\u5230 CPU \u5185\u5b58\u91cc\u3002<\/p>\n<p>\u4ee5\u4e0b\u662f\u4f7f\u7528\u793a\u4f8b\uff1a<\/p>\n<div class=\"gatsby-highlight\" data-language=\"python\">\n<pre class=\"language-python\"><code class=\"language-python\"><span class=\"token keyword\">from<\/span> haiscale<span class=\"token punctuation\">.<\/span>cpu_offload <span class=\"token keyword\">import<\/span> CPUOffload\r\n\r\nmodel <span class=\"token operator\">=<\/span> MyModel<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>cuda<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\noptimizer <span class=\"token operator\">=<\/span> torch<span class=\"token punctuation\">.<\/span>optim<span class=\"token punctuation\">.<\/span>AdamW<span class=\"token punctuation\">(<\/span>model<span class=\"token punctuation\">.<\/span>parameters<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">,<\/span> lr<span class=\"token operator\">=<\/span><span class=\"token number\">0.01<\/span><span class=\"token punctuation\">)<\/span>\r\n\r\n<span class=\"token keyword\">for<\/span> x<span class=\"token punctuation\">,<\/span> y <span class=\"token keyword\">in<\/span> dataloader<span class=\"token punctuation\">:<\/span>\r\n    optimizer<span class=\"token punctuation\">.<\/span>zero_grad<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n    <span class=\"token keyword\">with<\/span> CPUOffload<span class=\"token punctuation\">(<\/span>offload_ratio<span class=\"token operator\">=<\/span><span class=\"token number\">0.1<\/span><span class=\"token punctuation\">,<\/span> tag<span class=\"token operator\">=<\/span><span class=\"token string\">\"MyModel\"<\/span><span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">:<\/span>\r\n        output <span class=\"token operator\">=<\/span> model<span class=\"token punctuation\">(<\/span>x<span class=\"token punctuation\">)<\/span>\r\n    loss_fn<span class=\"token punctuation\">(<\/span>y<span class=\"token punctuation\">,<\/span> output<span class=\"token punctuation\">)<\/span><span class=\"token punctuation\">.<\/span>backward<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span>\r\n    optimizer<span class=\"token punctuation\">.<\/span>step<span class=\"token punctuation\">(<\/span><span class=\"token punctuation\">)<\/span><\/code><\/pre>\n<\/div>","protected":false},"excerpt":{"rendered":"<p>haiscale (Highflyer AI Scale) \u662f\u4e00\u4e2a\u8f7b\u91cf\u7ea7\u7684\u9ad8\u6027\u80fd\u5e76\u884c\u8bad\u7ec3\u5de5\u5177\u5e93\uff0c\u5176\u6574\u5408\u4e86\u5e7b\u65b9  [&hellip;]<\/p>","protected":false},"author":1,"featured_media":447,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"_acf_changed":false,"footnotes":""},"categories":[5],"tags":[],"class_list":["post-446","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-firefly-platform"],"acf":[],"_links":{"self":[{"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/posts\/446","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/comments?post=446"}],"version-history":[{"count":1,"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/posts\/446\/revisions"}],"predecessor-version":[{"id":452,"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/posts\/446\/revisions\/452"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/media\/447"}],"wp:attachment":[{"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/media?parent=446"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/categories?post=446"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/high-flyer.in.suopu.cc\/en\/wp-json\/wp\/v2\/tags?post=446"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}