Created
August 20, 2024 05:27
-
-
Save leslie-fang-intel/496fc4bc84e0bc544244eae8e6ebf061 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| import requests | |
| import torch | |
| print(torch.__version__) | |
| import torch.nn as nn | |
| import os, pickle | |
| import numpy as np | |
| import torch._inductor.config as config | |
| config.freezing = True | |
| config.max_autotune = True | |
| config.max_autotune_gemm_backends = "CPP,ATEN" | |
| config.fx_graph_cache = False | |
| config.fx_graph_remote_cache = False | |
| config.autotune_local_cache = False | |
| class M(torch.nn.Module): | |
| def __init__(self,): | |
| super().__init__() | |
| self.linear = torch.nn.Linear(1024, 4096 * 1024) | |
| # self.linear2 = torch.nn.Linear(4096 * 1024, 4096 * 1024) | |
| def forward(self, attn_weights): | |
| attn_weights = self.linear(attn_weights) | |
| # attn_weights = self.linear2(attn_weights) | |
| return attn_weights | |
| import time | |
| import psutil | |
| if __name__ == "__main__": | |
| time.sleep(30) | |
| print("Init psutil.virtual_memory() is: {}".format(psutil.virtual_memory()), flush=True) | |
| x = torch.randn(1, 1024) | |
| with torch.no_grad(), torch.autocast(device_type="cpu"): | |
| m = M().eval() | |
| print(x, flush=True) | |
| print(m(x), flush=True) | |
| cfn = torch.compile(m) | |
| print(cfn(x), flush=True) | |
| # time.sleep(30) | |
| print("After first run psutil.virtual_memory() is: {}".format(psutil.virtual_memory()), flush=True) | |
| for _ in range(10): | |
| cfn(x) | |
| # time.sleep(30) | |
| print("Final psutil.virtual_memory() is: {}".format(psutil.virtual_memory()), flush=True) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment