You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
on aarch64 linux platform, PyTorch inference latencies are increased on torch 2.1 and 2.2 compared to torch2.0 when openblas backend is used for multi-threaded configuration. The regression is higher for larger thread counts.
On AWS Graviton3, c7g.4xl, with 16 threads, the inference latency with torch2.0 is Time elapsed: 2.777902126312256 seconds
whereas with torch 2.1 and later, it is Time elapsed: 4.907686471939087 seconds