Memristor-based, non-von-Neumann architectures performing tensor operations directly in memory are a promising approach to address the ever-increasing demand for energy-efficient, high-throughput hardware accelerators for Machine Learning (ML) inference. A major challenge for the programmability and exploitation of such Computing-In-Memory (CIM) architectures consists in the efficient mapping of tensor operations from high-level ML frameworks to fixed-function hardware blocks implementing in-memory computations. We demonstrate the programmability of memristor-based accelerators with TC-CIM, a fully-automatic, end-to-end compilation flow from Tensor Comprehensions, a mathematical notation for tensor operations, to fixed-function memristor-ba...
Digital electronics has given rise to reliable, affordable, and scalable computing devices. However,...
In recent years, we are witnessing a trend toward in-memory computing for future generations of comp...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
International audienceMemristor-based, non-von-Neumann architectures performing tensor operations di...
Memristor-based, non-von-Neumann architectures performing tensor operations directly in memory are a...
140 pagesTensor algebra lives at the heart of big data applications. Where classical machine learnin...
Computation in-memory is a promising non-von Neumann approach aiming at completely diminishing the d...
Tensor computations are important mathematical operations for applications that rely on multidimensi...
Today's computing architectures and device technologies are unable to meet the increasingly stringen...
Tensors are higher-dimensional analogs of matrices, and represent a key data abstraction for many ap...
Computation-in-memory reverses the trend in von-Neumann processors by bringing the computation close...
This dissertation presents novel algorithmic techniques and data structures to help build scalable t...
Popular Machine Learning (ML) and High Performance Computing (HPC) workloads contribute to a signifi...
General-purpose computing systems have benefited from technology scaling for several decades but are...
Deploying deep learning models on various devices has become an important topic. The wave of hardwar...
Digital electronics has given rise to reliable, affordable, and scalable computing devices. However,...
In recent years, we are witnessing a trend toward in-memory computing for future generations of comp...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...
International audienceMemristor-based, non-von-Neumann architectures performing tensor operations di...
Memristor-based, non-von-Neumann architectures performing tensor operations directly in memory are a...
140 pagesTensor algebra lives at the heart of big data applications. Where classical machine learnin...
Computation in-memory is a promising non-von Neumann approach aiming at completely diminishing the d...
Tensor computations are important mathematical operations for applications that rely on multidimensi...
Today's computing architectures and device technologies are unable to meet the increasingly stringen...
Tensors are higher-dimensional analogs of matrices, and represent a key data abstraction for many ap...
Computation-in-memory reverses the trend in von-Neumann processors by bringing the computation close...
This dissertation presents novel algorithmic techniques and data structures to help build scalable t...
Popular Machine Learning (ML) and High Performance Computing (HPC) workloads contribute to a signifi...
General-purpose computing systems have benefited from technology scaling for several decades but are...
Deploying deep learning models on various devices has become an important topic. The wave of hardwar...
Digital electronics has given rise to reliable, affordable, and scalable computing devices. However,...
In recent years, we are witnessing a trend toward in-memory computing for future generations of comp...
In this thesis, we develop high performance algorithms for certain computations involving dense tens...