Massively parallel neural inference computing elements

R Appuswamy, JV Arthur, AS Cassidy, P Datta… - US Patent …, 2020 - Google Patents
Massively parallel neural inference computing elements are provided. A plurality of
multipliers is arranged in a plurality of equal-sized groups. Each of the plurality of multipliers …

Splitting neural network filters for implementation by neural network inference circuit

J Ko, K Duong, SL Teig - US Patent 11,250,326, 2022 - Google Patents
Some embodiments provide a method for compiling a neural network (NN) program for an
NN inference circuit (NNIC) that includes multiple partial dot product computation cir cuits …

Reduced dot product computation circuit

K Duong, J Ko, SL Teig - US Patent 10,740,434, 2020 - Google Patents
Some embodiments provide an IC for implementing a machine-trained network with multiple
layers. The IC includes a set of circuits to compute a dot product of (i) a first number of input …

Homogenizing data sparsity using a butterfly multiplexer

I Ovsiannikov, L Wang - US Patent 11,664,819, 2023 - Google Patents
US11664819B2 - Homogenizing data sparsity using a butterfly multiplexer - Google Patents
US11664819B2 - Homogenizing data sparsity using a butterfly multiplexer - Google Patents …

Dataflow all-reduce for reconfigurable processor systems

MR Raumann, Q Zheng, BB Shah, R Kumar… - US Patent …, 2022 - Google Patents
2021-09-08 Assigned to SambaNova Systems, Inc. reassignment SambaNova Systems, Inc.
ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors …

Adaptive and interchangeable neural networks

KA Abeloe - US Patent 10,691,133, 2020 - Google Patents
Methods and systems that allow neural network systems to maintain or increase operational
accuracy while being able to operate in various settings. A set of training data is collected …

Computation of neural network node by neural network inference circuit

K Duong, J Ko, SL Teig - US Patent 11,170,289, 2021 - Google Patents
Some embodiments provide a neural network inference circuit (NNIC) for executing a neural
network that includes multiple computation nodes, that include dot products, at multiple …

Compute optimizations for neural networks using bipolar binary weight

K Nealis, A Yao, X Chen, E Ould-Ahmed-Vall… - US Patent …, 2021 - Google Patents
One embodiment provides for a compute apparatus comprising a decode unit to decode a
single instruction into a decoded instruction that specifies multiple operands including a …

Storage of input values within core of neural network inference circuit

K Duong, J Ko, SL Teig - US Patent 11,468,145, 2022 - Google Patents
Some embodiments provide a neural network inference circuit (NNIC) for executing a NN
that includes multiple computation nodes at multiple layers. Each of a set of the computation …

Dataflow function offload to reconfigurable processors

MR Raumann, Q Zheng, BB Shah, R Kumar… - US Patent …, 2022 - Google Patents
Roughly described, the invention involves a system including a plurality of functional units
that execute different segments of a dataflow, and share intermediate results via a peer-to …