Abstract: Gate switching in deep learning operations can be reduced based on sparsity in the input data. A first element of an activation operand and a first element of a weight operand may be stored in input storage units associated with a multiplier in a processing element. The multiplier computes a product of the two elements, which may be stored in an output storage unit of the multiplier. After detecting that a second element of the activation operand or a second element of the weight operand is zero valued, gate switching is reduced by avoiding at least one gate switching needed for the multiply-accumulation operation. For instance, the input storage units may not be updated. A zero-valued data element may be stored in the output storage unit of the multiplier and used as a product of the second element of the activation operand and the second element of the weight operand.
| # | Name | Date |
|---|---|---|
| 1 | 202547096998-PRIORITY DOCUMENTS [08-10-2025(online)].pdf | 2025-10-08 |
| 2 | 202547096998-POWER OF AUTHORITY [08-10-2025(online)].pdf | 2025-10-08 |
| 3 | 202547096998-FORM 1 [08-10-2025(online)].pdf | 2025-10-08 |
| 4 | 202547096998-DRAWINGS [08-10-2025(online)].pdf | 2025-10-08 |
| 5 | 202547096998-DECLARATION OF INVENTORSHIP (FORM 5) [08-10-2025(online)].pdf | 2025-10-08 |
| 6 | 202547096998-COMPLETE SPECIFICATION [08-10-2025(online)].pdf | 2025-10-08 |