mirror of
https://github.com/ehw-fit/ariths-gen.git
synced 2025-04-19 13:30:56 +01:00

* #10 CGP Circuits as inputs (#11) * CGP Circuits as inputs * #10 support of signed output in general circuit * input as output works * output connected to input (c) * automated verilog testing * output rename * Implemented CSA and Wallace tree multiplier composing of CSAs. Also did some code cleanup. * Typos fix and code cleanup. * Added new (approximate) multiplier architectures and did some minor changes regarding sign extension for c output formats. * Updated automated testing scripts. * Small bugfix in python code generation (I initially thought this line is useless). * Updated generated circuits folder. Co-authored-by: Vojta Mrazek <mrazek@fit.vutbr.cz>
1615 lines
104 KiB
C
1615 lines
104 KiB
C
#include <stdio.h>
|
|
#include <stdint.h>
|
|
|
|
int64_t s_arrmul12(int64_t a, int64_t b){
|
|
int64_t s_arrmul12_out = 0;
|
|
uint8_t s_arrmul12_and0_0 = 0;
|
|
uint8_t s_arrmul12_and1_0 = 0;
|
|
uint8_t s_arrmul12_and2_0 = 0;
|
|
uint8_t s_arrmul12_and3_0 = 0;
|
|
uint8_t s_arrmul12_and4_0 = 0;
|
|
uint8_t s_arrmul12_and5_0 = 0;
|
|
uint8_t s_arrmul12_and6_0 = 0;
|
|
uint8_t s_arrmul12_and7_0 = 0;
|
|
uint8_t s_arrmul12_and8_0 = 0;
|
|
uint8_t s_arrmul12_and9_0 = 0;
|
|
uint8_t s_arrmul12_and10_0 = 0;
|
|
uint8_t s_arrmul12_nand11_0 = 0;
|
|
uint8_t s_arrmul12_and0_1 = 0;
|
|
uint8_t s_arrmul12_ha0_1_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_1_and0 = 0;
|
|
uint8_t s_arrmul12_and1_1 = 0;
|
|
uint8_t s_arrmul12_fa1_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_1_or0 = 0;
|
|
uint8_t s_arrmul12_and2_1 = 0;
|
|
uint8_t s_arrmul12_fa2_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_1_or0 = 0;
|
|
uint8_t s_arrmul12_and3_1 = 0;
|
|
uint8_t s_arrmul12_fa3_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_1_or0 = 0;
|
|
uint8_t s_arrmul12_and4_1 = 0;
|
|
uint8_t s_arrmul12_fa4_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_1_or0 = 0;
|
|
uint8_t s_arrmul12_and5_1 = 0;
|
|
uint8_t s_arrmul12_fa5_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_1_or0 = 0;
|
|
uint8_t s_arrmul12_and6_1 = 0;
|
|
uint8_t s_arrmul12_fa6_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_1_or0 = 0;
|
|
uint8_t s_arrmul12_and7_1 = 0;
|
|
uint8_t s_arrmul12_fa7_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_1_or0 = 0;
|
|
uint8_t s_arrmul12_and8_1 = 0;
|
|
uint8_t s_arrmul12_fa8_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_1_or0 = 0;
|
|
uint8_t s_arrmul12_and9_1 = 0;
|
|
uint8_t s_arrmul12_fa9_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_1_or0 = 0;
|
|
uint8_t s_arrmul12_and10_1 = 0;
|
|
uint8_t s_arrmul12_fa10_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_1_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_1_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_1 = 0;
|
|
uint8_t s_arrmul12_fa11_1_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_1_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_1_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_1_or0 = 0;
|
|
uint8_t s_arrmul12_and0_2 = 0;
|
|
uint8_t s_arrmul12_ha0_2_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_2_and0 = 0;
|
|
uint8_t s_arrmul12_and1_2 = 0;
|
|
uint8_t s_arrmul12_fa1_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_2_or0 = 0;
|
|
uint8_t s_arrmul12_and2_2 = 0;
|
|
uint8_t s_arrmul12_fa2_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_2_or0 = 0;
|
|
uint8_t s_arrmul12_and3_2 = 0;
|
|
uint8_t s_arrmul12_fa3_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_2_or0 = 0;
|
|
uint8_t s_arrmul12_and4_2 = 0;
|
|
uint8_t s_arrmul12_fa4_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_2_or0 = 0;
|
|
uint8_t s_arrmul12_and5_2 = 0;
|
|
uint8_t s_arrmul12_fa5_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_2_or0 = 0;
|
|
uint8_t s_arrmul12_and6_2 = 0;
|
|
uint8_t s_arrmul12_fa6_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_2_or0 = 0;
|
|
uint8_t s_arrmul12_and7_2 = 0;
|
|
uint8_t s_arrmul12_fa7_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_2_or0 = 0;
|
|
uint8_t s_arrmul12_and8_2 = 0;
|
|
uint8_t s_arrmul12_fa8_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_2_or0 = 0;
|
|
uint8_t s_arrmul12_and9_2 = 0;
|
|
uint8_t s_arrmul12_fa9_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_2_or0 = 0;
|
|
uint8_t s_arrmul12_and10_2 = 0;
|
|
uint8_t s_arrmul12_fa10_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_2_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_2 = 0;
|
|
uint8_t s_arrmul12_fa11_2_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_2_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_2_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_2_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_2_or0 = 0;
|
|
uint8_t s_arrmul12_and0_3 = 0;
|
|
uint8_t s_arrmul12_ha0_3_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_3_and0 = 0;
|
|
uint8_t s_arrmul12_and1_3 = 0;
|
|
uint8_t s_arrmul12_fa1_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_3_or0 = 0;
|
|
uint8_t s_arrmul12_and2_3 = 0;
|
|
uint8_t s_arrmul12_fa2_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_3_or0 = 0;
|
|
uint8_t s_arrmul12_and3_3 = 0;
|
|
uint8_t s_arrmul12_fa3_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_3_or0 = 0;
|
|
uint8_t s_arrmul12_and4_3 = 0;
|
|
uint8_t s_arrmul12_fa4_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_3_or0 = 0;
|
|
uint8_t s_arrmul12_and5_3 = 0;
|
|
uint8_t s_arrmul12_fa5_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_3_or0 = 0;
|
|
uint8_t s_arrmul12_and6_3 = 0;
|
|
uint8_t s_arrmul12_fa6_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_3_or0 = 0;
|
|
uint8_t s_arrmul12_and7_3 = 0;
|
|
uint8_t s_arrmul12_fa7_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_3_or0 = 0;
|
|
uint8_t s_arrmul12_and8_3 = 0;
|
|
uint8_t s_arrmul12_fa8_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_3_or0 = 0;
|
|
uint8_t s_arrmul12_and9_3 = 0;
|
|
uint8_t s_arrmul12_fa9_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_3_or0 = 0;
|
|
uint8_t s_arrmul12_and10_3 = 0;
|
|
uint8_t s_arrmul12_fa10_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_3_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_3 = 0;
|
|
uint8_t s_arrmul12_fa11_3_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_3_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_3_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_3_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_3_or0 = 0;
|
|
uint8_t s_arrmul12_and0_4 = 0;
|
|
uint8_t s_arrmul12_ha0_4_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_4_and0 = 0;
|
|
uint8_t s_arrmul12_and1_4 = 0;
|
|
uint8_t s_arrmul12_fa1_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_4_or0 = 0;
|
|
uint8_t s_arrmul12_and2_4 = 0;
|
|
uint8_t s_arrmul12_fa2_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_4_or0 = 0;
|
|
uint8_t s_arrmul12_and3_4 = 0;
|
|
uint8_t s_arrmul12_fa3_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_4_or0 = 0;
|
|
uint8_t s_arrmul12_and4_4 = 0;
|
|
uint8_t s_arrmul12_fa4_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_4_or0 = 0;
|
|
uint8_t s_arrmul12_and5_4 = 0;
|
|
uint8_t s_arrmul12_fa5_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_4_or0 = 0;
|
|
uint8_t s_arrmul12_and6_4 = 0;
|
|
uint8_t s_arrmul12_fa6_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_4_or0 = 0;
|
|
uint8_t s_arrmul12_and7_4 = 0;
|
|
uint8_t s_arrmul12_fa7_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_4_or0 = 0;
|
|
uint8_t s_arrmul12_and8_4 = 0;
|
|
uint8_t s_arrmul12_fa8_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_4_or0 = 0;
|
|
uint8_t s_arrmul12_and9_4 = 0;
|
|
uint8_t s_arrmul12_fa9_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_4_or0 = 0;
|
|
uint8_t s_arrmul12_and10_4 = 0;
|
|
uint8_t s_arrmul12_fa10_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_4_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_4 = 0;
|
|
uint8_t s_arrmul12_fa11_4_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_4_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_4_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_4_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_4_or0 = 0;
|
|
uint8_t s_arrmul12_and0_5 = 0;
|
|
uint8_t s_arrmul12_ha0_5_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_5_and0 = 0;
|
|
uint8_t s_arrmul12_and1_5 = 0;
|
|
uint8_t s_arrmul12_fa1_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_5_or0 = 0;
|
|
uint8_t s_arrmul12_and2_5 = 0;
|
|
uint8_t s_arrmul12_fa2_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_5_or0 = 0;
|
|
uint8_t s_arrmul12_and3_5 = 0;
|
|
uint8_t s_arrmul12_fa3_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_5_or0 = 0;
|
|
uint8_t s_arrmul12_and4_5 = 0;
|
|
uint8_t s_arrmul12_fa4_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_5_or0 = 0;
|
|
uint8_t s_arrmul12_and5_5 = 0;
|
|
uint8_t s_arrmul12_fa5_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_5_or0 = 0;
|
|
uint8_t s_arrmul12_and6_5 = 0;
|
|
uint8_t s_arrmul12_fa6_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_5_or0 = 0;
|
|
uint8_t s_arrmul12_and7_5 = 0;
|
|
uint8_t s_arrmul12_fa7_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_5_or0 = 0;
|
|
uint8_t s_arrmul12_and8_5 = 0;
|
|
uint8_t s_arrmul12_fa8_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_5_or0 = 0;
|
|
uint8_t s_arrmul12_and9_5 = 0;
|
|
uint8_t s_arrmul12_fa9_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_5_or0 = 0;
|
|
uint8_t s_arrmul12_and10_5 = 0;
|
|
uint8_t s_arrmul12_fa10_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_5_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_5 = 0;
|
|
uint8_t s_arrmul12_fa11_5_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_5_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_5_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_5_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_5_or0 = 0;
|
|
uint8_t s_arrmul12_and0_6 = 0;
|
|
uint8_t s_arrmul12_ha0_6_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_6_and0 = 0;
|
|
uint8_t s_arrmul12_and1_6 = 0;
|
|
uint8_t s_arrmul12_fa1_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_6_or0 = 0;
|
|
uint8_t s_arrmul12_and2_6 = 0;
|
|
uint8_t s_arrmul12_fa2_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_6_or0 = 0;
|
|
uint8_t s_arrmul12_and3_6 = 0;
|
|
uint8_t s_arrmul12_fa3_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_6_or0 = 0;
|
|
uint8_t s_arrmul12_and4_6 = 0;
|
|
uint8_t s_arrmul12_fa4_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_6_or0 = 0;
|
|
uint8_t s_arrmul12_and5_6 = 0;
|
|
uint8_t s_arrmul12_fa5_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_6_or0 = 0;
|
|
uint8_t s_arrmul12_and6_6 = 0;
|
|
uint8_t s_arrmul12_fa6_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_6_or0 = 0;
|
|
uint8_t s_arrmul12_and7_6 = 0;
|
|
uint8_t s_arrmul12_fa7_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_6_or0 = 0;
|
|
uint8_t s_arrmul12_and8_6 = 0;
|
|
uint8_t s_arrmul12_fa8_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_6_or0 = 0;
|
|
uint8_t s_arrmul12_and9_6 = 0;
|
|
uint8_t s_arrmul12_fa9_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_6_or0 = 0;
|
|
uint8_t s_arrmul12_and10_6 = 0;
|
|
uint8_t s_arrmul12_fa10_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_6_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_6 = 0;
|
|
uint8_t s_arrmul12_fa11_6_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_6_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_6_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_6_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_6_or0 = 0;
|
|
uint8_t s_arrmul12_and0_7 = 0;
|
|
uint8_t s_arrmul12_ha0_7_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_7_and0 = 0;
|
|
uint8_t s_arrmul12_and1_7 = 0;
|
|
uint8_t s_arrmul12_fa1_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_7_or0 = 0;
|
|
uint8_t s_arrmul12_and2_7 = 0;
|
|
uint8_t s_arrmul12_fa2_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_7_or0 = 0;
|
|
uint8_t s_arrmul12_and3_7 = 0;
|
|
uint8_t s_arrmul12_fa3_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_7_or0 = 0;
|
|
uint8_t s_arrmul12_and4_7 = 0;
|
|
uint8_t s_arrmul12_fa4_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_7_or0 = 0;
|
|
uint8_t s_arrmul12_and5_7 = 0;
|
|
uint8_t s_arrmul12_fa5_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_7_or0 = 0;
|
|
uint8_t s_arrmul12_and6_7 = 0;
|
|
uint8_t s_arrmul12_fa6_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_7_or0 = 0;
|
|
uint8_t s_arrmul12_and7_7 = 0;
|
|
uint8_t s_arrmul12_fa7_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_7_or0 = 0;
|
|
uint8_t s_arrmul12_and8_7 = 0;
|
|
uint8_t s_arrmul12_fa8_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_7_or0 = 0;
|
|
uint8_t s_arrmul12_and9_7 = 0;
|
|
uint8_t s_arrmul12_fa9_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_7_or0 = 0;
|
|
uint8_t s_arrmul12_and10_7 = 0;
|
|
uint8_t s_arrmul12_fa10_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_7_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_7 = 0;
|
|
uint8_t s_arrmul12_fa11_7_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_7_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_7_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_7_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_7_or0 = 0;
|
|
uint8_t s_arrmul12_and0_8 = 0;
|
|
uint8_t s_arrmul12_ha0_8_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_8_and0 = 0;
|
|
uint8_t s_arrmul12_and1_8 = 0;
|
|
uint8_t s_arrmul12_fa1_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_8_or0 = 0;
|
|
uint8_t s_arrmul12_and2_8 = 0;
|
|
uint8_t s_arrmul12_fa2_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_8_or0 = 0;
|
|
uint8_t s_arrmul12_and3_8 = 0;
|
|
uint8_t s_arrmul12_fa3_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_8_or0 = 0;
|
|
uint8_t s_arrmul12_and4_8 = 0;
|
|
uint8_t s_arrmul12_fa4_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_8_or0 = 0;
|
|
uint8_t s_arrmul12_and5_8 = 0;
|
|
uint8_t s_arrmul12_fa5_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_8_or0 = 0;
|
|
uint8_t s_arrmul12_and6_8 = 0;
|
|
uint8_t s_arrmul12_fa6_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_8_or0 = 0;
|
|
uint8_t s_arrmul12_and7_8 = 0;
|
|
uint8_t s_arrmul12_fa7_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_8_or0 = 0;
|
|
uint8_t s_arrmul12_and8_8 = 0;
|
|
uint8_t s_arrmul12_fa8_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_8_or0 = 0;
|
|
uint8_t s_arrmul12_and9_8 = 0;
|
|
uint8_t s_arrmul12_fa9_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_8_or0 = 0;
|
|
uint8_t s_arrmul12_and10_8 = 0;
|
|
uint8_t s_arrmul12_fa10_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_8_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_8 = 0;
|
|
uint8_t s_arrmul12_fa11_8_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_8_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_8_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_8_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_8_or0 = 0;
|
|
uint8_t s_arrmul12_and0_9 = 0;
|
|
uint8_t s_arrmul12_ha0_9_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_9_and0 = 0;
|
|
uint8_t s_arrmul12_and1_9 = 0;
|
|
uint8_t s_arrmul12_fa1_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_9_or0 = 0;
|
|
uint8_t s_arrmul12_and2_9 = 0;
|
|
uint8_t s_arrmul12_fa2_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_9_or0 = 0;
|
|
uint8_t s_arrmul12_and3_9 = 0;
|
|
uint8_t s_arrmul12_fa3_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_9_or0 = 0;
|
|
uint8_t s_arrmul12_and4_9 = 0;
|
|
uint8_t s_arrmul12_fa4_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_9_or0 = 0;
|
|
uint8_t s_arrmul12_and5_9 = 0;
|
|
uint8_t s_arrmul12_fa5_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_9_or0 = 0;
|
|
uint8_t s_arrmul12_and6_9 = 0;
|
|
uint8_t s_arrmul12_fa6_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_9_or0 = 0;
|
|
uint8_t s_arrmul12_and7_9 = 0;
|
|
uint8_t s_arrmul12_fa7_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_9_or0 = 0;
|
|
uint8_t s_arrmul12_and8_9 = 0;
|
|
uint8_t s_arrmul12_fa8_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_9_or0 = 0;
|
|
uint8_t s_arrmul12_and9_9 = 0;
|
|
uint8_t s_arrmul12_fa9_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_9_or0 = 0;
|
|
uint8_t s_arrmul12_and10_9 = 0;
|
|
uint8_t s_arrmul12_fa10_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_9_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_9 = 0;
|
|
uint8_t s_arrmul12_fa11_9_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_9_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_9_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_9_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_9_or0 = 0;
|
|
uint8_t s_arrmul12_and0_10 = 0;
|
|
uint8_t s_arrmul12_ha0_10_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_10_and0 = 0;
|
|
uint8_t s_arrmul12_and1_10 = 0;
|
|
uint8_t s_arrmul12_fa1_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_10_or0 = 0;
|
|
uint8_t s_arrmul12_and2_10 = 0;
|
|
uint8_t s_arrmul12_fa2_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_10_or0 = 0;
|
|
uint8_t s_arrmul12_and3_10 = 0;
|
|
uint8_t s_arrmul12_fa3_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_10_or0 = 0;
|
|
uint8_t s_arrmul12_and4_10 = 0;
|
|
uint8_t s_arrmul12_fa4_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_10_or0 = 0;
|
|
uint8_t s_arrmul12_and5_10 = 0;
|
|
uint8_t s_arrmul12_fa5_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_10_or0 = 0;
|
|
uint8_t s_arrmul12_and6_10 = 0;
|
|
uint8_t s_arrmul12_fa6_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_10_or0 = 0;
|
|
uint8_t s_arrmul12_and7_10 = 0;
|
|
uint8_t s_arrmul12_fa7_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_10_or0 = 0;
|
|
uint8_t s_arrmul12_and8_10 = 0;
|
|
uint8_t s_arrmul12_fa8_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_10_or0 = 0;
|
|
uint8_t s_arrmul12_and9_10 = 0;
|
|
uint8_t s_arrmul12_fa9_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_10_or0 = 0;
|
|
uint8_t s_arrmul12_and10_10 = 0;
|
|
uint8_t s_arrmul12_fa10_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_10_or0 = 0;
|
|
uint8_t s_arrmul12_nand11_10 = 0;
|
|
uint8_t s_arrmul12_fa11_10_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_10_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_10_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_10_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_10_or0 = 0;
|
|
uint8_t s_arrmul12_nand0_11 = 0;
|
|
uint8_t s_arrmul12_ha0_11_xor0 = 0;
|
|
uint8_t s_arrmul12_ha0_11_and0 = 0;
|
|
uint8_t s_arrmul12_nand1_11 = 0;
|
|
uint8_t s_arrmul12_fa1_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa1_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa1_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa1_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa1_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand2_11 = 0;
|
|
uint8_t s_arrmul12_fa2_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa2_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa2_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa2_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa2_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand3_11 = 0;
|
|
uint8_t s_arrmul12_fa3_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa3_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa3_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa3_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa3_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand4_11 = 0;
|
|
uint8_t s_arrmul12_fa4_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa4_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa4_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa4_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa4_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand5_11 = 0;
|
|
uint8_t s_arrmul12_fa5_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa5_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa5_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa5_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa5_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand6_11 = 0;
|
|
uint8_t s_arrmul12_fa6_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa6_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa6_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa6_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa6_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand7_11 = 0;
|
|
uint8_t s_arrmul12_fa7_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa7_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa7_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa7_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa7_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand8_11 = 0;
|
|
uint8_t s_arrmul12_fa8_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa8_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa8_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa8_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa8_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand9_11 = 0;
|
|
uint8_t s_arrmul12_fa9_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa9_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa9_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa9_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa9_11_or0 = 0;
|
|
uint8_t s_arrmul12_nand10_11 = 0;
|
|
uint8_t s_arrmul12_fa10_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa10_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa10_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa10_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa10_11_or0 = 0;
|
|
uint8_t s_arrmul12_and11_11 = 0;
|
|
uint8_t s_arrmul12_fa11_11_xor0 = 0;
|
|
uint8_t s_arrmul12_fa11_11_and0 = 0;
|
|
uint8_t s_arrmul12_fa11_11_xor1 = 0;
|
|
uint8_t s_arrmul12_fa11_11_and1 = 0;
|
|
uint8_t s_arrmul12_fa11_11_or0 = 0;
|
|
uint8_t s_arrmul12_xor12_11 = 0;
|
|
|
|
s_arrmul12_and0_0 = ((a >> 0) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and1_0 = ((a >> 1) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and2_0 = ((a >> 2) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and3_0 = ((a >> 3) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and4_0 = ((a >> 4) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and5_0 = ((a >> 5) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and6_0 = ((a >> 6) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and7_0 = ((a >> 7) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and8_0 = ((a >> 8) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and9_0 = ((a >> 9) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_and10_0 = ((a >> 10) & 0x01) & ((b >> 0) & 0x01);
|
|
s_arrmul12_nand11_0 = ~(((a >> 11) & 0x01) & ((b >> 0) & 0x01)) & 0x01;
|
|
s_arrmul12_and0_1 = ((a >> 0) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_ha0_1_xor0 = ((s_arrmul12_and0_1 >> 0) & 0x01) ^ ((s_arrmul12_and1_0 >> 0) & 0x01);
|
|
s_arrmul12_ha0_1_and0 = ((s_arrmul12_and0_1 >> 0) & 0x01) & ((s_arrmul12_and1_0 >> 0) & 0x01);
|
|
s_arrmul12_and1_1 = ((a >> 1) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa1_1_xor0 = ((s_arrmul12_and1_1 >> 0) & 0x01) ^ ((s_arrmul12_and2_0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_1_and0 = ((s_arrmul12_and1_1 >> 0) & 0x01) & ((s_arrmul12_and2_0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_1_xor1 = ((s_arrmul12_fa1_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_1_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_1_and1 = ((s_arrmul12_fa1_1_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_1_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_1_or0 = ((s_arrmul12_fa1_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_1 = ((a >> 2) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa2_1_xor0 = ((s_arrmul12_and2_1 >> 0) & 0x01) ^ ((s_arrmul12_and3_0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_1_and0 = ((s_arrmul12_and2_1 >> 0) & 0x01) & ((s_arrmul12_and3_0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_1_xor1 = ((s_arrmul12_fa2_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_1_and1 = ((s_arrmul12_fa2_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_1_or0 = ((s_arrmul12_fa2_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_1 = ((a >> 3) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa3_1_xor0 = ((s_arrmul12_and3_1 >> 0) & 0x01) ^ ((s_arrmul12_and4_0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_1_and0 = ((s_arrmul12_and3_1 >> 0) & 0x01) & ((s_arrmul12_and4_0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_1_xor1 = ((s_arrmul12_fa3_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_1_and1 = ((s_arrmul12_fa3_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_1_or0 = ((s_arrmul12_fa3_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_1 = ((a >> 4) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa4_1_xor0 = ((s_arrmul12_and4_1 >> 0) & 0x01) ^ ((s_arrmul12_and5_0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_1_and0 = ((s_arrmul12_and4_1 >> 0) & 0x01) & ((s_arrmul12_and5_0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_1_xor1 = ((s_arrmul12_fa4_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_1_and1 = ((s_arrmul12_fa4_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_1_or0 = ((s_arrmul12_fa4_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_1 = ((a >> 5) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa5_1_xor0 = ((s_arrmul12_and5_1 >> 0) & 0x01) ^ ((s_arrmul12_and6_0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_1_and0 = ((s_arrmul12_and5_1 >> 0) & 0x01) & ((s_arrmul12_and6_0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_1_xor1 = ((s_arrmul12_fa5_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_1_and1 = ((s_arrmul12_fa5_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_1_or0 = ((s_arrmul12_fa5_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_1 = ((a >> 6) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa6_1_xor0 = ((s_arrmul12_and6_1 >> 0) & 0x01) ^ ((s_arrmul12_and7_0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_1_and0 = ((s_arrmul12_and6_1 >> 0) & 0x01) & ((s_arrmul12_and7_0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_1_xor1 = ((s_arrmul12_fa6_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_1_and1 = ((s_arrmul12_fa6_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_1_or0 = ((s_arrmul12_fa6_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_1 = ((a >> 7) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa7_1_xor0 = ((s_arrmul12_and7_1 >> 0) & 0x01) ^ ((s_arrmul12_and8_0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_1_and0 = ((s_arrmul12_and7_1 >> 0) & 0x01) & ((s_arrmul12_and8_0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_1_xor1 = ((s_arrmul12_fa7_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_1_and1 = ((s_arrmul12_fa7_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_1_or0 = ((s_arrmul12_fa7_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_1 = ((a >> 8) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa8_1_xor0 = ((s_arrmul12_and8_1 >> 0) & 0x01) ^ ((s_arrmul12_and9_0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_1_and0 = ((s_arrmul12_and8_1 >> 0) & 0x01) & ((s_arrmul12_and9_0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_1_xor1 = ((s_arrmul12_fa8_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_1_and1 = ((s_arrmul12_fa8_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_1_or0 = ((s_arrmul12_fa8_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_1 = ((a >> 9) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa9_1_xor0 = ((s_arrmul12_and9_1 >> 0) & 0x01) ^ ((s_arrmul12_and10_0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_1_and0 = ((s_arrmul12_and9_1 >> 0) & 0x01) & ((s_arrmul12_and10_0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_1_xor1 = ((s_arrmul12_fa9_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_1_and1 = ((s_arrmul12_fa9_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_1_or0 = ((s_arrmul12_fa9_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_1 = ((a >> 10) & 0x01) & ((b >> 1) & 0x01);
|
|
s_arrmul12_fa10_1_xor0 = ((s_arrmul12_and10_1 >> 0) & 0x01) ^ ((s_arrmul12_nand11_0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_1_and0 = ((s_arrmul12_and10_1 >> 0) & 0x01) & ((s_arrmul12_nand11_0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_1_xor1 = ((s_arrmul12_fa10_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_1_and1 = ((s_arrmul12_fa10_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_1_or0 = ((s_arrmul12_fa10_1_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_1 = ~(((a >> 11) & 0x01) & ((b >> 1) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_1_xor0 = ~(((s_arrmul12_nand11_1 >> 0) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_1_xor1 = ((s_arrmul12_fa11_1_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_1_and1 = ((s_arrmul12_fa11_1_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_1_or0 = ((s_arrmul12_nand11_1 >> 0) & 0x01) | ((s_arrmul12_fa11_1_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_2 = ((a >> 0) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_ha0_2_xor0 = ((s_arrmul12_and0_2 >> 0) & 0x01) ^ ((s_arrmul12_fa1_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_2_and0 = ((s_arrmul12_and0_2 >> 0) & 0x01) & ((s_arrmul12_fa1_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_2 = ((a >> 1) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa1_2_xor0 = ((s_arrmul12_and1_2 >> 0) & 0x01) ^ ((s_arrmul12_fa2_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_2_and0 = ((s_arrmul12_and1_2 >> 0) & 0x01) & ((s_arrmul12_fa2_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_2_xor1 = ((s_arrmul12_fa1_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_2_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_2_and1 = ((s_arrmul12_fa1_2_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_2_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_2_or0 = ((s_arrmul12_fa1_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_2 = ((a >> 2) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa2_2_xor0 = ((s_arrmul12_and2_2 >> 0) & 0x01) ^ ((s_arrmul12_fa3_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_2_and0 = ((s_arrmul12_and2_2 >> 0) & 0x01) & ((s_arrmul12_fa3_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_2_xor1 = ((s_arrmul12_fa2_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_2_and1 = ((s_arrmul12_fa2_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_2_or0 = ((s_arrmul12_fa2_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_2 = ((a >> 3) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa3_2_xor0 = ((s_arrmul12_and3_2 >> 0) & 0x01) ^ ((s_arrmul12_fa4_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_2_and0 = ((s_arrmul12_and3_2 >> 0) & 0x01) & ((s_arrmul12_fa4_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_2_xor1 = ((s_arrmul12_fa3_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_2_and1 = ((s_arrmul12_fa3_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_2_or0 = ((s_arrmul12_fa3_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_2 = ((a >> 4) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa4_2_xor0 = ((s_arrmul12_and4_2 >> 0) & 0x01) ^ ((s_arrmul12_fa5_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_2_and0 = ((s_arrmul12_and4_2 >> 0) & 0x01) & ((s_arrmul12_fa5_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_2_xor1 = ((s_arrmul12_fa4_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_2_and1 = ((s_arrmul12_fa4_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_2_or0 = ((s_arrmul12_fa4_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_2 = ((a >> 5) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa5_2_xor0 = ((s_arrmul12_and5_2 >> 0) & 0x01) ^ ((s_arrmul12_fa6_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_2_and0 = ((s_arrmul12_and5_2 >> 0) & 0x01) & ((s_arrmul12_fa6_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_2_xor1 = ((s_arrmul12_fa5_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_2_and1 = ((s_arrmul12_fa5_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_2_or0 = ((s_arrmul12_fa5_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_2 = ((a >> 6) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa6_2_xor0 = ((s_arrmul12_and6_2 >> 0) & 0x01) ^ ((s_arrmul12_fa7_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_2_and0 = ((s_arrmul12_and6_2 >> 0) & 0x01) & ((s_arrmul12_fa7_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_2_xor1 = ((s_arrmul12_fa6_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_2_and1 = ((s_arrmul12_fa6_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_2_or0 = ((s_arrmul12_fa6_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_2 = ((a >> 7) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa7_2_xor0 = ((s_arrmul12_and7_2 >> 0) & 0x01) ^ ((s_arrmul12_fa8_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_2_and0 = ((s_arrmul12_and7_2 >> 0) & 0x01) & ((s_arrmul12_fa8_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_2_xor1 = ((s_arrmul12_fa7_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_2_and1 = ((s_arrmul12_fa7_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_2_or0 = ((s_arrmul12_fa7_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_2 = ((a >> 8) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa8_2_xor0 = ((s_arrmul12_and8_2 >> 0) & 0x01) ^ ((s_arrmul12_fa9_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_2_and0 = ((s_arrmul12_and8_2 >> 0) & 0x01) & ((s_arrmul12_fa9_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_2_xor1 = ((s_arrmul12_fa8_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_2_and1 = ((s_arrmul12_fa8_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_2_or0 = ((s_arrmul12_fa8_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_2 = ((a >> 9) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa9_2_xor0 = ((s_arrmul12_and9_2 >> 0) & 0x01) ^ ((s_arrmul12_fa10_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_2_and0 = ((s_arrmul12_and9_2 >> 0) & 0x01) & ((s_arrmul12_fa10_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_2_xor1 = ((s_arrmul12_fa9_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_2_and1 = ((s_arrmul12_fa9_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_2_or0 = ((s_arrmul12_fa9_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_2 = ((a >> 10) & 0x01) & ((b >> 2) & 0x01);
|
|
s_arrmul12_fa10_2_xor0 = ((s_arrmul12_and10_2 >> 0) & 0x01) ^ ((s_arrmul12_fa11_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_2_and0 = ((s_arrmul12_and10_2 >> 0) & 0x01) & ((s_arrmul12_fa11_1_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_2_xor1 = ((s_arrmul12_fa10_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_2_and1 = ((s_arrmul12_fa10_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_2_or0 = ((s_arrmul12_fa10_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_2 = ~(((a >> 11) & 0x01) & ((b >> 2) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_2_xor0 = ((s_arrmul12_nand11_2 >> 0) & 0x01) ^ ((s_arrmul12_fa11_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_2_and0 = ((s_arrmul12_nand11_2 >> 0) & 0x01) & ((s_arrmul12_fa11_1_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_2_xor1 = ((s_arrmul12_fa11_2_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_2_and1 = ((s_arrmul12_fa11_2_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_2_or0 = ((s_arrmul12_fa11_2_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_2_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_3 = ((a >> 0) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_ha0_3_xor0 = ((s_arrmul12_and0_3 >> 0) & 0x01) ^ ((s_arrmul12_fa1_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_3_and0 = ((s_arrmul12_and0_3 >> 0) & 0x01) & ((s_arrmul12_fa1_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_3 = ((a >> 1) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa1_3_xor0 = ((s_arrmul12_and1_3 >> 0) & 0x01) ^ ((s_arrmul12_fa2_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_3_and0 = ((s_arrmul12_and1_3 >> 0) & 0x01) & ((s_arrmul12_fa2_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_3_xor1 = ((s_arrmul12_fa1_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_3_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_3_and1 = ((s_arrmul12_fa1_3_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_3_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_3_or0 = ((s_arrmul12_fa1_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_3 = ((a >> 2) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa2_3_xor0 = ((s_arrmul12_and2_3 >> 0) & 0x01) ^ ((s_arrmul12_fa3_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_3_and0 = ((s_arrmul12_and2_3 >> 0) & 0x01) & ((s_arrmul12_fa3_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_3_xor1 = ((s_arrmul12_fa2_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_3_and1 = ((s_arrmul12_fa2_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_3_or0 = ((s_arrmul12_fa2_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_3 = ((a >> 3) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa3_3_xor0 = ((s_arrmul12_and3_3 >> 0) & 0x01) ^ ((s_arrmul12_fa4_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_3_and0 = ((s_arrmul12_and3_3 >> 0) & 0x01) & ((s_arrmul12_fa4_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_3_xor1 = ((s_arrmul12_fa3_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_3_and1 = ((s_arrmul12_fa3_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_3_or0 = ((s_arrmul12_fa3_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_3 = ((a >> 4) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa4_3_xor0 = ((s_arrmul12_and4_3 >> 0) & 0x01) ^ ((s_arrmul12_fa5_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_3_and0 = ((s_arrmul12_and4_3 >> 0) & 0x01) & ((s_arrmul12_fa5_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_3_xor1 = ((s_arrmul12_fa4_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_3_and1 = ((s_arrmul12_fa4_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_3_or0 = ((s_arrmul12_fa4_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_3 = ((a >> 5) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa5_3_xor0 = ((s_arrmul12_and5_3 >> 0) & 0x01) ^ ((s_arrmul12_fa6_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_3_and0 = ((s_arrmul12_and5_3 >> 0) & 0x01) & ((s_arrmul12_fa6_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_3_xor1 = ((s_arrmul12_fa5_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_3_and1 = ((s_arrmul12_fa5_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_3_or0 = ((s_arrmul12_fa5_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_3 = ((a >> 6) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa6_3_xor0 = ((s_arrmul12_and6_3 >> 0) & 0x01) ^ ((s_arrmul12_fa7_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_3_and0 = ((s_arrmul12_and6_3 >> 0) & 0x01) & ((s_arrmul12_fa7_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_3_xor1 = ((s_arrmul12_fa6_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_3_and1 = ((s_arrmul12_fa6_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_3_or0 = ((s_arrmul12_fa6_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_3 = ((a >> 7) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa7_3_xor0 = ((s_arrmul12_and7_3 >> 0) & 0x01) ^ ((s_arrmul12_fa8_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_3_and0 = ((s_arrmul12_and7_3 >> 0) & 0x01) & ((s_arrmul12_fa8_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_3_xor1 = ((s_arrmul12_fa7_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_3_and1 = ((s_arrmul12_fa7_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_3_or0 = ((s_arrmul12_fa7_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_3 = ((a >> 8) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa8_3_xor0 = ((s_arrmul12_and8_3 >> 0) & 0x01) ^ ((s_arrmul12_fa9_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_3_and0 = ((s_arrmul12_and8_3 >> 0) & 0x01) & ((s_arrmul12_fa9_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_3_xor1 = ((s_arrmul12_fa8_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_3_and1 = ((s_arrmul12_fa8_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_3_or0 = ((s_arrmul12_fa8_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_3 = ((a >> 9) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa9_3_xor0 = ((s_arrmul12_and9_3 >> 0) & 0x01) ^ ((s_arrmul12_fa10_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_3_and0 = ((s_arrmul12_and9_3 >> 0) & 0x01) & ((s_arrmul12_fa10_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_3_xor1 = ((s_arrmul12_fa9_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_3_and1 = ((s_arrmul12_fa9_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_3_or0 = ((s_arrmul12_fa9_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_3 = ((a >> 10) & 0x01) & ((b >> 3) & 0x01);
|
|
s_arrmul12_fa10_3_xor0 = ((s_arrmul12_and10_3 >> 0) & 0x01) ^ ((s_arrmul12_fa11_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_3_and0 = ((s_arrmul12_and10_3 >> 0) & 0x01) & ((s_arrmul12_fa11_2_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_3_xor1 = ((s_arrmul12_fa10_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_3_and1 = ((s_arrmul12_fa10_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_3_or0 = ((s_arrmul12_fa10_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_3 = ~(((a >> 11) & 0x01) & ((b >> 3) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_3_xor0 = ((s_arrmul12_nand11_3 >> 0) & 0x01) ^ ((s_arrmul12_fa11_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_3_and0 = ((s_arrmul12_nand11_3 >> 0) & 0x01) & ((s_arrmul12_fa11_2_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_3_xor1 = ((s_arrmul12_fa11_3_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_3_and1 = ((s_arrmul12_fa11_3_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_3_or0 = ((s_arrmul12_fa11_3_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_3_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_4 = ((a >> 0) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_ha0_4_xor0 = ((s_arrmul12_and0_4 >> 0) & 0x01) ^ ((s_arrmul12_fa1_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_4_and0 = ((s_arrmul12_and0_4 >> 0) & 0x01) & ((s_arrmul12_fa1_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_4 = ((a >> 1) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa1_4_xor0 = ((s_arrmul12_and1_4 >> 0) & 0x01) ^ ((s_arrmul12_fa2_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_4_and0 = ((s_arrmul12_and1_4 >> 0) & 0x01) & ((s_arrmul12_fa2_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_4_xor1 = ((s_arrmul12_fa1_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_4_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_4_and1 = ((s_arrmul12_fa1_4_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_4_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_4_or0 = ((s_arrmul12_fa1_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_4 = ((a >> 2) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa2_4_xor0 = ((s_arrmul12_and2_4 >> 0) & 0x01) ^ ((s_arrmul12_fa3_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_4_and0 = ((s_arrmul12_and2_4 >> 0) & 0x01) & ((s_arrmul12_fa3_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_4_xor1 = ((s_arrmul12_fa2_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_4_and1 = ((s_arrmul12_fa2_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_4_or0 = ((s_arrmul12_fa2_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_4 = ((a >> 3) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa3_4_xor0 = ((s_arrmul12_and3_4 >> 0) & 0x01) ^ ((s_arrmul12_fa4_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_4_and0 = ((s_arrmul12_and3_4 >> 0) & 0x01) & ((s_arrmul12_fa4_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_4_xor1 = ((s_arrmul12_fa3_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_4_and1 = ((s_arrmul12_fa3_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_4_or0 = ((s_arrmul12_fa3_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_4 = ((a >> 4) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa4_4_xor0 = ((s_arrmul12_and4_4 >> 0) & 0x01) ^ ((s_arrmul12_fa5_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_4_and0 = ((s_arrmul12_and4_4 >> 0) & 0x01) & ((s_arrmul12_fa5_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_4_xor1 = ((s_arrmul12_fa4_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_4_and1 = ((s_arrmul12_fa4_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_4_or0 = ((s_arrmul12_fa4_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_4 = ((a >> 5) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa5_4_xor0 = ((s_arrmul12_and5_4 >> 0) & 0x01) ^ ((s_arrmul12_fa6_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_4_and0 = ((s_arrmul12_and5_4 >> 0) & 0x01) & ((s_arrmul12_fa6_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_4_xor1 = ((s_arrmul12_fa5_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_4_and1 = ((s_arrmul12_fa5_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_4_or0 = ((s_arrmul12_fa5_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_4 = ((a >> 6) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa6_4_xor0 = ((s_arrmul12_and6_4 >> 0) & 0x01) ^ ((s_arrmul12_fa7_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_4_and0 = ((s_arrmul12_and6_4 >> 0) & 0x01) & ((s_arrmul12_fa7_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_4_xor1 = ((s_arrmul12_fa6_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_4_and1 = ((s_arrmul12_fa6_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_4_or0 = ((s_arrmul12_fa6_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_4 = ((a >> 7) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa7_4_xor0 = ((s_arrmul12_and7_4 >> 0) & 0x01) ^ ((s_arrmul12_fa8_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_4_and0 = ((s_arrmul12_and7_4 >> 0) & 0x01) & ((s_arrmul12_fa8_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_4_xor1 = ((s_arrmul12_fa7_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_4_and1 = ((s_arrmul12_fa7_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_4_or0 = ((s_arrmul12_fa7_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_4 = ((a >> 8) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa8_4_xor0 = ((s_arrmul12_and8_4 >> 0) & 0x01) ^ ((s_arrmul12_fa9_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_4_and0 = ((s_arrmul12_and8_4 >> 0) & 0x01) & ((s_arrmul12_fa9_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_4_xor1 = ((s_arrmul12_fa8_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_4_and1 = ((s_arrmul12_fa8_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_4_or0 = ((s_arrmul12_fa8_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_4 = ((a >> 9) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa9_4_xor0 = ((s_arrmul12_and9_4 >> 0) & 0x01) ^ ((s_arrmul12_fa10_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_4_and0 = ((s_arrmul12_and9_4 >> 0) & 0x01) & ((s_arrmul12_fa10_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_4_xor1 = ((s_arrmul12_fa9_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_4_and1 = ((s_arrmul12_fa9_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_4_or0 = ((s_arrmul12_fa9_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_4 = ((a >> 10) & 0x01) & ((b >> 4) & 0x01);
|
|
s_arrmul12_fa10_4_xor0 = ((s_arrmul12_and10_4 >> 0) & 0x01) ^ ((s_arrmul12_fa11_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_4_and0 = ((s_arrmul12_and10_4 >> 0) & 0x01) & ((s_arrmul12_fa11_3_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_4_xor1 = ((s_arrmul12_fa10_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_4_and1 = ((s_arrmul12_fa10_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_4_or0 = ((s_arrmul12_fa10_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_4 = ~(((a >> 11) & 0x01) & ((b >> 4) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_4_xor0 = ((s_arrmul12_nand11_4 >> 0) & 0x01) ^ ((s_arrmul12_fa11_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_4_and0 = ((s_arrmul12_nand11_4 >> 0) & 0x01) & ((s_arrmul12_fa11_3_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_4_xor1 = ((s_arrmul12_fa11_4_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_4_and1 = ((s_arrmul12_fa11_4_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_4_or0 = ((s_arrmul12_fa11_4_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_4_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_5 = ((a >> 0) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_ha0_5_xor0 = ((s_arrmul12_and0_5 >> 0) & 0x01) ^ ((s_arrmul12_fa1_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_5_and0 = ((s_arrmul12_and0_5 >> 0) & 0x01) & ((s_arrmul12_fa1_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_5 = ((a >> 1) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa1_5_xor0 = ((s_arrmul12_and1_5 >> 0) & 0x01) ^ ((s_arrmul12_fa2_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_5_and0 = ((s_arrmul12_and1_5 >> 0) & 0x01) & ((s_arrmul12_fa2_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_5_xor1 = ((s_arrmul12_fa1_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_5_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_5_and1 = ((s_arrmul12_fa1_5_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_5_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_5_or0 = ((s_arrmul12_fa1_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_5 = ((a >> 2) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa2_5_xor0 = ((s_arrmul12_and2_5 >> 0) & 0x01) ^ ((s_arrmul12_fa3_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_5_and0 = ((s_arrmul12_and2_5 >> 0) & 0x01) & ((s_arrmul12_fa3_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_5_xor1 = ((s_arrmul12_fa2_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_5_and1 = ((s_arrmul12_fa2_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_5_or0 = ((s_arrmul12_fa2_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_5 = ((a >> 3) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa3_5_xor0 = ((s_arrmul12_and3_5 >> 0) & 0x01) ^ ((s_arrmul12_fa4_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_5_and0 = ((s_arrmul12_and3_5 >> 0) & 0x01) & ((s_arrmul12_fa4_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_5_xor1 = ((s_arrmul12_fa3_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_5_and1 = ((s_arrmul12_fa3_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_5_or0 = ((s_arrmul12_fa3_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_5 = ((a >> 4) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa4_5_xor0 = ((s_arrmul12_and4_5 >> 0) & 0x01) ^ ((s_arrmul12_fa5_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_5_and0 = ((s_arrmul12_and4_5 >> 0) & 0x01) & ((s_arrmul12_fa5_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_5_xor1 = ((s_arrmul12_fa4_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_5_and1 = ((s_arrmul12_fa4_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_5_or0 = ((s_arrmul12_fa4_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_5 = ((a >> 5) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa5_5_xor0 = ((s_arrmul12_and5_5 >> 0) & 0x01) ^ ((s_arrmul12_fa6_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_5_and0 = ((s_arrmul12_and5_5 >> 0) & 0x01) & ((s_arrmul12_fa6_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_5_xor1 = ((s_arrmul12_fa5_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_5_and1 = ((s_arrmul12_fa5_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_5_or0 = ((s_arrmul12_fa5_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_5 = ((a >> 6) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa6_5_xor0 = ((s_arrmul12_and6_5 >> 0) & 0x01) ^ ((s_arrmul12_fa7_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_5_and0 = ((s_arrmul12_and6_5 >> 0) & 0x01) & ((s_arrmul12_fa7_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_5_xor1 = ((s_arrmul12_fa6_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_5_and1 = ((s_arrmul12_fa6_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_5_or0 = ((s_arrmul12_fa6_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_5 = ((a >> 7) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa7_5_xor0 = ((s_arrmul12_and7_5 >> 0) & 0x01) ^ ((s_arrmul12_fa8_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_5_and0 = ((s_arrmul12_and7_5 >> 0) & 0x01) & ((s_arrmul12_fa8_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_5_xor1 = ((s_arrmul12_fa7_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_5_and1 = ((s_arrmul12_fa7_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_5_or0 = ((s_arrmul12_fa7_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_5 = ((a >> 8) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa8_5_xor0 = ((s_arrmul12_and8_5 >> 0) & 0x01) ^ ((s_arrmul12_fa9_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_5_and0 = ((s_arrmul12_and8_5 >> 0) & 0x01) & ((s_arrmul12_fa9_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_5_xor1 = ((s_arrmul12_fa8_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_5_and1 = ((s_arrmul12_fa8_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_5_or0 = ((s_arrmul12_fa8_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_5 = ((a >> 9) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa9_5_xor0 = ((s_arrmul12_and9_5 >> 0) & 0x01) ^ ((s_arrmul12_fa10_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_5_and0 = ((s_arrmul12_and9_5 >> 0) & 0x01) & ((s_arrmul12_fa10_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_5_xor1 = ((s_arrmul12_fa9_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_5_and1 = ((s_arrmul12_fa9_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_5_or0 = ((s_arrmul12_fa9_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_5 = ((a >> 10) & 0x01) & ((b >> 5) & 0x01);
|
|
s_arrmul12_fa10_5_xor0 = ((s_arrmul12_and10_5 >> 0) & 0x01) ^ ((s_arrmul12_fa11_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_5_and0 = ((s_arrmul12_and10_5 >> 0) & 0x01) & ((s_arrmul12_fa11_4_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_5_xor1 = ((s_arrmul12_fa10_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_5_and1 = ((s_arrmul12_fa10_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_5_or0 = ((s_arrmul12_fa10_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_5 = ~(((a >> 11) & 0x01) & ((b >> 5) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_5_xor0 = ((s_arrmul12_nand11_5 >> 0) & 0x01) ^ ((s_arrmul12_fa11_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_5_and0 = ((s_arrmul12_nand11_5 >> 0) & 0x01) & ((s_arrmul12_fa11_4_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_5_xor1 = ((s_arrmul12_fa11_5_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_5_and1 = ((s_arrmul12_fa11_5_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_5_or0 = ((s_arrmul12_fa11_5_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_5_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_6 = ((a >> 0) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_ha0_6_xor0 = ((s_arrmul12_and0_6 >> 0) & 0x01) ^ ((s_arrmul12_fa1_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_6_and0 = ((s_arrmul12_and0_6 >> 0) & 0x01) & ((s_arrmul12_fa1_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_6 = ((a >> 1) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa1_6_xor0 = ((s_arrmul12_and1_6 >> 0) & 0x01) ^ ((s_arrmul12_fa2_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_6_and0 = ((s_arrmul12_and1_6 >> 0) & 0x01) & ((s_arrmul12_fa2_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_6_xor1 = ((s_arrmul12_fa1_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_6_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_6_and1 = ((s_arrmul12_fa1_6_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_6_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_6_or0 = ((s_arrmul12_fa1_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_6 = ((a >> 2) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa2_6_xor0 = ((s_arrmul12_and2_6 >> 0) & 0x01) ^ ((s_arrmul12_fa3_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_6_and0 = ((s_arrmul12_and2_6 >> 0) & 0x01) & ((s_arrmul12_fa3_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_6_xor1 = ((s_arrmul12_fa2_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_6_and1 = ((s_arrmul12_fa2_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_6_or0 = ((s_arrmul12_fa2_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_6 = ((a >> 3) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa3_6_xor0 = ((s_arrmul12_and3_6 >> 0) & 0x01) ^ ((s_arrmul12_fa4_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_6_and0 = ((s_arrmul12_and3_6 >> 0) & 0x01) & ((s_arrmul12_fa4_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_6_xor1 = ((s_arrmul12_fa3_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_6_and1 = ((s_arrmul12_fa3_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_6_or0 = ((s_arrmul12_fa3_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_6 = ((a >> 4) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa4_6_xor0 = ((s_arrmul12_and4_6 >> 0) & 0x01) ^ ((s_arrmul12_fa5_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_6_and0 = ((s_arrmul12_and4_6 >> 0) & 0x01) & ((s_arrmul12_fa5_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_6_xor1 = ((s_arrmul12_fa4_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_6_and1 = ((s_arrmul12_fa4_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_6_or0 = ((s_arrmul12_fa4_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_6 = ((a >> 5) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa5_6_xor0 = ((s_arrmul12_and5_6 >> 0) & 0x01) ^ ((s_arrmul12_fa6_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_6_and0 = ((s_arrmul12_and5_6 >> 0) & 0x01) & ((s_arrmul12_fa6_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_6_xor1 = ((s_arrmul12_fa5_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_6_and1 = ((s_arrmul12_fa5_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_6_or0 = ((s_arrmul12_fa5_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_6 = ((a >> 6) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa6_6_xor0 = ((s_arrmul12_and6_6 >> 0) & 0x01) ^ ((s_arrmul12_fa7_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_6_and0 = ((s_arrmul12_and6_6 >> 0) & 0x01) & ((s_arrmul12_fa7_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_6_xor1 = ((s_arrmul12_fa6_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_6_and1 = ((s_arrmul12_fa6_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_6_or0 = ((s_arrmul12_fa6_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_6 = ((a >> 7) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa7_6_xor0 = ((s_arrmul12_and7_6 >> 0) & 0x01) ^ ((s_arrmul12_fa8_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_6_and0 = ((s_arrmul12_and7_6 >> 0) & 0x01) & ((s_arrmul12_fa8_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_6_xor1 = ((s_arrmul12_fa7_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_6_and1 = ((s_arrmul12_fa7_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_6_or0 = ((s_arrmul12_fa7_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_6 = ((a >> 8) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa8_6_xor0 = ((s_arrmul12_and8_6 >> 0) & 0x01) ^ ((s_arrmul12_fa9_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_6_and0 = ((s_arrmul12_and8_6 >> 0) & 0x01) & ((s_arrmul12_fa9_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_6_xor1 = ((s_arrmul12_fa8_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_6_and1 = ((s_arrmul12_fa8_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_6_or0 = ((s_arrmul12_fa8_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_6 = ((a >> 9) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa9_6_xor0 = ((s_arrmul12_and9_6 >> 0) & 0x01) ^ ((s_arrmul12_fa10_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_6_and0 = ((s_arrmul12_and9_6 >> 0) & 0x01) & ((s_arrmul12_fa10_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_6_xor1 = ((s_arrmul12_fa9_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_6_and1 = ((s_arrmul12_fa9_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_6_or0 = ((s_arrmul12_fa9_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_6 = ((a >> 10) & 0x01) & ((b >> 6) & 0x01);
|
|
s_arrmul12_fa10_6_xor0 = ((s_arrmul12_and10_6 >> 0) & 0x01) ^ ((s_arrmul12_fa11_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_6_and0 = ((s_arrmul12_and10_6 >> 0) & 0x01) & ((s_arrmul12_fa11_5_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_6_xor1 = ((s_arrmul12_fa10_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_6_and1 = ((s_arrmul12_fa10_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_6_or0 = ((s_arrmul12_fa10_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_6 = ~(((a >> 11) & 0x01) & ((b >> 6) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_6_xor0 = ((s_arrmul12_nand11_6 >> 0) & 0x01) ^ ((s_arrmul12_fa11_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_6_and0 = ((s_arrmul12_nand11_6 >> 0) & 0x01) & ((s_arrmul12_fa11_5_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_6_xor1 = ((s_arrmul12_fa11_6_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_6_and1 = ((s_arrmul12_fa11_6_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_6_or0 = ((s_arrmul12_fa11_6_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_6_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_7 = ((a >> 0) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_ha0_7_xor0 = ((s_arrmul12_and0_7 >> 0) & 0x01) ^ ((s_arrmul12_fa1_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_7_and0 = ((s_arrmul12_and0_7 >> 0) & 0x01) & ((s_arrmul12_fa1_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_7 = ((a >> 1) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa1_7_xor0 = ((s_arrmul12_and1_7 >> 0) & 0x01) ^ ((s_arrmul12_fa2_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_7_and0 = ((s_arrmul12_and1_7 >> 0) & 0x01) & ((s_arrmul12_fa2_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_7_xor1 = ((s_arrmul12_fa1_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_7_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_7_and1 = ((s_arrmul12_fa1_7_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_7_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_7_or0 = ((s_arrmul12_fa1_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_7 = ((a >> 2) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa2_7_xor0 = ((s_arrmul12_and2_7 >> 0) & 0x01) ^ ((s_arrmul12_fa3_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_7_and0 = ((s_arrmul12_and2_7 >> 0) & 0x01) & ((s_arrmul12_fa3_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_7_xor1 = ((s_arrmul12_fa2_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_7_and1 = ((s_arrmul12_fa2_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_7_or0 = ((s_arrmul12_fa2_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_7 = ((a >> 3) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa3_7_xor0 = ((s_arrmul12_and3_7 >> 0) & 0x01) ^ ((s_arrmul12_fa4_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_7_and0 = ((s_arrmul12_and3_7 >> 0) & 0x01) & ((s_arrmul12_fa4_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_7_xor1 = ((s_arrmul12_fa3_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_7_and1 = ((s_arrmul12_fa3_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_7_or0 = ((s_arrmul12_fa3_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_7 = ((a >> 4) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa4_7_xor0 = ((s_arrmul12_and4_7 >> 0) & 0x01) ^ ((s_arrmul12_fa5_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_7_and0 = ((s_arrmul12_and4_7 >> 0) & 0x01) & ((s_arrmul12_fa5_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_7_xor1 = ((s_arrmul12_fa4_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_7_and1 = ((s_arrmul12_fa4_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_7_or0 = ((s_arrmul12_fa4_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_7 = ((a >> 5) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa5_7_xor0 = ((s_arrmul12_and5_7 >> 0) & 0x01) ^ ((s_arrmul12_fa6_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_7_and0 = ((s_arrmul12_and5_7 >> 0) & 0x01) & ((s_arrmul12_fa6_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_7_xor1 = ((s_arrmul12_fa5_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_7_and1 = ((s_arrmul12_fa5_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_7_or0 = ((s_arrmul12_fa5_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_7 = ((a >> 6) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa6_7_xor0 = ((s_arrmul12_and6_7 >> 0) & 0x01) ^ ((s_arrmul12_fa7_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_7_and0 = ((s_arrmul12_and6_7 >> 0) & 0x01) & ((s_arrmul12_fa7_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_7_xor1 = ((s_arrmul12_fa6_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_7_and1 = ((s_arrmul12_fa6_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_7_or0 = ((s_arrmul12_fa6_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_7 = ((a >> 7) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa7_7_xor0 = ((s_arrmul12_and7_7 >> 0) & 0x01) ^ ((s_arrmul12_fa8_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_7_and0 = ((s_arrmul12_and7_7 >> 0) & 0x01) & ((s_arrmul12_fa8_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_7_xor1 = ((s_arrmul12_fa7_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_7_and1 = ((s_arrmul12_fa7_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_7_or0 = ((s_arrmul12_fa7_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_7 = ((a >> 8) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa8_7_xor0 = ((s_arrmul12_and8_7 >> 0) & 0x01) ^ ((s_arrmul12_fa9_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_7_and0 = ((s_arrmul12_and8_7 >> 0) & 0x01) & ((s_arrmul12_fa9_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_7_xor1 = ((s_arrmul12_fa8_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_7_and1 = ((s_arrmul12_fa8_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_7_or0 = ((s_arrmul12_fa8_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_7 = ((a >> 9) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa9_7_xor0 = ((s_arrmul12_and9_7 >> 0) & 0x01) ^ ((s_arrmul12_fa10_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_7_and0 = ((s_arrmul12_and9_7 >> 0) & 0x01) & ((s_arrmul12_fa10_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_7_xor1 = ((s_arrmul12_fa9_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_7_and1 = ((s_arrmul12_fa9_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_7_or0 = ((s_arrmul12_fa9_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_7 = ((a >> 10) & 0x01) & ((b >> 7) & 0x01);
|
|
s_arrmul12_fa10_7_xor0 = ((s_arrmul12_and10_7 >> 0) & 0x01) ^ ((s_arrmul12_fa11_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_7_and0 = ((s_arrmul12_and10_7 >> 0) & 0x01) & ((s_arrmul12_fa11_6_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_7_xor1 = ((s_arrmul12_fa10_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_7_and1 = ((s_arrmul12_fa10_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_7_or0 = ((s_arrmul12_fa10_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_7 = ~(((a >> 11) & 0x01) & ((b >> 7) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_7_xor0 = ((s_arrmul12_nand11_7 >> 0) & 0x01) ^ ((s_arrmul12_fa11_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_7_and0 = ((s_arrmul12_nand11_7 >> 0) & 0x01) & ((s_arrmul12_fa11_6_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_7_xor1 = ((s_arrmul12_fa11_7_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_7_and1 = ((s_arrmul12_fa11_7_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_7_or0 = ((s_arrmul12_fa11_7_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_7_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_8 = ((a >> 0) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_ha0_8_xor0 = ((s_arrmul12_and0_8 >> 0) & 0x01) ^ ((s_arrmul12_fa1_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_8_and0 = ((s_arrmul12_and0_8 >> 0) & 0x01) & ((s_arrmul12_fa1_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_8 = ((a >> 1) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa1_8_xor0 = ((s_arrmul12_and1_8 >> 0) & 0x01) ^ ((s_arrmul12_fa2_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_8_and0 = ((s_arrmul12_and1_8 >> 0) & 0x01) & ((s_arrmul12_fa2_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_8_xor1 = ((s_arrmul12_fa1_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_8_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_8_and1 = ((s_arrmul12_fa1_8_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_8_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_8_or0 = ((s_arrmul12_fa1_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_8 = ((a >> 2) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa2_8_xor0 = ((s_arrmul12_and2_8 >> 0) & 0x01) ^ ((s_arrmul12_fa3_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_8_and0 = ((s_arrmul12_and2_8 >> 0) & 0x01) & ((s_arrmul12_fa3_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_8_xor1 = ((s_arrmul12_fa2_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_8_and1 = ((s_arrmul12_fa2_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_8_or0 = ((s_arrmul12_fa2_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_8 = ((a >> 3) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa3_8_xor0 = ((s_arrmul12_and3_8 >> 0) & 0x01) ^ ((s_arrmul12_fa4_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_8_and0 = ((s_arrmul12_and3_8 >> 0) & 0x01) & ((s_arrmul12_fa4_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_8_xor1 = ((s_arrmul12_fa3_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_8_and1 = ((s_arrmul12_fa3_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_8_or0 = ((s_arrmul12_fa3_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_8 = ((a >> 4) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa4_8_xor0 = ((s_arrmul12_and4_8 >> 0) & 0x01) ^ ((s_arrmul12_fa5_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_8_and0 = ((s_arrmul12_and4_8 >> 0) & 0x01) & ((s_arrmul12_fa5_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_8_xor1 = ((s_arrmul12_fa4_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_8_and1 = ((s_arrmul12_fa4_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_8_or0 = ((s_arrmul12_fa4_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_8 = ((a >> 5) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa5_8_xor0 = ((s_arrmul12_and5_8 >> 0) & 0x01) ^ ((s_arrmul12_fa6_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_8_and0 = ((s_arrmul12_and5_8 >> 0) & 0x01) & ((s_arrmul12_fa6_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_8_xor1 = ((s_arrmul12_fa5_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_8_and1 = ((s_arrmul12_fa5_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_8_or0 = ((s_arrmul12_fa5_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_8 = ((a >> 6) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa6_8_xor0 = ((s_arrmul12_and6_8 >> 0) & 0x01) ^ ((s_arrmul12_fa7_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_8_and0 = ((s_arrmul12_and6_8 >> 0) & 0x01) & ((s_arrmul12_fa7_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_8_xor1 = ((s_arrmul12_fa6_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_8_and1 = ((s_arrmul12_fa6_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_8_or0 = ((s_arrmul12_fa6_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_8 = ((a >> 7) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa7_8_xor0 = ((s_arrmul12_and7_8 >> 0) & 0x01) ^ ((s_arrmul12_fa8_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_8_and0 = ((s_arrmul12_and7_8 >> 0) & 0x01) & ((s_arrmul12_fa8_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_8_xor1 = ((s_arrmul12_fa7_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_8_and1 = ((s_arrmul12_fa7_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_8_or0 = ((s_arrmul12_fa7_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_8 = ((a >> 8) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa8_8_xor0 = ((s_arrmul12_and8_8 >> 0) & 0x01) ^ ((s_arrmul12_fa9_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_8_and0 = ((s_arrmul12_and8_8 >> 0) & 0x01) & ((s_arrmul12_fa9_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_8_xor1 = ((s_arrmul12_fa8_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_8_and1 = ((s_arrmul12_fa8_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_8_or0 = ((s_arrmul12_fa8_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_8 = ((a >> 9) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa9_8_xor0 = ((s_arrmul12_and9_8 >> 0) & 0x01) ^ ((s_arrmul12_fa10_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_8_and0 = ((s_arrmul12_and9_8 >> 0) & 0x01) & ((s_arrmul12_fa10_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_8_xor1 = ((s_arrmul12_fa9_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_8_and1 = ((s_arrmul12_fa9_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_8_or0 = ((s_arrmul12_fa9_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_8 = ((a >> 10) & 0x01) & ((b >> 8) & 0x01);
|
|
s_arrmul12_fa10_8_xor0 = ((s_arrmul12_and10_8 >> 0) & 0x01) ^ ((s_arrmul12_fa11_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_8_and0 = ((s_arrmul12_and10_8 >> 0) & 0x01) & ((s_arrmul12_fa11_7_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_8_xor1 = ((s_arrmul12_fa10_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_8_and1 = ((s_arrmul12_fa10_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_8_or0 = ((s_arrmul12_fa10_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_8 = ~(((a >> 11) & 0x01) & ((b >> 8) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_8_xor0 = ((s_arrmul12_nand11_8 >> 0) & 0x01) ^ ((s_arrmul12_fa11_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_8_and0 = ((s_arrmul12_nand11_8 >> 0) & 0x01) & ((s_arrmul12_fa11_7_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_8_xor1 = ((s_arrmul12_fa11_8_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_8_and1 = ((s_arrmul12_fa11_8_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_8_or0 = ((s_arrmul12_fa11_8_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_8_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_9 = ((a >> 0) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_ha0_9_xor0 = ((s_arrmul12_and0_9 >> 0) & 0x01) ^ ((s_arrmul12_fa1_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_9_and0 = ((s_arrmul12_and0_9 >> 0) & 0x01) & ((s_arrmul12_fa1_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_9 = ((a >> 1) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa1_9_xor0 = ((s_arrmul12_and1_9 >> 0) & 0x01) ^ ((s_arrmul12_fa2_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_9_and0 = ((s_arrmul12_and1_9 >> 0) & 0x01) & ((s_arrmul12_fa2_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_9_xor1 = ((s_arrmul12_fa1_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_9_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_9_and1 = ((s_arrmul12_fa1_9_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_9_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_9_or0 = ((s_arrmul12_fa1_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_9 = ((a >> 2) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa2_9_xor0 = ((s_arrmul12_and2_9 >> 0) & 0x01) ^ ((s_arrmul12_fa3_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_9_and0 = ((s_arrmul12_and2_9 >> 0) & 0x01) & ((s_arrmul12_fa3_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_9_xor1 = ((s_arrmul12_fa2_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_9_and1 = ((s_arrmul12_fa2_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_9_or0 = ((s_arrmul12_fa2_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_9 = ((a >> 3) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa3_9_xor0 = ((s_arrmul12_and3_9 >> 0) & 0x01) ^ ((s_arrmul12_fa4_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_9_and0 = ((s_arrmul12_and3_9 >> 0) & 0x01) & ((s_arrmul12_fa4_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_9_xor1 = ((s_arrmul12_fa3_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_9_and1 = ((s_arrmul12_fa3_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_9_or0 = ((s_arrmul12_fa3_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_9 = ((a >> 4) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa4_9_xor0 = ((s_arrmul12_and4_9 >> 0) & 0x01) ^ ((s_arrmul12_fa5_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_9_and0 = ((s_arrmul12_and4_9 >> 0) & 0x01) & ((s_arrmul12_fa5_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_9_xor1 = ((s_arrmul12_fa4_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_9_and1 = ((s_arrmul12_fa4_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_9_or0 = ((s_arrmul12_fa4_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_9 = ((a >> 5) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa5_9_xor0 = ((s_arrmul12_and5_9 >> 0) & 0x01) ^ ((s_arrmul12_fa6_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_9_and0 = ((s_arrmul12_and5_9 >> 0) & 0x01) & ((s_arrmul12_fa6_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_9_xor1 = ((s_arrmul12_fa5_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_9_and1 = ((s_arrmul12_fa5_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_9_or0 = ((s_arrmul12_fa5_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_9 = ((a >> 6) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa6_9_xor0 = ((s_arrmul12_and6_9 >> 0) & 0x01) ^ ((s_arrmul12_fa7_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_9_and0 = ((s_arrmul12_and6_9 >> 0) & 0x01) & ((s_arrmul12_fa7_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_9_xor1 = ((s_arrmul12_fa6_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_9_and1 = ((s_arrmul12_fa6_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_9_or0 = ((s_arrmul12_fa6_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_9 = ((a >> 7) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa7_9_xor0 = ((s_arrmul12_and7_9 >> 0) & 0x01) ^ ((s_arrmul12_fa8_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_9_and0 = ((s_arrmul12_and7_9 >> 0) & 0x01) & ((s_arrmul12_fa8_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_9_xor1 = ((s_arrmul12_fa7_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_9_and1 = ((s_arrmul12_fa7_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_9_or0 = ((s_arrmul12_fa7_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_9 = ((a >> 8) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa8_9_xor0 = ((s_arrmul12_and8_9 >> 0) & 0x01) ^ ((s_arrmul12_fa9_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_9_and0 = ((s_arrmul12_and8_9 >> 0) & 0x01) & ((s_arrmul12_fa9_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_9_xor1 = ((s_arrmul12_fa8_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_9_and1 = ((s_arrmul12_fa8_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_9_or0 = ((s_arrmul12_fa8_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_9 = ((a >> 9) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa9_9_xor0 = ((s_arrmul12_and9_9 >> 0) & 0x01) ^ ((s_arrmul12_fa10_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_9_and0 = ((s_arrmul12_and9_9 >> 0) & 0x01) & ((s_arrmul12_fa10_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_9_xor1 = ((s_arrmul12_fa9_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_9_and1 = ((s_arrmul12_fa9_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_9_or0 = ((s_arrmul12_fa9_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_9 = ((a >> 10) & 0x01) & ((b >> 9) & 0x01);
|
|
s_arrmul12_fa10_9_xor0 = ((s_arrmul12_and10_9 >> 0) & 0x01) ^ ((s_arrmul12_fa11_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_9_and0 = ((s_arrmul12_and10_9 >> 0) & 0x01) & ((s_arrmul12_fa11_8_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_9_xor1 = ((s_arrmul12_fa10_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_9_and1 = ((s_arrmul12_fa10_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_9_or0 = ((s_arrmul12_fa10_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_9 = ~(((a >> 11) & 0x01) & ((b >> 9) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_9_xor0 = ((s_arrmul12_nand11_9 >> 0) & 0x01) ^ ((s_arrmul12_fa11_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_9_and0 = ((s_arrmul12_nand11_9 >> 0) & 0x01) & ((s_arrmul12_fa11_8_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_9_xor1 = ((s_arrmul12_fa11_9_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_9_and1 = ((s_arrmul12_fa11_9_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_9_or0 = ((s_arrmul12_fa11_9_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_9_and1 >> 0) & 0x01);
|
|
s_arrmul12_and0_10 = ((a >> 0) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_ha0_10_xor0 = ((s_arrmul12_and0_10 >> 0) & 0x01) ^ ((s_arrmul12_fa1_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_10_and0 = ((s_arrmul12_and0_10 >> 0) & 0x01) & ((s_arrmul12_fa1_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_and1_10 = ((a >> 1) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa1_10_xor0 = ((s_arrmul12_and1_10 >> 0) & 0x01) ^ ((s_arrmul12_fa2_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_10_and0 = ((s_arrmul12_and1_10 >> 0) & 0x01) & ((s_arrmul12_fa2_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_10_xor1 = ((s_arrmul12_fa1_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_10_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_10_and1 = ((s_arrmul12_fa1_10_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_10_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_10_or0 = ((s_arrmul12_fa1_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and2_10 = ((a >> 2) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa2_10_xor0 = ((s_arrmul12_and2_10 >> 0) & 0x01) ^ ((s_arrmul12_fa3_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_10_and0 = ((s_arrmul12_and2_10 >> 0) & 0x01) & ((s_arrmul12_fa3_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_10_xor1 = ((s_arrmul12_fa2_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_10_and1 = ((s_arrmul12_fa2_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_10_or0 = ((s_arrmul12_fa2_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and3_10 = ((a >> 3) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa3_10_xor0 = ((s_arrmul12_and3_10 >> 0) & 0x01) ^ ((s_arrmul12_fa4_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_10_and0 = ((s_arrmul12_and3_10 >> 0) & 0x01) & ((s_arrmul12_fa4_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_10_xor1 = ((s_arrmul12_fa3_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_10_and1 = ((s_arrmul12_fa3_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_10_or0 = ((s_arrmul12_fa3_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and4_10 = ((a >> 4) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa4_10_xor0 = ((s_arrmul12_and4_10 >> 0) & 0x01) ^ ((s_arrmul12_fa5_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_10_and0 = ((s_arrmul12_and4_10 >> 0) & 0x01) & ((s_arrmul12_fa5_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_10_xor1 = ((s_arrmul12_fa4_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_10_and1 = ((s_arrmul12_fa4_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_10_or0 = ((s_arrmul12_fa4_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and5_10 = ((a >> 5) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa5_10_xor0 = ((s_arrmul12_and5_10 >> 0) & 0x01) ^ ((s_arrmul12_fa6_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_10_and0 = ((s_arrmul12_and5_10 >> 0) & 0x01) & ((s_arrmul12_fa6_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_10_xor1 = ((s_arrmul12_fa5_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_10_and1 = ((s_arrmul12_fa5_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_10_or0 = ((s_arrmul12_fa5_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and6_10 = ((a >> 6) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa6_10_xor0 = ((s_arrmul12_and6_10 >> 0) & 0x01) ^ ((s_arrmul12_fa7_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_10_and0 = ((s_arrmul12_and6_10 >> 0) & 0x01) & ((s_arrmul12_fa7_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_10_xor1 = ((s_arrmul12_fa6_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_10_and1 = ((s_arrmul12_fa6_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_10_or0 = ((s_arrmul12_fa6_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and7_10 = ((a >> 7) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa7_10_xor0 = ((s_arrmul12_and7_10 >> 0) & 0x01) ^ ((s_arrmul12_fa8_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_10_and0 = ((s_arrmul12_and7_10 >> 0) & 0x01) & ((s_arrmul12_fa8_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_10_xor1 = ((s_arrmul12_fa7_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_10_and1 = ((s_arrmul12_fa7_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_10_or0 = ((s_arrmul12_fa7_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and8_10 = ((a >> 8) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa8_10_xor0 = ((s_arrmul12_and8_10 >> 0) & 0x01) ^ ((s_arrmul12_fa9_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_10_and0 = ((s_arrmul12_and8_10 >> 0) & 0x01) & ((s_arrmul12_fa9_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_10_xor1 = ((s_arrmul12_fa8_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_10_and1 = ((s_arrmul12_fa8_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_10_or0 = ((s_arrmul12_fa8_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and9_10 = ((a >> 9) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa9_10_xor0 = ((s_arrmul12_and9_10 >> 0) & 0x01) ^ ((s_arrmul12_fa10_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_10_and0 = ((s_arrmul12_and9_10 >> 0) & 0x01) & ((s_arrmul12_fa10_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_10_xor1 = ((s_arrmul12_fa9_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_10_and1 = ((s_arrmul12_fa9_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_10_or0 = ((s_arrmul12_fa9_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_and10_10 = ((a >> 10) & 0x01) & ((b >> 10) & 0x01);
|
|
s_arrmul12_fa10_10_xor0 = ((s_arrmul12_and10_10 >> 0) & 0x01) ^ ((s_arrmul12_fa11_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_10_and0 = ((s_arrmul12_and10_10 >> 0) & 0x01) & ((s_arrmul12_fa11_9_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_10_xor1 = ((s_arrmul12_fa10_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_10_and1 = ((s_arrmul12_fa10_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_10_or0 = ((s_arrmul12_fa10_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand11_10 = ~(((a >> 11) & 0x01) & ((b >> 10) & 0x01)) & 0x01;
|
|
s_arrmul12_fa11_10_xor0 = ((s_arrmul12_nand11_10 >> 0) & 0x01) ^ ((s_arrmul12_fa11_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_10_and0 = ((s_arrmul12_nand11_10 >> 0) & 0x01) & ((s_arrmul12_fa11_9_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_10_xor1 = ((s_arrmul12_fa11_10_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_10_and1 = ((s_arrmul12_fa11_10_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_10_or0 = ((s_arrmul12_fa11_10_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_10_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand0_11 = ~(((a >> 0) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_ha0_11_xor0 = ((s_arrmul12_nand0_11 >> 0) & 0x01) ^ ((s_arrmul12_fa1_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_ha0_11_and0 = ((s_arrmul12_nand0_11 >> 0) & 0x01) & ((s_arrmul12_fa1_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_nand1_11 = ~(((a >> 1) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa1_11_xor0 = ((s_arrmul12_nand1_11 >> 0) & 0x01) ^ ((s_arrmul12_fa2_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_11_and0 = ((s_arrmul12_nand1_11 >> 0) & 0x01) & ((s_arrmul12_fa2_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa1_11_xor1 = ((s_arrmul12_fa1_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_ha0_11_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_11_and1 = ((s_arrmul12_fa1_11_xor0 >> 0) & 0x01) & ((s_arrmul12_ha0_11_and0 >> 0) & 0x01);
|
|
s_arrmul12_fa1_11_or0 = ((s_arrmul12_fa1_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa1_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand2_11 = ~(((a >> 2) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa2_11_xor0 = ((s_arrmul12_nand2_11 >> 0) & 0x01) ^ ((s_arrmul12_fa3_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_11_and0 = ((s_arrmul12_nand2_11 >> 0) & 0x01) & ((s_arrmul12_fa3_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa2_11_xor1 = ((s_arrmul12_fa2_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa1_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_11_and1 = ((s_arrmul12_fa2_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa1_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa2_11_or0 = ((s_arrmul12_fa2_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa2_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand3_11 = ~(((a >> 3) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa3_11_xor0 = ((s_arrmul12_nand3_11 >> 0) & 0x01) ^ ((s_arrmul12_fa4_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_11_and0 = ((s_arrmul12_nand3_11 >> 0) & 0x01) & ((s_arrmul12_fa4_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa3_11_xor1 = ((s_arrmul12_fa3_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa2_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_11_and1 = ((s_arrmul12_fa3_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa2_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa3_11_or0 = ((s_arrmul12_fa3_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa3_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand4_11 = ~(((a >> 4) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa4_11_xor0 = ((s_arrmul12_nand4_11 >> 0) & 0x01) ^ ((s_arrmul12_fa5_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_11_and0 = ((s_arrmul12_nand4_11 >> 0) & 0x01) & ((s_arrmul12_fa5_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa4_11_xor1 = ((s_arrmul12_fa4_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa3_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_11_and1 = ((s_arrmul12_fa4_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa3_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa4_11_or0 = ((s_arrmul12_fa4_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa4_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand5_11 = ~(((a >> 5) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa5_11_xor0 = ((s_arrmul12_nand5_11 >> 0) & 0x01) ^ ((s_arrmul12_fa6_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_11_and0 = ((s_arrmul12_nand5_11 >> 0) & 0x01) & ((s_arrmul12_fa6_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa5_11_xor1 = ((s_arrmul12_fa5_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa4_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_11_and1 = ((s_arrmul12_fa5_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa4_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa5_11_or0 = ((s_arrmul12_fa5_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa5_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand6_11 = ~(((a >> 6) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa6_11_xor0 = ((s_arrmul12_nand6_11 >> 0) & 0x01) ^ ((s_arrmul12_fa7_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_11_and0 = ((s_arrmul12_nand6_11 >> 0) & 0x01) & ((s_arrmul12_fa7_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa6_11_xor1 = ((s_arrmul12_fa6_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa5_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_11_and1 = ((s_arrmul12_fa6_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa5_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa6_11_or0 = ((s_arrmul12_fa6_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa6_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand7_11 = ~(((a >> 7) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa7_11_xor0 = ((s_arrmul12_nand7_11 >> 0) & 0x01) ^ ((s_arrmul12_fa8_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_11_and0 = ((s_arrmul12_nand7_11 >> 0) & 0x01) & ((s_arrmul12_fa8_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa7_11_xor1 = ((s_arrmul12_fa7_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa6_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_11_and1 = ((s_arrmul12_fa7_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa6_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa7_11_or0 = ((s_arrmul12_fa7_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa7_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand8_11 = ~(((a >> 8) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa8_11_xor0 = ((s_arrmul12_nand8_11 >> 0) & 0x01) ^ ((s_arrmul12_fa9_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_11_and0 = ((s_arrmul12_nand8_11 >> 0) & 0x01) & ((s_arrmul12_fa9_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa8_11_xor1 = ((s_arrmul12_fa8_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa7_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_11_and1 = ((s_arrmul12_fa8_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa7_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa8_11_or0 = ((s_arrmul12_fa8_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa8_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand9_11 = ~(((a >> 9) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa9_11_xor0 = ((s_arrmul12_nand9_11 >> 0) & 0x01) ^ ((s_arrmul12_fa10_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_11_and0 = ((s_arrmul12_nand9_11 >> 0) & 0x01) & ((s_arrmul12_fa10_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa9_11_xor1 = ((s_arrmul12_fa9_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa8_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_11_and1 = ((s_arrmul12_fa9_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa8_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa9_11_or0 = ((s_arrmul12_fa9_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa9_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_nand10_11 = ~(((a >> 10) & 0x01) & ((b >> 11) & 0x01)) & 0x01;
|
|
s_arrmul12_fa10_11_xor0 = ((s_arrmul12_nand10_11 >> 0) & 0x01) ^ ((s_arrmul12_fa11_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_11_and0 = ((s_arrmul12_nand10_11 >> 0) & 0x01) & ((s_arrmul12_fa11_10_xor1 >> 0) & 0x01);
|
|
s_arrmul12_fa10_11_xor1 = ((s_arrmul12_fa10_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa9_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_11_and1 = ((s_arrmul12_fa10_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa9_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa10_11_or0 = ((s_arrmul12_fa10_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa10_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_and11_11 = ((a >> 11) & 0x01) & ((b >> 11) & 0x01);
|
|
s_arrmul12_fa11_11_xor0 = ((s_arrmul12_and11_11 >> 0) & 0x01) ^ ((s_arrmul12_fa11_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_11_and0 = ((s_arrmul12_and11_11 >> 0) & 0x01) & ((s_arrmul12_fa11_10_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_11_xor1 = ((s_arrmul12_fa11_11_xor0 >> 0) & 0x01) ^ ((s_arrmul12_fa10_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_11_and1 = ((s_arrmul12_fa11_11_xor0 >> 0) & 0x01) & ((s_arrmul12_fa10_11_or0 >> 0) & 0x01);
|
|
s_arrmul12_fa11_11_or0 = ((s_arrmul12_fa11_11_and0 >> 0) & 0x01) | ((s_arrmul12_fa11_11_and1 >> 0) & 0x01);
|
|
s_arrmul12_xor12_11 = ~(((s_arrmul12_fa11_11_or0 >> 0) & 0x01)) & 0x01;
|
|
|
|
s_arrmul12_out |= ((s_arrmul12_and0_0 >> 0) & 0x01ull) << 0;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_1_xor0 >> 0) & 0x01ull) << 1;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_2_xor0 >> 0) & 0x01ull) << 2;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_3_xor0 >> 0) & 0x01ull) << 3;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_4_xor0 >> 0) & 0x01ull) << 4;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_5_xor0 >> 0) & 0x01ull) << 5;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_6_xor0 >> 0) & 0x01ull) << 6;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_7_xor0 >> 0) & 0x01ull) << 7;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_8_xor0 >> 0) & 0x01ull) << 8;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_9_xor0 >> 0) & 0x01ull) << 9;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_10_xor0 >> 0) & 0x01ull) << 10;
|
|
s_arrmul12_out |= ((s_arrmul12_ha0_11_xor0 >> 0) & 0x01ull) << 11;
|
|
s_arrmul12_out |= ((s_arrmul12_fa1_11_xor1 >> 0) & 0x01ull) << 12;
|
|
s_arrmul12_out |= ((s_arrmul12_fa2_11_xor1 >> 0) & 0x01ull) << 13;
|
|
s_arrmul12_out |= ((s_arrmul12_fa3_11_xor1 >> 0) & 0x01ull) << 14;
|
|
s_arrmul12_out |= ((s_arrmul12_fa4_11_xor1 >> 0) & 0x01ull) << 15;
|
|
s_arrmul12_out |= ((s_arrmul12_fa5_11_xor1 >> 0) & 0x01ull) << 16;
|
|
s_arrmul12_out |= ((s_arrmul12_fa6_11_xor1 >> 0) & 0x01ull) << 17;
|
|
s_arrmul12_out |= ((s_arrmul12_fa7_11_xor1 >> 0) & 0x01ull) << 18;
|
|
s_arrmul12_out |= ((s_arrmul12_fa8_11_xor1 >> 0) & 0x01ull) << 19;
|
|
s_arrmul12_out |= ((s_arrmul12_fa9_11_xor1 >> 0) & 0x01ull) << 20;
|
|
s_arrmul12_out |= ((s_arrmul12_fa10_11_xor1 >> 0) & 0x01ull) << 21;
|
|
s_arrmul12_out |= ((s_arrmul12_fa11_11_xor1 >> 0) & 0x01ull) << 22;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 23;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 24;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 25;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 26;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 27;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 28;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 29;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 30;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 31;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 32;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 33;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 34;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 35;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 36;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 37;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 38;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 39;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 40;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 41;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 42;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 43;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 44;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 45;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 46;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 47;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 48;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 49;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 50;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 51;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 52;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 53;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 54;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 55;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 56;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 57;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 58;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 59;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 60;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 61;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 62;
|
|
s_arrmul12_out |= ((s_arrmul12_xor12_11 >> 0) & 0x01ull) << 63;
|
|
return s_arrmul12_out;
|
|
} |