- {"little32", 0, {1,2,4,4,8, 4,4, 4,8}, /* win32 */
- {1,2,4,4,8, 4,4, 4,8}},
-
- {"little32_4", 0, {1,2,4,4,8, 4,4, 4,8}, /* linux x86 */
- {1,2,4,4,4, 4,4, 4,4}},
-
- {"little64", 0, {1,2,4,8,8, 8,8, 4,8}, /* alpha, ia64 */
- {1,2,4,8,8, 8,8, 4,8}},
-
- {"big32", 1, {1,2,4,4,8, 4,4, 4,8},
- {1,2,4,4,8, 4,4, 4,8}},
-
- {"big32_8_4", 1, {1,2,4,4,8, 4,4, 4,8}, /* AIX */
- {1,2,4,4,8, 4,4, 4,4}},
-
- {"big32_4", 1, {1,2,4,4,8, 4,4, 4,8}, /* G5 */
- {1,2,4,4,4, 4,4, 4,4}},
-
- {"big32_2", 1, {1,2,4,4,8, 4,4, 4,8}, /* ARM */
- {1,2,2,2,2, 2,2, 2,2}},
-
- {"big64", 1, {1,2,4,8,8, 8,8, 4,8}, /* sparc */
- {1,2,4,8,8, 8,8, 4,8}},
-
- {"big64_8_4", 1, {1,2,4,8,8, 8,8, 4,8}, /* aix with -maix64 */
- {1,2,4,8,8, 8,8, 4,4}}
+
+ {"little32_1", 0, {1, 2, 4, 4, 8, 4, 4, 4, 8}, /* little endian, 1 byte alignement (win32) */
+ {1, 1, 1, 1, 1, 1, 1, 1, 1}},
+
+ {"little32_2", 0, {1, 2, 4, 4, 8, 4, 4, 4, 8}, /* little endian, 2 bytes alignements (win32) */
+ {1, 2, 2, 2, 2, 2, 2, 2, 2}},
+
+ {"little32_4", 0, {1, 2, 4, 4, 8, 4, 4, 4, 8}, /* little endian, 4 bytes alignements (win32 and linux x86) */
+ {1, 2, 4, 4, 4, 4, 4, 4, 4}},
+
+ {"little32_8", 0, {1, 2, 4, 4, 8, 4, 4, 4, 8}, /* little endian, 8 bytes alignement (win32) */
+ {1, 2, 4, 4, 8, 4, 4, 4, 8}},
+
+ {"little64", 0, {1, 2, 4, 8, 8, 8, 8, 4, 8}, /* alpha, ia64 */
+ {1, 2, 4, 8, 8, 8, 8, 4, 8}},
+
+ {"big32_8", 1, {1, 2, 4, 4, 8, 4, 4, 4, 8},
+ {1, 2, 4, 4, 8, 4, 4, 4, 8}},
+
+ {"big32_8_4", 1, {1, 2, 4, 4, 8, 4, 4, 4, 8}, /* AIX */
+ {1, 2, 4, 4, 8, 4, 4, 4, 4}},
+
+ {"big32_4", 1, {1, 2, 4, 4, 8, 4, 4, 4, 8}, /* G5 */
+ {1, 2, 4, 4, 4, 4, 4, 4, 4}},
+
+ {"big32_2", 1, {1, 2, 4, 4, 8, 4, 4, 4, 8}, /* ARM */
+ {1, 2, 2, 2, 2, 2, 2, 2, 2}},
+
+ {"big64", 1, {1, 2, 4, 8, 8, 8, 8, 4, 8}, /* sparc */
+ {1, 2, 4, 8, 8, 8, 8, 4, 8}},
+
+ {"big64_8_4", 1, {1, 2, 4, 8, 8, 8, 8, 4, 8}, /* aix with -maix64 */
+ {1, 2, 4, 8, 8, 8, 8, 4, 4}}