generate_error_codes.py 10 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273
  1. #!/usr/bin/env python
  2. # Copyright 2015 Cloudera Inc.
  3. #
  4. # Licensed under the Apache License, Version 2.0 (the "License");
  5. # you may not use this file except in compliance with the License.
  6. # You may obtain a copy of the License at
  7. #
  8. # http://www.apache.org/licenses/LICENSE-2.0
  9. #
  10. # Unless required by applicable law or agreed to in writing, software
  11. # distributed under the License is distributed on an "AS IS" BASIS,
  12. # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  13. # See the License for the specific language governing permissions and
  14. # limitations under the License.
  15. # For readability purposes we define the error codes and messages at the top of the
  16. # file. New codes and messages must be added here. Old error messages MUST NEVER BE
  17. # DELETED, but can be renamed. The tuple layout for a new entry is: error code enum name,
  18. # numeric error code, format string of the message.
  19. #
  20. # TODO Add support for SQL Error Codes
  21. # https://msdn.microsoft.com/en-us/library/ms714687%28v=vs.85%29.aspx
  22. from __future__ import print_function
  23. error_codes = (
  24. ("OK", 1, ""),
  25. ("GENERAL", 2, "$0"),
  26. ("CANCELLED", 3, "$0"),
  27. ("ANALYSIS_ERROR", 4, "$0"),
  28. ("NOT_IMPLEMENTED_ERROR", 5, "$0"),
  29. ("RUNTIME_ERROR", 6, "$0"),
  30. ("MEM_LIMIT_EXCEEDED", 7, "$0"),
  31. ("INTERNAL_ERROR", 8, "$0"),
  32. ("RECOVERABLE_ERROR", 9, "$0"),
  33. ("PARQUET_MULTIPLE_BLOCKS", 10,
  34. "Parquet files should not be split into multiple hdfs-blocks. file=$0"),
  35. ("PARQUET_COLUMN_METADATA_INVALID", 11,
  36. "Column metadata states there are $0 values, but read $1 values from column $2. "
  37. "file=$3"),
  38. ("PARQUET_HEADER_PAGE_SIZE_EXCEEDED", 12, "(unused)"),
  39. ("PARQUET_HEADER_EOF", 13,
  40. "ParquetScanner: reached EOF while deserializing data page header. file=$0"),
  41. ("PARQUET_GROUP_ROW_COUNT_ERROR", 14,
  42. "Metadata states that in group $0($1) there are $2 rows, but $3 rows were read."),
  43. ("PARQUET_GROUP_ROW_COUNT_OVERFLOW", 15, "(unused)"),
  44. ("PARQUET_MISSING_PRECISION", 16,
  45. "File '$0' column '$1' does not have the decimal precision set."),
  46. ("PARQUET_WRONG_PRECISION", 17,
  47. "File '$0' column '$1' has a precision that does not match the table metadata "
  48. " precision. File metadata precision: $2, table metadata precision: $3."),
  49. ("PARQUET_BAD_CONVERTED_TYPE", 18,
  50. "File '$0' column '$1' does not have converted type set to DECIMAL"),
  51. ("PARQUET_INCOMPATIBLE_DECIMAL", 19,
  52. "File '$0' column '$1' contains decimal data but the table metadata has type $2"),
  53. ("SEQUENCE_SCANNER_PARSE_ERROR", 20,
  54. "Problem parsing file $0 at $1$2"),
  55. ("SNAPPY_DECOMPRESS_INVALID_BLOCK_SIZE", 21,
  56. "Decompressor: block size is too big. Data is likely corrupt. Size: $0"),
  57. ("SNAPPY_DECOMPRESS_INVALID_COMPRESSED_LENGTH", 22,
  58. "Decompressor: invalid compressed length. Data is likely corrupt."),
  59. ("SNAPPY_DECOMPRESS_UNCOMPRESSED_LENGTH_FAILED", 23,
  60. "Snappy: GetUncompressedLength failed"),
  61. ("SNAPPY_DECOMPRESS_RAW_UNCOMPRESS_FAILED", 24,
  62. "SnappyBlock: RawUncompress failed"),
  63. ("SNAPPY_DECOMPRESS_DECOMPRESS_SIZE_INCORRECT", 25,
  64. "Snappy: Decompressed size is not correct."),
  65. ("HDFS_SCAN_NODE_UNKNOWN_DISK", 26, "Unknown disk id. "
  66. "This will negatively affect performance. "
  67. "Check your hdfs settings to enable block location metadata."),
  68. ("FRAGMENT_EXECUTOR", 27, "Reserved resource size ($0) is larger than "
  69. "query mem limit ($1), and will be restricted to $1. Configure the reservation "
  70. "size by setting RM_INITIAL_MEM."),
  71. ("PARTITIONED_HASH_JOIN_MAX_PARTITION_DEPTH", 28,
  72. "Cannot perform join at hash join node with id $0."
  73. " The input data was partitioned the maximum number of $1 times."
  74. " This could mean there is significant skew in the data or the memory limit is"
  75. " set too low."),
  76. ("PARTITIONED_AGG_MAX_PARTITION_DEPTH", 29,
  77. "Cannot perform aggregation at hash aggregation node with id $0."
  78. " The input data was partitioned the maximum number of $1 times."
  79. " This could mean there is significant skew in the data or the memory limit is"
  80. " set too low."),
  81. ("MISSING_BUILTIN", 30, "Builtin '$0' with symbol '$1' does not exist. "
  82. "Verify that all your impalads are the same version."),
  83. ("RPC_GENERAL_ERROR", 31, "RPC Error: $0"),
  84. ("RPC_TIMEOUT", 32, "RPC timed out"),
  85. ("UDF_VERIFY_FAILED", 33,
  86. "Failed to verify function $0 from LLVM module $1, see log for more details."),
  87. ("PARQUET_CORRUPT_VALUE", 34, "File $0 corrupt. RLE level data bytes = $1"),
  88. ("AVRO_DECIMAL_RESOLUTION_ERROR", 35, "Column '$0' has conflicting Avro decimal types. "
  89. "Table schema $1: $2, file schema $1: $3"),
  90. ("AVRO_DECIMAL_METADATA_MISMATCH", 36, "Column '$0' has conflicting Avro decimal types. "
  91. "Declared $1: $2, $1 in table's Avro schema: $3"),
  92. ("AVRO_SCHEMA_RESOLUTION_ERROR", 37, "Unresolvable types for column '$0': "
  93. "table type: $1, file type: $2"),
  94. ("AVRO_SCHEMA_METADATA_MISMATCH", 38, "Unresolvable types for column '$0': "
  95. "declared column type: $1, table's Avro schema type: $2"),
  96. ("AVRO_UNSUPPORTED_DEFAULT_VALUE", 39, "Field $0 is missing from file and default "
  97. "values of type $1 are not yet supported."),
  98. ("AVRO_MISSING_FIELD", 40, "Inconsistent table metadata. Mismatch between column "
  99. "definition and Avro schema: cannot read field $0 because there are only $1 fields."),
  100. ("AVRO_MISSING_DEFAULT", 41,
  101. "Field $0 is missing from file and does not have a default value."),
  102. ("AVRO_NULLABILITY_MISMATCH", 42,
  103. "Field $0 is nullable in the file schema but not the table schema."),
  104. ("AVRO_NOT_A_RECORD", 43,
  105. "Inconsistent table metadata. Field $0 is not a record in the Avro schema."),
  106. ("PARQUET_DEF_LEVEL_ERROR", 44, "Could not read definition level, even though metadata"
  107. " states there are $0 values remaining in data page. file=$1"),
  108. ("PARQUET_NUM_COL_VALS_ERROR", 45, "Mismatched number of values in column index $0 "
  109. "($1 vs. $2). file=$3"),
  110. ("PARQUET_DICT_DECODE_FAILURE", 46, "Failed to decode dictionary-encoded value. "
  111. "file=$0"),
  112. ("SSL_PASSWORD_CMD_FAILED", 47,
  113. "SSL private-key password command ('$0') failed with error: $1"),
  114. ("SSL_CERTIFICATE_PATH_BLANK", 48, "The SSL certificate path is blank"),
  115. ("SSL_PRIVATE_KEY_PATH_BLANK", 49, "The SSL private key path is blank"),
  116. ("SSL_CERTIFICATE_NOT_FOUND", 50, "The SSL certificate file does not exist at path $0"),
  117. ("SSL_PRIVATE_KEY_NOT_FOUND", 51, "The SSL private key file does not exist at path $0"),
  118. ("SSL_SOCKET_CREATION_FAILED", 52, "SSL socket creation failed: $0"),
  119. ("MEM_ALLOC_FAILED", 53, "Memory allocation of $0 bytes failed"),
  120. ("PARQUET_REP_LEVEL_ERROR", 54, "Could not read repetition level, even though metadata"
  121. " states there are $0 values remaining in data page. file=$1"),
  122. ("PARQUET_UNRECOGNIZED_SCHEMA", 55, "File '$0' has an incompatible Parquet schema for "
  123. "column '$1'. Column type: $2, Parquet schema:\\n$3"),
  124. ("COLLECTION_ALLOC_FAILED", 56, "Failed to allocate buffer for collection '$0'."),
  125. ("TMP_DEVICE_BLACKLISTED", 57,
  126. "Temporary device for directory $0 is blacklisted from a previous error and cannot "
  127. "be used."),
  128. ("TMP_FILE_BLACKLISTED", 58,
  129. "Temporary file $0 is blacklisted from a previous error and cannot be expanded."),
  130. ("RPC_CLIENT_CONNECT_FAILURE", 59,
  131. "RPC client failed to connect: $0"),
  132. ("STALE_METADATA_FILE_TOO_SHORT", 60, "Metadata for file '$0' appears stale. "
  133. "Try running \\\"refresh $1\\\" to reload the file metadata."),
  134. ("PARQUET_BAD_VERSION_NUMBER", 61, "File '$0' has an invalid version number: $1\\n"
  135. "This could be due to stale metadata. Try running \\\"refresh $2\\\"."),
  136. ("SCANNER_INCOMPLETE_READ", 62, "Tried to read $0 bytes but could only read $1 bytes. "
  137. "This may indicate data file corruption. (file $2, byte offset: $3)"),
  138. ("SCANNER_INVALID_READ", 63, "Invalid read of $0 bytes. This may indicate data file "
  139. "corruption. (file $1, byte offset: $2)"),
  140. ("AVRO_BAD_VERSION_HEADER", 64, "File '$0' has an invalid version header: $1\\n"
  141. "Make sure the file is an Avro data file."),
  142. ("IMPALA_2598_KERBEROS_SSL_DISALLOWED", 65, "Enabling server-to-server SSL connections "
  143. "in conjunction with Kerberos authentication is not supported at the same time. "
  144. "Disable server-to-server SSL by unsetting --ssl_client_ca_certificate."),
  145. )
  146. import sys
  147. import os
  148. # Verifies the uniqueness of the error constants and numeric error codes.
  149. def check_duplicates(codes):
  150. constants = {}
  151. num_codes = {}
  152. for row in codes:
  153. if row[0] in constants:
  154. print("Constant %s already used, please check definition of '%s'!" % \
  155. (row[0], constants[row[0]]))
  156. exit(1)
  157. if row[1] in num_codes:
  158. print("Numeric error code %d already used, please check definition of '%s'!" % \
  159. (row[1], num_codes[row[1]]))
  160. exit(1)
  161. constants[row[0]] = row[2]
  162. num_codes[row[1]] = row[2]
  163. preamble = """
  164. // Copyright 2015 Cloudera Inc.
  165. //
  166. // Licensed under the Apache License, Version 2.0 (the "License");
  167. // you may not use this file except in compliance with the License.
  168. // You may obtain a copy of the License at
  169. //
  170. // http://www.apache.org/licenses/LICENSE-2.0
  171. //
  172. // Unless required by applicable law or agreed to in writing, software
  173. // distributed under the License is distributed on an "AS IS" BASIS,
  174. // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  175. // See the License for the specific language governing permissions and
  176. // limitations under the License.
  177. //
  178. //
  179. // THIS FILE IS AUTO GENERATED BY generated_error_codes.py DO NOT MODIFY
  180. // IT BY HAND.
  181. //
  182. namespace cpp impala
  183. namespace java com.cloudera.impala.thrift
  184. """
  185. # The script will always generate the file, CMake will take care of running it only if
  186. # necessary.
  187. target_file = "ErrorCodes.thrift"
  188. # Check uniqueness of error constants and numeric codes
  189. check_duplicates(error_codes)
  190. fid = open(target_file, "w+")
  191. try:
  192. fid.write(preamble)
  193. fid.write("""\nenum TErrorCode {\n""")
  194. fid.write(",\n".join([" %s" % x[0] for x in error_codes]))
  195. fid.write("\n}")
  196. fid.write("\n")
  197. fid.write("const list<string> TErrorMessage = [\n")
  198. fid.write(",\n".join([" // %s\n \"%s\"" %(x[0], x[2]) for x in error_codes]))
  199. fid.write("\n]")
  200. finally:
  201. fid.close()
  202. print("%s created." % target_file)