hive 報錯return code 40000 from org.apache.hadoop.hive.ql.exec.MoveTask解決思路

參考學習
https://github.com/apache/hive/blob/2b57dd27ad61e552f93817ac69313066af6562d9/ql/src/java/org/apache/hadoop/hive/ql/ErrorMsg.java#L47

為啥學習error code
開發過程中遇到以下錯誤,大家覺得應該怎么辦?從哪方面入手呢?
1.百度?
2.源碼查看報錯地方
3.忽略(這個錯是偶發的)

Error: Error while compiling statement: FAILED: Execution Error, return code 40000 from org.apache.hadoop.hive.ql.exec.MoveTask. org.apache.hadoop.hive.ql.metadata.HiveExceptionat org.apache.hadoop.hive.ql.metadata.Hive.copyFiles(Hive.java:4792)at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3180)at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:412)at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:213)at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:105)at org.apache.hadoop.hive.ql.Executor.launchTask(Executor.java:357)at org.apache.hadoop.hive.ql.Executor.launchTasks(Executor.java:330)at org.apache.hadoop.hive.ql.Executor.runTasks(Executor.java:246)at org.apache.hadoop.hive.ql.Executor.execute(Executor.java:109)at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:749)at org.apache.hadoop.hive.ql.Driver.run(Driver.java:504)at org.apache.hadoop.hive.ql.Driver.run(Driver.java:498)at org.apache.hadoop.hive.ql.reexec.ReExecDriver.run(ReExecDriver.java:166)at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:226)at org.apache.hive.service.cli.operation.SQLOperation.access$700(SQLOperation.java:88)at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:327)at java.security.AccessController.doPrivileged(Native Method)at javax.security.auth.Subject.doAs(Subject.java:422)at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1898)at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:345)at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)at java.util.concurrent.FutureTask.run(FutureTask.java:266)at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)at java.util.concurrent.FutureTask.run(FutureTask.java:266)at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)at java.lang.Thread.run(Thread.java:748) (state=08S01,code=40000)

但是這個錯是hive的錯,我們開發過程 中會遇到很多hive問題,都怎么解決呢?我建議看源碼,但先不要看報錯源碼,直接看errorCode源碼,別人hive的開發者也是人,也知道這里會報錯,可能會好心告訴我們怎么解決。

 // 10000 to 19999: Errors occurring during semantic analysis and compilation of the query.// 20000 to 29999: Runtime errors where Hive believes that retries are unlikely to succeed.// 30000 to 39999: Runtime errors which Hive thinks may be transient and retrying may succeed.// 40000 to 49999: Errors where Hive is unable to advise about retries.

比如上面我們的措施40000,別人明確告訴我們 retries就行,實際情況就是重試確實ok。
但是還是想知道為什么報錯,能不能有更好的解決辦法。還是得看源碼 哈哈哈。
以上述為例,因為我使用的集群是cdp3.1的hive版本
在這里插入圖片描述
源碼cdp肯定是不會給的,或者我不知道。那么直接看hive的源碼,畢竟cdp也是抄的hive官方的代碼。根據報錯信息

at org.apache.hadoop.hive.ql.metadata.Hive.copyFiles(Hive.java:4792)
at org.apache.hadoop.hive.ql.metadata.Hive.loadTable(Hive.java:3180)
at org.apache.hadoop.hive.ql.exec.MoveTask.execute(MoveTask.java:412)

定位到
在這里插入圖片描述
看到這里獲取配置項,就舒服了,畢竟就算我們發現假設hive源碼有問題,我也沒法修改cdp的源碼。但是配置有問題我們可以修改配置項。

    HIVE_MOVE_FILES_THREAD_COUNT("hive.mv.files.thread", 15, new  SizeValidator(0L, true, 1024L, true), "Number of threads"+ " used to move files in move task. Set it to 0 to disable multi-threaded file moves. This parameter is also used by"+ " MSCK to check tables."),HIVE_LOAD_DYNAMIC_PARTITIONS_THREAD_COUNT("hive.load.dynamic.partitions.thread", 15,new  SizeValidator(1L, true, 1024L, true),"Number of threads used to load dynamic partitions.")

顧名思義 這個配置項是move files的線程數,后面會根據這個線程數new一個線程池newFixedThreadPool
然后開始move file
在這里插入圖片描述
感受了 就是這里真正move的時候出錯了。繼續看mvFile干嘛。
在這里插入圖片描述
這里翻譯下代碼做了啥,和我們在干嘛。
我們做的是
insert into tableA select * from table1
insert into tableA select * from table2
比如說 table1中間有十個文件 table2也有10個文件 此時要把20個文件都mv到tableA目錄下。此時默認開啟了兩個線程池各有15個線程。
以上代碼邏輯是
從0開始命名 如果存在則+1
如果一個文件insert沒有問題,但是此時有兩個pool
pool1 發現沒有0文件 pool2也發現沒有0文件。
pool1 mv table1/0000 to tableA/0001
pool2 mv table2/0000 to tableA/0001
此時產生了沖突,這種在多線池情況下沒法避免。。。所以hive保留這個bug,告訴我們就是重試吧。
那我們怎么解決呢?還記得上面的配置項嗎?
1.設置線程數為1,線程數越少,產生bug的幾率就越小 。
set hive.mv.files.thread=1
2.insert into 非分區表時 串行執行
3.改為分區表,此時不在一個目錄了,就不會有這個問題了。
4.union 將多個操作合并為一個操作

居然這么多人點贊,我又想到一個問題,我舉例的是非分區表,那么分區表會不會也出現這個問題呢?答案是可能的,和上面的案例同理,同時向一個分區插入數據也會報錯,解決還是如上
1.設置線程數set hive.mv.files.thread=1
2.串行執行 多個插入同一個分區的操作
3.union 將多個操作合并為一個操作
如果幫到你,點個贊是對我最大的支持

附上hive的errorMsg

public enum ErrorMsg {// The error codes are Hive-specific and partitioned into the following ranges:// 10000 to 19999: Errors occurring during semantic analysis and compilation of the query.// 20000 to 29999: Runtime errors where Hive believes that retries are unlikely to succeed.// 30000 to 39999: Runtime errors which Hive thinks may be transient and retrying may succeed.// 40000 to 49999: Errors where Hive is unable to advise about retries.// In addition to the error code, ErrorMsg also has a SQLState field.// SQLStates are taken from Section 22.1 of ISO-9075.// See http://www.contrib.andrew.cmu.edu/~shadow/sql/sql1992.txt// Most will just rollup to the generic syntax error state of 42000, but// specific errors can override the that state.// See this page for how MySQL uses SQLState codes:// http://dev.mysql.com/doc/refman/5.0/en/connector-j-reference-error-sqlstates.htmlGENERIC_ERROR(40000, "Exception while processing"),//========================== 10000 range starts here ========================//INVALID_TABLE(10001, "Table not found", "42S02"),INVALID_COLUMN(10002, "Invalid column reference"),INVALID_TABLE_OR_COLUMN(10004, "Invalid table alias or column reference"),AMBIGUOUS_TABLE_OR_COLUMN(10005, "Ambiguous table alias or column reference"),INVALID_PARTITION(10006, "Partition not found"),AMBIGUOUS_COLUMN(10007, "Ambiguous column reference"),AMBIGUOUS_TABLE_ALIAS(10008, "Ambiguous table alias"),INVALID_TABLE_ALIAS(10009, "Invalid table alias"),NO_TABLE_ALIAS(10010, "No table alias"),INVALID_FUNCTION(10011, "Invalid function"),INVALID_FUNCTION_SIGNATURE(10012, "Function argument type mismatch"),INVALID_OPERATOR_SIGNATURE(10013, "Operator argument type mismatch"),INVALID_ARGUMENT(10014, "Wrong arguments"),INVALID_ARGUMENT_LENGTH(10015, "Arguments length mismatch", "21000"),INVALID_ARGUMENT_TYPE(10016, "Argument type mismatch"),@Deprecated INVALID_JOIN_CONDITION_1(10017, "Both left and right aliases encountered in JOIN"),@Deprecated INVALID_JOIN_CONDITION_2(10018, "Neither left nor right aliases encountered in JOIN"),@Deprecated INVALID_JOIN_CONDITION_3(10019, "OR not supported in JOIN currently"),INVALID_TRANSFORM(10020, "TRANSFORM with other SELECT columns not supported"),UNSUPPORTED_MULTIPLE_DISTINCTS(10022, "DISTINCT on different columns not supported" +" with skew in data"),NO_SUBQUERY_ALIAS(10023, "No alias for subquery"),NO_INSERT_INSUBQUERY(10024, "Cannot insert in a subquery. Inserting to table "),NON_KEY_EXPR_IN_GROUPBY(10025, "Expression not in GROUP BY key"),INVALID_XPATH(10026, "General . and [] operators are not supported"),INVALID_PATH(10027, "Invalid path"),ILLEGAL_PATH(10028, "Path is not legal"),INVALID_NUMERICAL_CONSTANT(10029, "Invalid numerical constant"),INVALID_ARRAYINDEX_TYPE(10030,"Not proper type for index of ARRAY. Currently, only integer type is supported"),INVALID_MAPINDEX_CONSTANT(10031, "Non-constant expression for map indexes not supported"),INVALID_MAPINDEX_TYPE(10032, "MAP key type does not match index expression type"),NON_COLLECTION_TYPE(10033, "[] not valid on non-collection types"),SELECT_DISTINCT_WITH_GROUPBY(10034, "SELECT DISTINCT and GROUP BY can not be in the same query"),COLUMN_REPEATED_IN_PARTITIONING_COLS(10035, "Column repeated in partitioning columns"),DUPLICATE_COLUMN_NAMES(10036, "Duplicate column name:"),INVALID_BUCKET_NUMBER(10037, "Bucket number should be bigger than zero"),COLUMN_REPEATED_IN_CLUSTER_SORT(10038, "Same column cannot appear in CLUSTER BY and SORT BY"),SAMPLE_RESTRICTION(10039, "Cannot SAMPLE on more than two columns"),SAMPLE_COLUMN_NOT_FOUND(10040, "SAMPLE column not found"),NO_PARTITION_PREDICATE(10041, "No partition predicate found"),INVALID_DOT(10042, ". Operator is only supported on struct or list of struct types"),INVALID_TBL_DDL_SERDE(10043, "Either list of columns or a custom serializer should be specified"),TARGET_TABLE_COLUMN_MISMATCH(10044,"Cannot insert into target table because column number/types are different"),TABLE_ALIAS_NOT_ALLOWED(10045, "Table alias not allowed in sampling clause"),CLUSTERBY_DISTRIBUTEBY_CONFLICT(10046, "Cannot have both CLUSTER BY and DISTRIBUTE BY clauses"),ORDERBY_DISTRIBUTEBY_CONFLICT(10047, "Cannot have both ORDER BY and DISTRIBUTE BY clauses"),CLUSTERBY_SORTBY_CONFLICT(10048, "Cannot have both CLUSTER BY and SORT BY clauses"),ORDERBY_SORTBY_CONFLICT(10049, "Cannot have both ORDER BY and SORT BY clauses"),CLUSTERBY_ORDERBY_CONFLICT(10050, "Cannot have both CLUSTER BY and ORDER BY clauses"),NO_LIMIT_WITH_ORDERBY(10051, "In strict mode, if ORDER BY is specified, "+ "LIMIT must also be specified"),UNION_NOTIN_SUBQ(10053, "Top level UNION is not supported currently; "+ "use a subquery for the UNION"),INVALID_INPUT_FORMAT_TYPE(10054, "Input format must implement InputFormat"),INVALID_OUTPUT_FORMAT_TYPE(10055, "Output Format must implement HiveOutputFormat, "+ "otherwise it should be either IgnoreKeyTextOutputFormat or SequenceFileOutputFormat"),NO_VALID_PARTN(10056, HiveConf.StrictChecks.NO_PARTITIONLESS_MSG),NO_OUTER_MAPJOIN(10057, "MAPJOIN cannot be performed with OUTER JOIN"),INVALID_MAPJOIN_HINT(10058, "All tables are specified as map-table for join"),INVALID_MAPJOIN_TABLE(10059, "Result of a union cannot be a map table"),NON_BUCKETED_TABLE(10060, "Sampling expression needed for non-bucketed table"),BUCKETED_NUMERATOR_BIGGER_DENOMINATOR(10061, "Numerator should not be bigger than "+ "denominator in sample clause for table"),NEED_PARTITION_ERROR(10062, "Need to specify partition columns because the destination "+ "table is partitioned"),CTAS_CTLT_COEXISTENCE(10063, "Create table command does not allow LIKE and AS-SELECT in "+ "the same command"),LINES_TERMINATED_BY_NON_NEWLINE(10064, "LINES TERMINATED BY only supports "+ "newline '\\n' right now"),CTAS_COLLST_COEXISTENCE(10065, "CREATE TABLE AS SELECT command cannot specify "+ "the list of columns "+ "for the target table"),CTLT_COLLST_COEXISTENCE(10066, "CREATE TABLE LIKE command cannot specify the list of columns for "+ "the target table"),INVALID_SELECT_SCHEMA(10067, "Cannot derive schema from the select-clause"),CTAS_PARCOL_COEXISTENCE(10068, "CREATE-TABLE-AS-SELECT does not support "+ "partitioning in the target table "),CTAS_MULTI_LOADFILE(10069, "CREATE-TABLE-AS-SELECT results in multiple file load"),CTAS_EXTTBL_COEXISTENCE(10070, "CREATE-TABLE-AS-SELECT cannot create external table"),INSERT_EXTERNAL_TABLE(10071, "Inserting into a external table is not allowed"),DATABASE_NOT_EXISTS(10072, "Database does not exist:"),TABLE_ALREADY_EXISTS(10073, "Table already exists:", "42S02"),COLUMN_ALIAS_ALREADY_EXISTS(10074, "Column alias already exists:", "42S02"),UDTF_MULTIPLE_EXPR(10075, "Only a single expression in the SELECT clause is "+ "supported with UDTF's"),@Deprecated UDTF_REQUIRE_AS(10076, "UDTF's require an AS clause"),UDTF_NO_GROUP_BY(10077, "GROUP BY is not supported with a UDTF in the SELECT clause"),UDTF_NO_SORT_BY(10078, "SORT BY is not supported with a UDTF in the SELECT clause"),UDTF_NO_CLUSTER_BY(10079, "CLUSTER BY is not supported with a UDTF in the SELECT clause"),UDTF_NO_DISTRIBUTE_BY(10080, "DISTRUBTE BY is not supported with a UDTF in the SELECT clause"),UDTF_INVALID_LOCATION(10081, "UDTF's are not supported outside the SELECT clause, nor nested "+ "in expressions"),UDTF_LATERAL_VIEW(10082, "UDTF's cannot be in a select expression when there is a lateral view"),UDTF_ALIAS_MISMATCH(10083, "The number of aliases supplied in the AS clause does not match the "+ "number of columns output by the UDTF"),UDF_STATEFUL_INVALID_LOCATION(10084, "Stateful UDF's can only be invoked in the SELECT list"),LATERAL_VIEW_WITH_JOIN(10085, "JOIN with a LATERAL VIEW is not supported"),LATERAL_VIEW_INVALID_CHILD(10086, "LATERAL VIEW AST with invalid child"),OUTPUT_SPECIFIED_MULTIPLE_TIMES(10087, "The same output cannot be present multiple times: "),INVALID_AS(10088, "AS clause has an invalid number of aliases"),VIEW_COL_MISMATCH(10089, "The number of columns produced by the SELECT clause does not match the "+ "number of column names specified by CREATE VIEW"),DML_AGAINST_VIEW(10090, "A view cannot be used as target table for LOAD or INSERT"),ANALYZE_VIEW(10091, "ANALYZE is not supported for views"),VIEW_PARTITION_TOTAL(10092, "At least one non-partitioning column must be present in view"),VIEW_PARTITION_MISMATCH(10093, "Rightmost columns in view output do not match "+ "PARTITIONED ON clause"),PARTITION_DYN_STA_ORDER(10094, "Dynamic partition cannot be the parent of a static partition"),DYNAMIC_PARTITION_DISABLED(10095, "Dynamic partition is disabled. Either enable it by setting "+ "hive.exec.dynamic.partition=true or specify partition column values"),DYNAMIC_PARTITION_STRICT_MODE(10096, "Dynamic partition strict mode requires at least one "+ "static partition column. To turn this off set hive.exec.dynamic.partition.mode=nonstrict"),NONEXISTPARTCOL(10098, "Non-Partition column appears in the partition specification: "),UNSUPPORTED_TYPE(10099, "DATETIME type isn't supported yet. Please use "+ "DATE or TIMESTAMP instead"),CREATE_NON_NATIVE_AS(10100, "CREATE TABLE AS SELECT cannot be used for a non-native table"),LOAD_INTO_NON_NATIVE(10101, "A non-native table cannot be used as target for LOAD"),LOCKMGR_NOT_SPECIFIED(10102, "Lock manager not specified correctly, set hive.lock.manager"),LOCKMGR_NOT_INITIALIZED(10103, "Lock manager could not be initialized, check hive.lock.manager "),LOCK_CANNOT_BE_ACQUIRED(10104, "Locks on the underlying objects cannot be acquired. "+ "retry after some time"),ZOOKEEPER_CLIENT_COULD_NOT_BE_INITIALIZED(10105, "Check hive.zookeeper.quorum "+ "and hive.zookeeper.client.port"),OVERWRITE_ARCHIVED_PART(10106, "Cannot overwrite an archived partition. " +"Unarchive before running this command"),ARCHIVE_METHODS_DISABLED(10107, "Archiving methods are currently disabled. " +"Please see the Hive wiki for more information about enabling archiving"),ARCHIVE_ON_MULI_PARTS(10108, "ARCHIVE can only be run on a single partition"),UNARCHIVE_ON_MULI_PARTS(10109, "ARCHIVE can only be run on a single partition"),ARCHIVE_ON_TABLE(10110, "ARCHIVE can only be run on partitions"),RESERVED_PART_VAL(10111, "Partition value contains a reserved substring"),OFFLINE_TABLE_OR_PARTITION(10113, "Query against an offline table or partition"),NEED_PARTITION_SPECIFICATION(10115, "Table is partitioned and partition specification is needed"),INVALID_METADATA(10116, "The metadata file could not be parsed "),NEED_TABLE_SPECIFICATION(10117, "Table name could be determined; It should be specified "),PARTITION_EXISTS(10118, "Partition already exists"),TABLE_DATA_EXISTS(10119, "Table exists and contains data files"),INCOMPATIBLE_SCHEMA(10120, "The existing table is not compatible with the import spec. "),EXIM_FOR_NON_NATIVE(10121, "Export/Import cannot be done for a non-native table. "),INSERT_INTO_BUCKETIZED_TABLE(10122, "Bucketized tables do not support INSERT INTO:"),PARTSPEC_DIFFER_FROM_SCHEMA(10125, "Partition columns in partition specification are "+ "not the same as that defined in the table schema. "+ "The names and orders have to be exactly the same."),PARTITION_COLUMN_NON_PRIMITIVE(10126, "Partition column must be of primitive type."),INSERT_INTO_DYNAMICPARTITION_IFNOTEXISTS(10127,"Dynamic partitions do not support IF NOT EXISTS. Specified partitions with value :"),UDAF_INVALID_LOCATION(10128, "Not yet supported place for UDAF"),DROP_PARTITION_NON_STRING_PARTCOLS_NONEQUALITY(10129,"Drop partitions for a non-string partition column is only allowed using equality"),ALTER_COMMAND_FOR_VIEWS(10131, "To alter a view you need to use the ALTER VIEW command."),ALTER_COMMAND_FOR_TABLES(10132, "To alter a base table you need to use the ALTER TABLE command."),ALTER_VIEW_DISALLOWED_OP(10133, "Cannot use this form of ALTER on a view"),ALTER_TABLE_NON_NATIVE(10134, "ALTER TABLE can only be used for " + AlterTableTypes.nonNativeTableAllowedTypes + " to a non-native table "),SORTMERGE_MAPJOIN_FAILED(10135,"Sort merge bucketed join could not be performed. " +"If you really want to perform the operation, either set " +"hive.optimize.bucketmapjoin.sortedmerge=false, or set " +"hive.enforce.sortmergebucketmapjoin=false."),BUCKET_MAPJOIN_NOT_POSSIBLE(10136,"Bucketed mapjoin cannot be performed. " +"This can be due to multiple reasons: " +" . Join columns dont match bucketed columns. " +" . Number of buckets are not a multiple of each other. " +"If you really want to perform the operation, either remove the " +"mapjoin hint from your query or set hive.enforce.bucketmapjoin to false."),BUCKETED_TABLE_METADATA_INCORRECT(10141,"Bucketed table metadata is not correct. " +"Fix the metadata or don't use bucketed mapjoin, by setting " +"hive.enforce.bucketmapjoin to false."),JOINNODE_OUTERJOIN_MORETHAN_16(10142, "Single join node containing outer join(s) " +"cannot have more than 16 aliases"),INVALID_JDO_FILTER_EXPRESSION(10143, "Invalid expression for JDO filter"),ALTER_BUCKETNUM_NONBUCKETIZED_TBL(10145, "Table is not bucketized."),TRUNCATE_FOR_NON_MANAGED_TABLE(10146, "Cannot truncate non-managed table {0}.", true),TRUNCATE_FOR_NON_NATIVE_TABLE(10147, "Cannot truncate non-native table {0}.", true),PARTSPEC_FOR_NON_PARTITIONED_TABLE(10148, "Partition spec for non partitioned table {0}.", true),INVALID_TABLE_IN_ON_CLAUSE_OF_MERGE(10149, "No columns from target table ''{0}'' found in ON " +"clause ''{1}'' of MERGE statement.", true),LOAD_INTO_STORED_AS_DIR(10195, "A stored-as-directories table cannot be used as target for LOAD"),ALTER_TBL_STOREDASDIR_NOT_SKEWED(10196, "This operation is only valid on skewed table."),ALTER_TBL_SKEWED_LOC_NO_LOC(10197, "Alter table skewed location doesn't have locations."),ALTER_TBL_SKEWED_LOC_NO_MAP(10198, "Alter table skewed location doesn't have location map."),SKEWED_TABLE_NO_COLUMN_NAME(10200, "No skewed column name."),SKEWED_TABLE_NO_COLUMN_VALUE(10201, "No skewed values."),SKEWED_TABLE_DUPLICATE_COLUMN_NAMES(10202,"Duplicate skewed column name:"),SKEWED_TABLE_INVALID_COLUMN(10203,"Invalid skewed column name:"),SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_1(10204,"Skewed column name is empty but skewed value is not."),SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_2(10205,"Skewed column value is empty but skewed name is not."),SKEWED_TABLE_SKEWED_COL_NAME_VALUE_MISMATCH_3(10206,"The number of skewed column names and the number of " +"skewed column values are different: "),ALTER_TABLE_NOT_ALLOWED_RENAME_SKEWED_COLUMN(10207," is a skewed column. It's not allowed to rename skewed column"+ " or change skewed column type."),HIVE_GROUPING_SETS_AGGR_NOMAPAGGR(10209,"Grouping sets aggregations (with rollups or cubes) are not allowed if map-side " +" aggregation is turned off. Set hive.map.aggr=true if you want to use grouping sets"),HIVE_GROUPING_SETS_AGGR_EXPRESSION_INVALID(10210,"Grouping sets aggregations (with rollups or cubes) are not allowed if aggregation function " +"parameters overlap with the aggregation functions columns"),HIVE_GROUPING_SETS_EMPTY(10211,"Empty grouping sets not allowed"),HIVE_UNION_REMOVE_OPTIMIZATION_NEEDS_SUBDIRECTORIES(10212,"In order to use hive.optimize.union.remove, the hadoop version that you are using " +"should support sub-directories for tables/partitions. If that is true, set " +"hive.hadoop.supports.subdirectories to true. Otherwise, set hive.optimize.union.remove " +"to false"),HIVE_GROUPING_SETS_EXPR_NOT_IN_GROUPBY(10213,"Grouping sets expression is not in GROUP BY key"),INVALID_PARTITION_SPEC(10214, "Invalid partition spec specified"),ALTER_TBL_UNSET_NON_EXIST_PROPERTY(10215,"Please use the following syntax if not sure " +"whether the property existed or not:\n" +"ALTER TABLE tableName UNSET TBLPROPERTIES IF EXISTS (key1, key2, ...)\n"),ALTER_VIEW_AS_SELECT_NOT_EXIST(10216,"Cannot ALTER VIEW AS SELECT if view currently does not exist\n"),REPLACE_VIEW_WITH_PARTITION(10217,"Cannot replace a view with CREATE VIEW or REPLACE VIEW or " +"ALTER VIEW AS SELECT if the view has partitions\n"),EXISTING_TABLE_IS_NOT_VIEW(10218,"Existing table is not a view\n"),NO_SUPPORTED_ORDERBY_ALLCOLREF_POS(10219,"Position in ORDER BY is not supported when using SELECT *"),INVALID_POSITION_ALIAS_IN_GROUPBY(10220,"Invalid position alias in Group By\n"),INVALID_POSITION_ALIAS_IN_ORDERBY(10221,"Invalid position alias in Order By\n"),HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_SKEW(10225,"An additional MR job is introduced since the number of rows created per input row " +"due to grouping sets is more than hive.new.job.grouping.set.cardinality. There is no need " +"to handle skew separately. set hive.groupby.skewindata to false."),HIVE_GROUPING_SETS_THRESHOLD_NOT_ALLOWED_WITH_DISTINCTS(10226,"An additional MR job is introduced since the cardinality of grouping sets " +"is more than hive.new.job.grouping.set.cardinality. This functionality is not supported " +"with distincts. Either set hive.new.job.grouping.set.cardinality to a high number " +"(higher than the number of rows per input row due to grouping sets in the query), or " +"rewrite the query to not use distincts."),OPERATOR_NOT_ALLOWED_WITH_MAPJOIN(10227,"Not all clauses are supported with mapjoin hint. Please remove mapjoin hint."),ANALYZE_TABLE_NOSCAN_NON_NATIVE(10228, "ANALYZE TABLE NOSCAN cannot be used for " + "a non-native table"),PARTITION_VALUE_NOT_CONTINUOUS(10234, "Partition values specified are not continuous." +" A subpartition value is specified without specifying the parent partition's value"),TABLES_INCOMPATIBLE_SCHEMAS(10235, "Tables have incompatible schemas and their partitions " +" cannot be exchanged."),EXCHANGE_PARTITION_NOT_ALLOWED_WITH_TRANSACTIONAL_TABLES(10236, "Exchange partition is not allowed with "+ "transactional tables. Alternatively, shall use load data or insert overwrite to move partitions."),TRUNCATE_COLUMN_NOT_RC(10237, "Only RCFileFormat supports column truncation."),TRUNCATE_COLUMN_ARCHIVED(10238, "Column truncation cannot be performed on archived partitions."),TRUNCATE_BUCKETED_COLUMN(10239,"A column on which a partition/table is bucketed cannot be truncated."),TRUNCATE_LIST_BUCKETED_COLUMN(10240,"A column on which a partition/table is list bucketed cannot be truncated."),TABLE_NOT_PARTITIONED(10241, "Table {0} is not a partitioned table", true),DATABSAE_ALREADY_EXISTS(10242, "Database {0} already exists", true),CANNOT_REPLACE_COLUMNS(10243, "Replace columns is not supported for table {0}. SerDe may be incompatible.", true),BAD_LOCATION_VALUE(10244, "{0}  is not absolute.  Please specify a complete absolute uri."),UNSUPPORTED_ALTER_TBL_OP(10245, "{0} alter table options is not supported"),INVALID_BIGTABLE_MAPJOIN(10246, "{0} table chosen for streaming is not valid", true),MISSING_OVER_CLAUSE(10247, "Missing over clause for function : "),PARTITION_SPEC_TYPE_MISMATCH(10248, "Cannot add partition column {0} of type {1} as it cannot be converted to type {2}", true),UNSUPPORTED_SUBQUERY_EXPRESSION(10249, "Unsupported SubQuery Expression"),INVALID_SUBQUERY_EXPRESSION(10250, "Invalid SubQuery expression"),INVALID_HDFS_URI(10251, "{0} is not a hdfs uri", true),INVALID_DIR(10252, "{0} is not a directory", true),NO_VALID_LOCATIONS(10253, "Could not find any valid location to place the jars. " +"Please update hive.jar.directory or hive.user.install.directory with a valid location", false),UNSUPPORTED_AUTHORIZATION_PRINCIPAL_TYPE_GROUP(10254,"Principal type GROUP is not supported in this authorization setting", "28000"),INVALID_TABLE_NAME(10255, "Invalid table name {0}", true),INSERT_INTO_IMMUTABLE_TABLE(10256, "Inserting into a non-empty immutable table is not allowed"),UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_GLOBAL(10257,"Resource type GLOBAL is not supported in this authorization setting", "28000"),UNSUPPORTED_AUTHORIZATION_RESOURCE_TYPE_COLUMN(10258,"Resource type COLUMN is not supported in this authorization setting", "28000"),TXNMGR_NOT_SPECIFIED(10260, "Transaction manager not specified correctly, " +"set hive.txn.manager"),TXNMGR_NOT_INSTANTIATED(10261, "Transaction manager could not be " +"instantiated, check hive.txn.manager"),TXN_NO_SUCH_TRANSACTION(10262, "No record of transaction {0} could be found, " +"may have timed out", true),TXN_ABORTED(10263, "Transaction manager has aborted the transaction {0}.  Reason: {1}", true),DBTXNMGR_REQUIRES_CONCURRENCY(10264,"To use DbTxnManager you must set hive.support.concurrency=true"),TXNMGR_NOT_ACID(10265, "This command is not allowed on an ACID table {0}.{1} with a non-ACID transaction manager", true),LOCK_NO_SUCH_LOCK(10270, "No record of lock {0} could be found, " +"may have timed out", true),LOCK_REQUEST_UNSUPPORTED(10271, "Current transaction manager does not " +"support explicit lock requests.  Transaction manager:  "),METASTORE_COMMUNICATION_FAILED(10280, "Error communicating with the " +"metastore"),METASTORE_COULD_NOT_INITIATE(10281, "Unable to initiate connection to the " +"metastore."),INVALID_COMPACTION_TYPE(10282, "Invalid compaction type, supported values are 'major' and " +"'minor'"),NO_COMPACTION_PARTITION(10283, "You must specify a partition to compact for partitioned tables"),TOO_MANY_COMPACTION_PARTITIONS(10284, "Compaction can only be requested on one partition at a " +"time."),DISTINCT_NOT_SUPPORTED(10285, "Distinct keyword is not support in current context"),NONACID_COMPACTION_NOT_SUPPORTED(10286, "Compaction is not allowed on non-ACID table {0}.{1}", true),MASKING_FILTERING_ON_ACID_NOT_SUPPORTED(10287,"Detected {0}.{1} has row masking/column filtering enabled, " +"which is not supported for query involving ACID operations", true),UPDATEDELETE_PARSE_ERROR(10290, "Encountered parse error while parsing rewritten merge/update or " +"delete query"),UPDATEDELETE_IO_ERROR(10291, "Encountered I/O error while parsing rewritten update or " +"delete query"),UPDATE_CANNOT_UPDATE_PART_VALUE(10292, "Updating values of partition columns is not supported"),INSERT_CANNOT_CREATE_TEMP_FILE(10293, "Unable to create temp file for insert values "),ACID_OP_ON_NONACID_TXNMGR(10294, "Attempt to do update or delete using transaction manager that" +" does not support these operations."),VALUES_TABLE_CONSTRUCTOR_NOT_SUPPORTED(10296,"Values clause with table constructor not yet supported"),ACID_OP_ON_NONACID_TABLE(10297, "Attempt to do update or delete on table {0} that is " +"not transactional", true),ACID_NO_SORTED_BUCKETS(10298, "ACID insert, update, delete not supported on tables that are " +"sorted, table {0}", true),ALTER_TABLE_TYPE_PARTIAL_PARTITION_SPEC_NO_SUPPORTED(10299,"Alter table partition type {0} does not allow partial partition spec", true),ALTER_TABLE_PARTITION_CASCADE_NOT_SUPPORTED(10300,"Alter table partition type {0} does not support cascade", true),DROP_NATIVE_FUNCTION(10301, "Cannot drop native function"),UPDATE_CANNOT_UPDATE_BUCKET_VALUE(10302, "Updating values of bucketing columns is not supported.  Column {0}.", true),IMPORT_INTO_STRICT_REPL_TABLE(10303,"Non-repl import disallowed against table that is a destination of replication."),CTAS_LOCATION_NONEMPTY(10304, "CREATE-TABLE-AS-SELECT cannot create table with location to a non-empty directory."),CTAS_CREATES_VOID_TYPE(10305, "CREATE-TABLE-AS-SELECT creates a VOID type, please use CAST to specify the type, near field: "),TBL_SORTED_NOT_BUCKETED(10306, "Destination table {0} found to be sorted but not bucketed.", true),//{2} should be lockidLOCK_ACQUIRE_TIMEDOUT(10307, "Lock acquisition for {0} timed out after {1}ms.  {2}", true),COMPILE_LOCK_TIMED_OUT(10308, "Attempt to acquire compile lock timed out.", true),CANNOT_CHANGE_SERDE(10309, "Changing SerDe (from {0}) is not supported for table {1}. File format may be incompatible", true),CANNOT_CHANGE_FILEFORMAT(10310, "Changing file format (from {0}) is not supported for table {1}", true),CANNOT_REORDER_COLUMNS(10311, "Reordering columns is not supported for table {0}. SerDe may be incompatible", true),CANNOT_CHANGE_COLUMN_TYPE(10312, "Changing from type {0} to {1} is not supported for column {2}. SerDe may be incompatible", true),REPLACE_CANNOT_DROP_COLUMNS(10313, "Replacing columns cannot drop columns for table {0}. SerDe may be incompatible", true),REPLACE_UNSUPPORTED_TYPE_CONVERSION(10314, "Replacing columns with unsupported type conversion (from {0} to {1}) for column {2}. SerDe may be incompatible", true),HIVE_GROUPING_SETS_AGGR_NOMAPAGGR_MULTIGBY(10315,"Grouping sets aggregations (with rollups or cubes) are not allowed when " +"HIVEMULTIGROUPBYSINGLEREDUCER is turned on. Set hive.multigroupby.singlereducer=false if you want to use grouping sets"),CANNOT_RETRIEVE_TABLE_METADATA(10316, "Error while retrieving table metadata"),INVALID_AST_TREE(10318, "Internal error : Invalid AST"),ERROR_SERIALIZE_METASTORE(10319, "Error while serializing the metastore objects"),IO_ERROR(10320, "Error while performing IO operation "),ERROR_SERIALIZE_METADATA(10321, "Error while serializing the metadata"),INVALID_LOAD_TABLE_FILE_WORK(10322, "Invalid Load Table Work or Load File Work"),CLASSPATH_ERROR(10323, "Classpath error"),IMPORT_SEMANTIC_ERROR(10324, "Import Semantic Analyzer Error"),INVALID_FK_SYNTAX(10325, "Invalid Foreign Key syntax"),INVALID_CSTR_SYNTAX(10326, "Invalid Constraint syntax"),ACID_NOT_ENOUGH_HISTORY(10327, "Not enough history available for ({0},{1}).  " +"Oldest available base: {2}", true),INVALID_COLUMN_NAME(10328, "Invalid column name"),UNSUPPORTED_SET_OPERATOR(10329, "Unsupported set operator"),LOCK_ACQUIRE_CANCELLED(10330, "Query was cancelled while acquiring locks on the underlying objects. "),NOT_RECOGNIZED_CONSTRAINT(10331, "Constraint not recognized"),INVALID_CONSTRAINT(10332, "Invalid constraint definition"),REPLACE_VIEW_WITH_MATERIALIZED(10400, "Attempt to replace view {0} with materialized view", true),REPLACE_MATERIALIZED_WITH_VIEW(10401, "Attempt to replace materialized view {0} with view", true),UPDATE_DELETE_VIEW(10402, "You cannot update or delete records in a view"),MATERIALIZED_VIEW_DEF_EMPTY(10403, "Query for the materialized view rebuild could not be retrieved"),MERGE_PREDIACTE_REQUIRED(10404, "MERGE statement with both UPDATE and DELETE clauses " +"requires \"AND <boolean>\" on the 1st WHEN MATCHED clause of <{0}>", true),MERGE_TOO_MANY_DELETE(10405, "MERGE statment can have at most 1 WHEN MATCHED ... DELETE clause: <{0}>", true),MERGE_TOO_MANY_UPDATE(10406, "MERGE statment can have at most 1 WHEN MATCHED ... UPDATE clause: <{0}>", true),INVALID_JOIN_CONDITION(10407, "Error parsing condition in join"),INVALID_TARGET_COLUMN_IN_SET_CLAUSE(10408, "Target column \"{0}\" of set clause is not found in table \"{1}\".", true),HIVE_GROUPING_FUNCTION_EXPR_NOT_IN_GROUPBY(10409, "Expression in GROUPING function not present in GROUP BY"),ALTER_TABLE_NON_PARTITIONED_TABLE_CASCADE_NOT_SUPPORTED(10410,"Alter table with non-partitioned table does not support cascade"),HIVE_GROUPING_SETS_SIZE_LIMIT(10411,"Grouping sets size cannot be greater than 64"),REBUILD_NO_MATERIALIZED_VIEW(10412, "Rebuild command only valid for materialized views"),LOAD_DATA_ACID_FILE(10413,"\"{0}\" was created created by Acid write - it cannot be loaded into anther Acid table",true),ACID_OP_ON_INSERTONLYTRAN_TABLE(10414, "Attempt to do update or delete on table {0} that is " +"insert-only transactional", true),LOAD_DATA_LAUNCH_JOB_IO_ERROR(10415, "Encountered I/O error while parsing rewritten load data into insert query"),LOAD_DATA_LAUNCH_JOB_PARSE_ERROR(10416, "Encountered parse error while parsing rewritten load data into insert query"),//========================== 20000 range starts here ========================//SCRIPT_INIT_ERROR(20000, "Unable to initialize custom script."),SCRIPT_IO_ERROR(20001, "An error occurred while reading or writing to your custom script. "+ "It may have crashed with an error."),SCRIPT_GENERIC_ERROR(20002, "Hive encountered some unknown error while "+ "running your custom script."),SCRIPT_CLOSING_ERROR(20003, "An error occurred when trying to close the Operator " +"running your custom script."),DYNAMIC_PARTITIONS_TOO_MANY_PER_NODE_ERROR(20004, "Fatal error occurred when node " +"tried to create too many dynamic partitions. The maximum number of dynamic partitions " +"is controlled by hive.exec.max.dynamic.partitions and hive.exec.max.dynamic.partitions.pernode. "),/*** {1} is the transaction id;* use {@link org.apache.hadoop.hive.common.JavaUtils#txnIdToString(long)} to format*/OP_NOT_ALLOWED_IN_IMPLICIT_TXN(20006, "Operation {0} is not allowed in an implicit transaction ({1}).", true),/*** {1} is the transaction id;* use {@link org.apache.hadoop.hive.common.JavaUtils#txnIdToString(long)} to format*/OP_NOT_ALLOWED_IN_TXN(20007, "Operation {0} is not allowed in a transaction ({1},queryId={2}).", true),OP_NOT_ALLOWED_WITHOUT_TXN(20008, "Operation {0} is not allowed without an active transaction", true),ACCESS_DENIED(20009, "Access denied: {0}", "42000", true),QUOTA_EXCEEDED(20010, "Quota exceeded: {0}", "64000", true),UNRESOLVED_PATH(20011, "Unresolved path: {0}", "64000", true),FILE_NOT_FOUND(20012, "File not found: {0}", "64000", true),WRONG_FILE_FORMAT(20013, "Wrong file format. Please check the file's format.", "64000", true),SPARK_CREATE_CLIENT_INVALID_QUEUE(20014, "Spark job is submitted to an invalid queue: {0}."+ " Please fix and try again.", true),SPARK_RUNTIME_OOM(20015, "Spark job failed because of out of memory."),//if the error message is changed for REPL_EVENTS_MISSING_IN_METASTORE, then need modification in getNextNotification//method in HiveMetaStoreClientREPL_EVENTS_MISSING_IN_METASTORE(20016, "Notification events are missing in the meta store."),REPL_BOOTSTRAP_LOAD_PATH_NOT_VALID(20017, "Target database is bootstrapped from some other path."),REPL_FILE_MISSING_FROM_SRC_AND_CM_PATH(20018, "File is missing from both source and cm path."),REPL_LOAD_PATH_NOT_FOUND(20019, "Load path does not exist."),REPL_DATABASE_IS_NOT_SOURCE_OF_REPLICATION(20020,"Source of replication (repl.source.for) is not set in the database properties."),// An exception from runtime that will show the full stack to clientUNRESOLVED_RT_EXCEPTION(29999, "Runtime Error: {0}", "58004", true),//========================== 30000 range starts here ========================//STATSPUBLISHER_NOT_OBTAINED(30000, "StatsPublisher cannot be obtained. " +"There was a error to retrieve the StatsPublisher, and retrying " +"might help. If you dont want the query to fail because accurate statistics " +"could not be collected, set hive.stats.reliable=false"),STATSPUBLISHER_INITIALIZATION_ERROR(30001, "StatsPublisher cannot be initialized. " +"There was a error in the initialization of StatsPublisher, and retrying " +"might help. If you dont want the query to fail because accurate statistics " +"could not be collected, set hive.stats.reliable=false"),STATSPUBLISHER_CONNECTION_ERROR(30002, "StatsPublisher cannot be connected to." +"There was a error while connecting to the StatsPublisher, and retrying " +"might help. If you dont want the query to fail because accurate statistics " +"could not be collected, set hive.stats.reliable=false"),STATSPUBLISHER_PUBLISHING_ERROR(30003, "Error in publishing stats. There was an " +"error in publishing stats via StatsPublisher, and retrying " +"might help. If you dont want the query to fail because accurate statistics " +"could not be collected, set hive.stats.reliable=false"),STATSPUBLISHER_CLOSING_ERROR(30004, "StatsPublisher cannot be closed." +"There was a error while closing the StatsPublisher, and retrying " +"might help. If you dont want the query to fail because accurate statistics " +"could not be collected, set hive.stats.reliable=false"),COLUMNSTATSCOLLECTOR_INVALID_PART_KEY(30005, "Invalid partitioning key specified in ANALYZE " +"statement"),COLUMNSTATSCOLLECTOR_INVALID_PARTITION(30007, "Invalid partitioning key/value specified in " +"ANALYZE statement"),COLUMNSTATSCOLLECTOR_PARSE_ERROR(30009, "Encountered parse error while parsing rewritten query"),COLUMNSTATSCOLLECTOR_IO_ERROR(30010, "Encountered I/O exception while parsing rewritten query"),DROP_COMMAND_NOT_ALLOWED_FOR_PARTITION(30011, "Partition protected from being dropped"),COLUMNSTATSCOLLECTOR_INVALID_COLUMN(30012, "Column statistics are not supported "+ "for partition columns"),STATSAGGREGATOR_SOURCETASK_NULL(30014, "SourceTask of StatsTask should not be null"),STATSAGGREGATOR_CONNECTION_ERROR(30015,"Stats aggregator of type {0} cannot be connected to", true),STATS_SKIPPING_BY_ERROR(30017, "Skipping stats aggregation by error {0}", true),INVALID_FILE_FORMAT_IN_LOAD(30019, "The file that you are trying to load does not match the" +" file format of the destination table."),SCHEMA_REQUIRED_TO_READ_ACID_TABLES(30020, "Neither the configuration variables " +"schema.evolution.columns / schema.evolution.columns.types " +"nor the " +"columns / columns.types " +"are set.  Table schema information is required to read ACID tables"),ACID_TABLES_MUST_BE_READ_WITH_ACID_READER(30021, "An ORC ACID reader required to read ACID tables"),ACID_TABLES_MUST_BE_READ_WITH_HIVEINPUTFORMAT(30022, "Must use HiveInputFormat to read ACID tables " +"(set hive.input.format=org.apache.hadoop.hive.ql.io.HiveInputFormat)"),ACID_LOAD_DATA_INVALID_FILE_NAME(30023, "{0} file name is not valid in Load Data into Acid " +"table {1}.  Examples of valid names are: 00000_0, 00000_0_copy_1", true),CONCATENATE_UNSUPPORTED_FILE_FORMAT(30030, "Concatenate/Merge only supported for RCFile and ORCFile formats"),CONCATENATE_UNSUPPORTED_TABLE_BUCKETED(30031, "Concatenate/Merge can not be performed on bucketed tables"),CONCATENATE_UNSUPPORTED_PARTITION_ARCHIVED(30032, "Concatenate/Merge can not be performed on archived partitions"),CONCATENATE_UNSUPPORTED_TABLE_NON_NATIVE(30033, "Concatenate/Merge can not be performed on non-native tables"),CONCATENATE_UNSUPPORTED_TABLE_NOT_MANAGED(30034, "Concatenate/Merge can only be performed on managed tables"),CONCATENATE_UNSUPPORTED_TABLE_TRANSACTIONAL(30035,"Concatenate/Merge can not be performed on transactional tables"),SPARK_GET_JOB_INFO_TIMEOUT(30036,"Spark job timed out after {0} seconds while getting job info", true),SPARK_JOB_MONITOR_TIMEOUT(30037, "Job hasn''t been submitted after {0}s." +" Aborting it.\nPossible reasons include network issues, " +"errors in remote driver or the cluster has no available resources, etc.\n" +"Please check YARN or Spark driver''s logs for further information.\n" +"The timeout is controlled by " + HiveConf.ConfVars.SPARK_JOB_MONITOR_TIMEOUT + ".", true),// Various errors when creating Spark clientSPARK_CREATE_CLIENT_TIMEOUT(30038,"Timed out while creating Spark client for session {0}.", true),SPARK_CREATE_CLIENT_QUEUE_FULL(30039,"Failed to create Spark client because job queue is full: {0}.", true),SPARK_CREATE_CLIENT_INTERRUPTED(30040,"Interrupted while creating Spark client for session {0}", true),SPARK_CREATE_CLIENT_ERROR(30041,"Failed to create Spark client for Spark session {0}", true),SPARK_CREATE_CLIENT_INVALID_RESOURCE_REQUEST(30042,"Failed to create Spark client due to invalid resource request: {0}", true),SPARK_CREATE_CLIENT_CLOSED_SESSION(30043,"Cannot create Spark client on a closed session {0}", true),SPARK_JOB_INTERRUPTED(30044, "Spark job was interrupted while executing"),REPL_FILE_SYSTEM_OPERATION_RETRY(30045, "Replication file system operation retry expired."),//========================== 40000 range starts here ========================//SPARK_JOB_RUNTIME_ERROR(40001,"Spark job failed during runtime. Please check stacktrace for the root cause.");

本文來自互聯網用戶投稿,該文觀點僅代表作者本人,不代表本站立場。本站僅提供信息存儲空間服務,不擁有所有權,不承擔相關法律責任。
如若轉載,請注明出處:http://www.pswp.cn/news/167746.shtml
繁體地址,請注明出處:http://hk.pswp.cn/news/167746.shtml
英文地址,請注明出處:http://en.pswp.cn/news/167746.shtml

如若內容造成侵權/違法違規/事實不符,請聯系多彩編程網進行投訴反饋email:809451989@qq.com,一經查實,立即刪除!

相關文章

解決在Windows10或Windows11下無權限修改hosts文件

解決在Windows10或Windows11下無權限修改hosts文件&#xff0c;無法寫入內容 1、首先在開始菜單中找到這個 2、接著輸入&#xff1a; C:\Windows\System32\drivers\etc3、再次輸入以下命令行&#xff1a;notepad hosts &#xff0c;并回車&#xff1a; notepad hosts 4、然后…

DataFunSummit:2023年現代數據棧技術峰會-核心PPT資料下載

一、峰會簡介 現代數據棧&#xff08;Modern Data Stack&#xff09;是一種集合了多種技術和工具的軟件基礎設施&#xff0c;旨在更好地管理和處理數據&#xff0c;并為企業提供數據驅動的洞察和決策。包含以下幾個組件&#xff1a;數據采集、數據處理、數據存儲、數據查詢和分…

區塊鏈技術與應用 【全國職業院校技能大賽國賽題目解析】第四套區塊鏈應用后端開發

第四套區塊鏈應用后端開發 環境 : ubuntu20 fisco : 2.8.0 springboot 2.1.1 fisco-java-sdk: 2.7.2 maven 3.8.8 前言 這套后端樣題,只涉及調用fisco的系統接口,不涉及此食品溯源項目的業務接口,所以我就直接生成一個springboot項目進行完成此題目。 請提前準備好一…

Docker的項目資源參考

Docker的項目資源包括以下內容&#xff1a; Docker官方網站&#xff1a;https://www.docker.com/ Docker Hub&#xff1a;https://hub.docker.com/ Docker文檔&#xff1a;https://docs.docker.com/ Docker GitHub倉庫&#xff1a;https://github.com/docker Docker官方博客…

Unity中Shader的Standard材質解析(二)

文章目錄 前言一、我們對 Standard 的 PBR 的 GI 進行解析1、我們先創建一個PBR的.cginc文件&#xff0c;用于整理用到的函數2、然后在Standard的Shader中引用該cginc文件 二、依次整理函數到該cginc文件中我們來看一下PBR中GI的鏡面反射做了些什么 二、最終代碼.cginc代碼&…

OpenGL 繪制旋轉球(Qt)

文章目錄 一、簡介二、實現代碼三、實現效果一、簡介 這里其實就是指三個互相垂直的三個圓形,正好之前已經完成了圓形平面的繪制,那么這里就需要對之前的圓形進行一些改造,使得它們可以以任意一種姿態在OpenGL中進行繪制(添加變換矩陣)。 這里同樣對其進行封裝,具體內容如…

【教學類-06-07】20231124 (55格版)X-X之間的加法、減法、加減混合題

背景需求 在大四班里&#xff0c;預測試55格“5以內、10以內、20以內的加法題、減法題、加減混合題”的“實用性”。 由于只打印一份20以內加法減法混合題。 “這套20以內的加減法最難”&#xff0c;我詢問誰會做&#xff08;摸底幼兒的水平&#xff09; 有兩位男孩舉手想挑…

joplin筆記同步 到騰訊云S3

創建存儲桶 打開騰訊云的存儲桶列表&#xff0c;點擊“創建存儲桶”&#xff0c;輸入名稱&#xff0c;選擇地域&#xff08;建議選擇離自己較近的地域以降低訪問時延&#xff09;和訪問權限&#xff08;建議選擇“私有讀寫”&#xff09;。 s3 存儲桶&#xff1a; 存儲桶的名稱…

【經典小練習】簡單的文件加密解密

文章目錄 &#x1f339;什么是文件加密?應用場景 &#x1f6f8;案例&#x1f33a;描述&#x1f33a;代碼 &#x1f339;什么是文件加密 Java文件加密是指使用Java編程語言和相關的加密算法對文件進行加密處理。通過這種方式&#xff0c;可以將文件內容轉換為一種非常規的形式…

Halcon Solution Guide I basics(4): Blob Analysis(連通性解析)

文章目錄 文章專欄前言文章解析開頭步驟分析簡單案例進階方案 進階代碼案例crystal&#xff0c;結晶匹配需求分析 文章專欄 Halcon開發 Halcon學習 練習項目gitee倉庫 CSDN Major 博主Halcon文章推薦 前言 今天來看第三章內容&#xff0c;既然是零基礎&#xff0c;而且我還有大…

希寶貓罐頭怎么樣?專業人士告訴你口碑好的貓罐頭推薦

作為一個從業寵物營養師7年的人&#xff0c;可以說對于貓咪的食物很有研究和貓罐頭品牌選購上&#xff0c;我有自己的見解。那么希寶貓罐頭怎么樣呢&#xff1f; 希寶貓罐頭采用了先進的加工工藝&#xff0c;注重產品的包裝和密封性&#xff0c;其包裝設計簡潔時尚&#xff0c…

Java定時任務 ScheduledThreadPoolExecutor

ScheduledThreadPoolExecutor 的創建 ScheduledThreadPoolExecutor executorService new ScheduledThreadPoolExecutor(1, // 核心線程數new BasicThreadFactory.Builder().namingPattern("example-schedule-pool-%d") // 線程命名規則.daemon(true) // 設置線程為…

STM32 中斷系統

單片機學習 目錄 文章目錄 前言 一、中斷系統 1.1 什么是中斷 1.2 中斷優先級 1.3 中斷嵌套 1.4 C語言中的中斷程序 二、STM32的中斷通道和中斷向量 2.1 中斷通道 2.2 嵌套向量中斷控制器NVIC 2.2.1 什么是NVIC 2.2.2 NVIC基本結構 2.2.3搶占優先級和響應優先級 2.2.4 NVIC的優…

間隔分區表(DM8:達夢數據庫)

DM8:達夢數據庫 - 間隔分區表 環境介紹1 按 年 - 間隔分區表2 按 月 - 間隔分區3 按 日 - 間隔分區4 按 數值 - 間隔分區表5 達夢數據庫學習使用列表 環境介紹 間隔分區表使用說明&#xff1a; 僅支持一級范圍分區創建間隔分區。 只能有一個分區列&#xff0c;且分區列類型為…

究竟什么是阻塞與非阻塞、同步與異步

文章目錄 前言阻塞與非阻塞同步與異步復雜的網絡IO真正的異步IOIO分類與示例總結 前言 這幾個名詞在程序開發時經常聽到&#xff0c;但是突然問起來各個詞的含義一時間還真是說不清楚&#xff0c;貌似這幾個詞都是翻譯過來的&#xff0c;每個人的解釋都不太一樣&#xff0c;我…

深度學習卷積神經網絡參數計算難點重點

目錄 一、卷積層圖像輸出尺寸 二、池化層圖像輸出尺寸 三、全連接層輸出尺寸 四、卷積層參數數量 五、全連接層參數數量 六、代碼實現與驗證 以LeNet5經典模型為例子并且通道數為1 LeNet5網絡有7層&#xff1a; ? 1.第1層&#xff1a;卷積層 ? 輸入&#xff1a;原始的圖片像素…

c語言數字轉圈

數字轉圈 題干輸入整數 N&#xff08;1≤N≤9&#xff09;&#xff0c;輸出如下 N 階方陣。 若輸入5顯示如下方陣&#xff1a; * 1** 2** 3** 4** 5* *16**17**18**19** 6* *15**24**25**20** 7* *14**23**22**21** 8* *13**12**11**10** 9*輸入樣例3輸出樣例* 1*…

PTA 海盜分贓

P 個海盜偷了 D 顆鉆石后來到公海分贓&#xff0c;一致同意如下分贓策略&#xff1a; 首先&#xff0c;P 個海盜通過抽簽決定 1 - P 的序號。然后由第 1 號海盜提出一個分配方案&#xff08;方案應給出每個海盜分得的具體數量&#xff09;&#xff0c;如果能夠得到包括 1 號在…

linux高級篇基礎理論六(firewalld,防火墻類型,,區域,服務端口,富語言)

??作者&#xff1a;小劉在C站 ??個人主頁&#xff1a; 小劉主頁 ??不能因為人生的道路坎坷,就使自己的身軀變得彎曲;不能因為生活的歷程漫長,就使求索的 腳步遲緩。 ??學習兩年總結出的運維經驗&#xff0c;以及思科模擬器全套網絡實驗教程。專欄&#xff1a;云計算技…

基于戰爭策略算法優化概率神經網絡PNN的分類預測 - 附代碼

基于戰爭策略算法優化概率神經網絡PNN的分類預測 - 附代碼 文章目錄 基于戰爭策略算法優化概率神經網絡PNN的分類預測 - 附代碼1.PNN網絡概述2.變壓器故障診街系統相關背景2.1 模型建立 3.基于戰爭策略優化的PNN網絡5.測試結果6.參考文獻7.Matlab代碼 摘要&#xff1a;針對PNN神…