-
Notifications
You must be signed in to change notification settings - Fork 2.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[SUPPORT]I can see the Hudi table and data in spark-shell, but in Hive, the table is visible, yet no data can be queried. #12673
Comments
@TigerTORA Did you add hive bundle for version 1.0.0 to your hive libs, or you can do add jar under hive cli too. |
yes I have add hive bundle for version 1.0.0 to my hive libs
|
Hi @TigerTORA I have successfully read data from Hive using your code. Could you please review the Hive query logs to identify any potential issues? 0: jdbc:hive2://hive-server:10000/default> use my_db;
0: jdbc:hive2://hive-server:10000/default> show tables;
+-----------+
| tab_name |
+-----------+
| hudi_cow |
+-----------+
0: jdbc:hive2://hive-server:10000/default> select * from hudi_cow;
INFO : Compiling command(queryId=hive_20250120110910_bacc7a57-aabb-467f-93b7-da01cf279020): select * from hudi_cow
INFO : No Stats for my_db@hudi_cow, Columns: _hoodie_commit_time, inttolong, longtoint, _hoodie_partition_path, versionid, precomb, _hoodie_record_key, name, tobedeletedstr, _hoodie_commit_seqno, _hoodie_file_name, rowid
INFO : Semantic Analysis Completed (retrial = false)
INFO : Created Hive schema: Schema(fieldSchemas:[FieldSchema(name:hudi_cow._hoodie_commit_time, type:string, comment:null), FieldSchema(name:hudi_cow._hoodie_commit_seqno, type:string, comment:null), FieldSchema(name:hudi_cow._hoodie_record_key, type:string, comment:null), FieldSchema(name:hudi_cow._hoodie_partition_path, type:string, comment:null), FieldSchema(name:hudi_cow._hoodie_file_name, type:string, comment:null), FieldSchema(name:hudi_cow.rowid, type:string, comment:null), FieldSchema(name:hudi_cow.precomb, type:bigint, comment:null), FieldSchema(name:hudi_cow.name, type:string, comment:null), FieldSchema(name:hudi_cow.versionid, type:string, comment:null), FieldSchema(name:hudi_cow.tobedeletedstr, type:string, comment:null), FieldSchema(name:hudi_cow.inttolong, type:int, comment:null), FieldSchema(name:hudi_cow.longtoint, type:bigint, comment:null), FieldSchema(name:hudi_cow.partitionid, type:string, comment:null)], properties:null)
INFO : Completed compiling command(queryId=hive_20250120110910_bacc7a57-aabb-467f-93b7-da01cf279020); Time taken: 0.31 seconds
INFO : Concurrency mode is disabled, not creating a lock manager
INFO : Executing command(queryId=hive_20250120110910_bacc7a57-aabb-467f-93b7-da01cf279020): select * from hudi_cow
INFO : Completed executing command(queryId=hive_20250120110910_bacc7a57-aabb-467f-93b7-da01cf279020); Time taken: 0.001 seconds
+-------------------------------+--------------------------------+------------------------------+----------------------------------+----------------------------------------------------+-----------------+-------------------+----------------+---------------------+--------------------------+---------------------+---------------------+-----------------------+
| hudi_cow._hoodie_commit_time | hudi_cow._hoodie_commit_seqno | hudi_cow._hoodie_record_key | hudi_cow._hoodie_partition_path | hudi_cow._hoodie_file_name | hudi_cow.rowid | hudi_cow.precomb | hudi_cow.name | hudi_cow.versionid | hudi_cow.tobedeletedstr | hudi_cow.inttolong | hudi_cow.longtoint | hudi_cow.partitionid |
+-------------------------------+--------------------------------+------------------------------+----------------------------------+----------------------------------------------------+-----------------+-------------------+----------------+---------------------+--------------------------+---------------------+---------------------+-----------------------+
| 20250120110832477 | 20250120110832477_1_0 | row_1 | partitionId=2021/01/01 | 8273822c-3db4-4f41-9369-52a48b289ec9-0_1-68-121_20250120110832477.parquet | row_1 | 0 | bob | v_0 | toBeDel0 | 0 | 1000000 | 2021/01/01 |
| 20250120110832477 | 20250120110832477_1_1 | row_2 | partitionId=2021/01/01 | 8273822c-3db4-4f41-9369-52a48b289ec9-0_1-68-121_20250120110832477.parquet | row_2 | 0 | john | v_0 | toBeDel0 | 0 | 1000000 | 2021/01/01 |
| 20250120110832477 | 20250120110832477_0_0 | row_3 | partitionId=2021/01/02 | 328f6651-ed04-49db-be45-9c3671bcda8d-0_0-68-120_20250120110832477.parquet | row_3 | 0 | tom | v_0 | toBeDel0 | 0 | 1000000 | 2021/01/02 |
+-------------------------------+--------------------------------+------------------------------+----------------------------------+----------------------------------------------------+-----------------+-------------------+----------------+---------------------+--------------------------+---------------------+---------------------+-----------------------+
3 rows selected (1.282 seconds)
0: jdbc:hive2://hive-server:10000/default> I have placed hudi-hadoop-mr-bundle-1.0.0-rc1.jar to the hive lib. hive@9b304fb95236:/opt/hive$ ls lib/*hudi*
lib/hudi-hadoop-mr-bundle-1.0.0-rc1.jar |
First of all, I would like to sincerely thank everyone for their help. This is my query log |
Hi @TigerTORA Could you please check the following:
|
Tips before filing an issue
Have you gone through our FAQs?
Join the mailing list to engage in conversations and get faster support at [email protected].
If you have triaged this as a bug, then file an issue directly.
Describe the problem you faced
I can see the Hudi table and data in spark-shell, but in Hive, the table is visible, yet no data can be queried.
To Reproduce
Steps to reproduce the behavior:
show create table
in hiveExpected behavior
Environment Description
Hudi version :1.0.0
Spark version :3.4.1.7.3.1.0-197
Hive version :3.1.3000.7.3.1.0-197
Hadoop version :3.1.1.7.3.1.0-197
Storage (HDFS/S3/GCS..) : HDFS
Running on Docker? (yes/no) :no
Additional context
Add any other context about the problem here.
Stacktrace
Add the stacktrace of the error.
The text was updated successfully, but these errors were encountered: