Hive not loading data from HDFS to HIVE external table


10 Answer(s)


hi Menal,

It's not clear from the error. Can you share the table creation script and the data you are tying to load to Hive

This is the table creation script i used:


And my data is (tweets) stored in HDFS:


Hi Menal,
Thanks for the Table script.
You are using the partitioned variables that are not defined in the table/Tweets.
Please use the correct names.
Hope this helps.
Thanks.

I'm sorry I am still a bit confused, I tried replacing it with (%Y INT, %m INT, %d INT, %H INT) as per flume.conf and that did not work either. What else can I use to partition it?

Would I use the column names on Cloudera manager? (type, size, replication, block size, modification time, etc..)

I keep getting this error now.


Hi Menal,
As per last error context:
Copy the JsonSerde Jar into Hive lib
You can find the copy of the jar in the Twitter project under Module-15 Project on your Dezyre Dashboard.
Hope this helps.
Let me know if you get any error again.
Thanks.

I fixed the SerDe problem but I am still having the issue of exactly what my partitioned variables are. I tried both (y, m, d, h) and the variables shown in the table (name, type, size, replication, etc.) yet I keep getting the same error as I initially stated..

Hi Menal,
Open FlumeData.xxxxxxxxxxx and you will see the tag name. They are column names that you can use for partition.

Hope this helps.
Thanks.