hadoop - How to store special characters in Hive? -


i've been playing around spark, hive , parquet, have data in hive table , here how looks ( warning french language ahead ) :

bloqu� � l'arriv�e      null probl�me de connexion   bloqu� en hub 

obviously there's wrong here.

what : read teradata table dataframe spark, store parquet file , use file store hive, here's create table script :

create table `table`(    `lib` varchar(255),    `libelle_sous_cause` varchar(255),    )  stored parquet  location    'hdfs://location'; 

i don't know cause this, might caused special encoding between teradata > parquet or parquet > hive, i'm not sure.

any appreciated, thanks.

i've run same problem when doing sqoop teradata hadoop. when extracting teradata, in select, please try wrapping varchar columns, may have issue, line:

select name, age, trim(cast(translate(columnofinterest using latin_to_unicode error) varchar(50))) tablename; 

columnofinterest column have special characters.

let me know if works.


Comments

Popular posts from this blog

php - Wordpress website dashboard page or post editor content is not showing but front end data is showing properly -

How to get the ip address of VM and use it to configure SSH connection dynamically in Ansible -

javascript - Get parameter of GET request -