How do I load a big csv file into a partitioned table?
Answer / Mahendra Kumar Meena
To load a big CSV file into a partitioned table in Impala, you can use the LOAD DATA INPATH command. Here's an example:
CREATE EXTERNAL TABLE mytable (...)
PARTITIONED BY (partition_column1 STRING, partition_column2 INT)
LOCATION 'hdfs://namenode/path/to/data';
LOAD DATA INPATH 'hdfs://namenode/path/to/bigcsvfile.csv'
INTO TABLE mytable
PARTITION (partition_column1='part1', partition_column2=1);
| Is This Answer Correct ? | 0 Yes | 0 No |
How is impala metadata managed?
Describe impala shell (impala-shell command)?
Is there a dual table?
How does impala achieve its performance improvements?
State some advantages of impala?
State some impala hadoop benefits?
What is the maximum number of rows in a table?
Are results returned as they become available, or all at once when a query completes?
Does impala support generic jdbc?
What happens when the data set exceeds available memory?
Can impala be used for complex event processing?
What are the names of daemons in impala?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)