Reply To: Is there an Cloudera HdfsFindTool equivalent in HDP?
Thanks Chris. I am looking to use find to delete old files in hdfs space. Something with -mtime and -exec and the find in HDP 2.3.0 is just for -name Do you know if thru Solr I could emulate that...
View ArticleHDFS permissions reset for hive.metastore.warehouse.dir after hive restart
Hello, Since we’ve upgraded to 2.3 we’ve noticed that our hive.metastore.warehouse.dir permissions reset to whatever the default is. We have many files in the location (/user/hive/warehouse) which...
View ArticleReply To: Sandbox – Pig Basic Tutorial example is nbot working
Hi, No, Ambari is not required to be enabled for running Pig scripts. Ambari is a f<span style=”color: #000000; font-family: Arial; font-size: 14.6666666666667px; white-space: pre-wrap;...
View Articleerror running hive query "S020 Data storage error"
hi I am running a simple query SELECT avg(sk_daily_hydrometric_eng.waterlevelinmeters) FROM sk_daily_hydrometric_eng LIMIT 100; the table itself has ~1,000,000 + records so maybe that the issue …but...
View ArticleReply To: odbc drivers for HDP 2.3 windows 7 and MacOS
I am using the microsoft hive ODBC driver but you need to sign up for Microsoft azure to get it which requires a credit card (I used a prepaid credit card i got from winning at a coffee shop)
View ArticleReading Data from pdf files?
Hi All, I need to load data from multiple pdf files and use it for analysis in hadoop. I don’t have java background hence can’t write complete map redue program. Can some one please help.Its urgent. I...
View ArticleUpgrade HDP-2.2.4.2-2 to HDP-2.3.0.0 using Ambari
Hi, When I try upgrade I get the following requirements error: The SNameNode component must be deleted from all hosts Reason: The SNameNode component must be deleted from host: SECONDARY_NAMENODE....
View ArticleReply To: Reading Data from pdf files?
Hi Prafull, I’m not sure what processing you want to do. But if you are dealing with PDFs then you might want to look at HDP Search (Solr): http://doc.lucidworks.com/hdpsearch23/Guide-Solr.html...
View ArticleReply To: Reading Data from pdf files?
HI Rafael, Thanks fr this. My requirement is very simple. 1. read data from PDF file. 2. Store it in HDFS 3. Extract data from HDFS using Hive tables and queries.
View ArticleHistory Server/MapReduce2 stopped
I’m tried to do the hdp 2.2 to 2.3 rolling upgrade through Ambari 2.1.1. It had problems due to the second namenode being stuck in maintenance mode. So I rolled back. Ever since I did that, mapreduce2...
View Articlewhich JAR files are required for remote JDBC clients
Using 2.3.0.0-2557 have started the SPARK thriftserver. There does not appear to be any documentation on which collection of JAR files would be required for a remote client using a non-kerberized...
View ArticleSpark jars in Oozie sharelib not built for YARN (HDP 2.3)
Environment: HDP 2.3. Oozie v4.2 (Oozie server 4.2.0.2.3.0.0-2557) Hi, Oozie job with Spark action workflow fails with the following log: Error: Could not load YARN classes. This copy of Spark may not...
View ArticleProblem with LDAP
Hi, I have installed knox with ambari on a centos machine. When i try to start LDAP with “/usr/hdp/current/knox-server/bin/ldap.sh start” i receive “Starting LDAP failed” How i can resolve it? Anyone...
View ArticleSpark action job tries to connect to Resource Mgr on wrong (default) address
Hi, Environment: HDP 2.3. Oozie v4.2 (Oozie server 4.2.0.2.3.0.0-2557) on a cluster setup in EC2 Spark workflow job is stuck in RUNNING state as it tries to connect to the Resource Manager using the...
View ArticleSynchronizing Ambari cluster configurations
We have been exploring Apache Ambari with HDP 2.2 to setup a cluster. Our backend features three environments: testing, staging and production which is a standard practice in our industry. When we...
View ArticleReply To: insert select with regex serde failing
I have the same problem, you can fix them? Thanks Sonia
View ArticleFlume – how to create a custom key for a HDFS SequenceFile?
I’m using Flume’s HDFS SequenceFile sink for writing data to HDFS. I’m looking for a possibility to create “custom keys”. Per default, Flume is using the Timestamp as key within a SequenceFile....
View ArticleRename Namenode HA
Hi everyone, I have have set up a HA cluster which is fully working without any problem. My problem is that my client (a Java client) is using the name of my cluster as a variable. This name contains...
View ArticleReply To: Problem with LDAP
Can you look in the {GATEWAY_HOME}/logs/ldap.log, ldap.err and/or ldap.out files? You may already have an instance or some other server running on the port.
View ArticleReply To: Authentication Problem with Knox
I was able to start LDAP, but now i receive “Error 403 Forbidden”. You know why?
View Article