Quantcast
Channel: Hortonworks » All Replies
Viewing all 3435 articles
Browse latest View live

Cannot start hive – and get Exception

$
0
0

Hi
i’ve downloaded Sandbox of 2.1 some time ago.
I’m sure it was running. I’ve been playing a bit with it and than I’ve installed spark, sbt, scala.

when I wanted to run hive, I got this error:

Logging initialized using configuration in file:/etc/hive/conf.dist/hive-log4j.properties
Exception in thread “main” java.lang.RuntimeException: java.lang.RuntimeException: Unable to instantiate org.apache.hadoop.hive.metastore.HiveMetaStoreClient
at org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:347)
at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:681)
at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:625)
etc…

I’ve restarted the system and noticed that hive server nad hive2 served gave WARNING while starting system with something like: strying to start script but cannot fing it. I dodn’t copy the full text.

Can you please help on how to start the hive server?

Regards
Pawel


HIve 'SDS' table is full

$
0
0

Hello All,
We have a 10 node cluster with hive installed in it. I have table in hive with 1.9 billion records in it. I am doing dynamic partitions on the table around 92K partitions. I am getting below error. Could you please tell me how to avoid this error?

FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:javax.jdo.JDODataStoreException: Insert of object "org.apache.hadoop.hive.metastore.model.MStorageDescriptor@3967eaa4" using statement "INSERT INTO SDS(SD_ID,INPUT_FORMAT,IS_STOREDASSUBDIRECTORIES,NUM_BUCKETS,LOCATION,IS_COMPRESSED,SERDE_ID,CD_ID,OUTPUT_FORMAT
) VALUES (?,?,?,?,?,?,?,?,?)" failed : The table 'SDS' is full
	at org.datanucleus.api.jdo.NucleusJDOHelper.getJDOExceptionForNucleusException(NucleusJDOHelper.java:451)
	at org.datanucleus.api.jdo.JDOPersistenceManager.jdoMakePersistent(JDOPersistenceManager.java:732)
	at org.datanucleus.api.jdo.JDOPersistenceManager.makePersistent(JDOPersistenceManager.java:752)
	at org.apache.hadoop.hive.metastore.ObjectStore.createTable(ObjectStore.java:719)
	at sun.reflect.GeneratedMethodAccessor31.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.hive.metastore.RawStoreProxy.invoke(RawStoreProxy.java:108)
	at com.sun.proxy.$Proxy0.createTable(Unknown Source)
	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_core(HiveMetaStore.java:1262)
	at org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.create_table_with_environment_context(HiveMetaStore.java:1295)
	at sun.reflect.GeneratedMethodAccessor32.invoke(Unknown Source)
	at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
	at java.lang.reflect.Method.invoke(Method.java:606)
	at org.apache.hadoop.hive.metastore.RetryingHMSHandler.invoke(RetryingHMSHandler.java:105)
	at com.sun.proxy.$Proxy1.create_table_with_environment_context(Unknown Source)
	at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$create_table_with_environment_context.getResult(ThriftHiveMetastore.java:8016)
	at org.apache.hadoop.hive.metastore.api.ThriftHiveMetastore$Processor$create_table_with_environment_context.getResult(ThriftHiveMetastore.java:8000)
	at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
	at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:110)
	at org.apache.hadoop.hive.metastore.TUGIBasedProcessor$1.run(TUGIBasedProcessor.java:107)
	at java.security.AccessController.doPrivileged(Native Method)
	at javax.security.auth.Subject.doAs(Subject.java:415)
	at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)

Could not locate executable null\bin\winutils.exe in the Hadoop binaries

$
0
0

What are the necessary steps required to resolve the issue?? kindly help.

Reply To: Could not locate executable null\bin\winutils.exe in the Hadoop binaries

$
0
0

Mohan,

Are you hitting an issue while trying to run a Hadoop command? If so, can you share the command you tried running and the error log?
Or are you looking to locate the winutils.exe binary in a directory on disk?

Let us know what you are trying to do or where you are hitting an error and we can help.

Reply To: Could not locate executable null\bin\winutils.exe in the Hadoop binaries

$
0
0

I am running the Hadoop code from Talend ETL tool.

Reply To: Spark 1.0.1 Tech preview available

$
0
0

Hi,

I went through the preview instructions and everything worked great. For those who would like to write a java/scala client, which spark-core version should we use via maven? I couldn’t find the jars in the Hortonworks maven repo, is one there? I was getting serialization errors with some of the spark classes when using “org.apache.spark” % “spark-streaming_2.10″ % “1.0.2″. Should I use 1.0.1 from apache?

org.apache.spark.SparkException: Job aborted due to stage failure: Task 0.0:0 failed 4 times, most recent failure: Exception failure in TID 3 on host ip-172-31-128-8.ec2.internal: java.io.InvalidClassException: org.apache.spark.rdd.RDD; local class incompatible: stream classdesc serialVersionUID = -6766554341038829528, local class serialVersionUID = 385418487991259089

Best,

-Mike

Data import from SQL Server

$
0
0

Hi I am hadoop beginner and exploring some hadoop options from our existing project. here is my question.
We have Hortonnetwork version of hadoop on windows.
Iam trying to use below command to load data from on-premisis sql server to HDFS :

sqoop list-databases –connect jdbc::sqlserver://<servername>:1433 –username <MyUserName> –password <MyPassword>
This is just to check my connection before pulling actual tables in HDFS..but the command throws below error :

“Got error creating database manager: java.io.IOException: No manager for connect string: jdbc::sqlserver:<Servername>”
few background points:
- Able to ping the server from hadoop server.
- Have copied sqljdbc4.jar file to sqoop/lib directory.
-sql server 2012

Can anyone share their experiences/solutions if you have any such.

Thanks
Sreeni

Hadoop tools advice

$
0
0

We have a large integrated database containing a very diverse range of variables. We identify a master population using simple business rules to determine if they meet initial parameters, then keep narrowing down the population using more logic until we know whether they meet the given requirement. This logic is not terribly complex.

After the population is identified we create a large list of transactions associated with these people, then apply further business rules to these transactions to determine if the associated people are ‘in’ or ‘out’ of a final population of interest.

Our current t-sql based solution does not perform well and we’re looking for alternatives in the hadoop stack, particularly around using the parallel processing capabilities in hadoop to speed us up and enable larger data problems.

We’re looking for suggestions on which of the hadoop or surrounding technologies (even spark, r integrated with hadoop etc etc) would help us with this problem and others like it.

We would really appreciate it :-)


A New Nook, a Good, a Lock With No Key

Reply To: Data import from SQL Server

$
0
0

Hello,

Can you try with that syntax :
sqoop list-databases –connect ‘jdbc:sqlserver://<servername>:1433;username=<username>;password=<password>’

Hope it work.

HiveServer2 failed to start

$
0
0

Hi,
I deployed hadoop using ambari 1.6.1.While starting HiveServer2,I am getting below error.Similar error for YARN and MapReduce2 too.Please help me out.

Fail: Execution of ‘hadoop fs -mkdir rpm -q hadoop | grep -q "hadoop-1" || echo "-p" /apps/hive/warehouse /user/hive && hadoop fs -chmod 777 /apps/hive/warehouse && hadoop fs -chmod 700 /user/hive && hadoop fs -chown hive /apps/hive/warehouse /user/hive’ returned 1. mkdir: Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message contained an invalid tag (zero).

MR job jusing AvroMultipleOutputs failing with Replication error

$
0
0

Hello,

I am using AvroMultipleOutputs class to dynamically sort the data to output files in reducer but when the number of output files is higher than it fails with following eorror:
Error: org.apache.hadoop.ipc.RemoteException(java.io.IOException): File /data/pif-dob-categorize/2014/08/26/14/_temporary/1/_temporary/attempt_1408977992653_0120_r_000000_0/HISTORY/20131204/64619-r-00000.avro could only be replicated to 0 nodes instead of minReplication (=1). There are 2 datanode(s) running and no node(s) are excluded in this operation. at org.apache.hadoop.hdfs.server.blockmanagement.BlockManager.chooseTarget(BlockManager.java:1384) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:2503) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addBlock(NameNodeRpcServer.java:555) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.addBlock(ClientNamenodeProtocolServerSideTranslatorPB.java:387) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java:…
org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.locateFollowingBlock(DFSOutputStream.java:1231) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1078) at org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:514) Container killed by the ApplicationMaster. Container killed on request. Exit code is 143

I checked HDFS and datanodes all seems ok, enough space no network partition etc. Details can be found on SO

http://stackoverflow.com/questions/25509161/hadoop-avromultipleoutputs-error-could-only-be-replicated-to-0-nodes-instead-o

It seems somehow related to the number of files created/opened as for smaller amount of data job is ok! What settings can limit the number of output files?

Thx

File References in Java UDF

$
0
0

I have uploaded to the Sandbox files I need to read from inside a Java UDF.
Trouble is, I have no idea of the path to the files so they can be opened.
I can’t seem to simply use the file name, as I do with the LOAD command in pig.

Moreover, I don’t have the source code to the package that actually opens and processes the file, so I can’t take any special steps to read the data,
Best I can do is pass the file path to the constructor and hope for the best.

Can more learned colleagues offer any help?
Thanks in advance.

NodeManager won't start on Ambrari 1.6.1

$
0
0

Ambrari 1.6.1 building and HDP 2.1 cluster, JDK 1.7, all hosts RedHat 6.5. All checks passed, iptables and selinux off on all machines. I’ve tried several different cluster topologies and I always get the same result–the NodeManager won’t start. It initially showed as HiveServer2 failing to start, but by pairing back to the bare minimum services, it looks like it’s actually the NodeManager. It manifests in the logs as a connection refused when the slaves try to reach back to the master on 8020 (which, according to the HDFS > Config > Advanced section, is the location of the NDFS service). Now, I know it is tempting to write this off as something wrong with the way I configured my hosts or my network, but then the following maybe suggests otherwise:
The services seem to be running ok on the master:

[root@master ~]# jps
6229 QuorumPeerMain
8736 SecondaryNameNode
7515 ApplicationHistoryServer
8135 JobHistoryServer
10550 Jps
8317 ResourceManager
6530 NameNode
[root@master ~]#

There is listening on selected hadoop ports:

[root@master ~]# netstat -l | grep tcp | grep ":8"
tcp master.jhuapl:8141 *:* LISTEN
tcp master.jhuapl:8050 *:* LISTEN
tcp master.jhuapl:8188 *:* LISTEN
tcp master.jhuapl:8030 *:* LISTEN
tcp *:8670 *:* LISTEN

And here’s the kicker. If I run a little listener on the master:

[root@master ~]# nc -l 12345

I can communicate with it from the slave:

[root@slave1 ~]# nc -z master 12345
Connection to master 12345 port [tcp/italk] succeeded!

But not on the expected Hadoop service ports!

[root@slave1 ~]# nc -z master 8020
[root@slave1 ~]# nc -z master 8020
[root@slave1 ~]#

A wild guess might be that there is something wrong with the services on the master, but I don’t know how to prove that or not.

Please suggest anything I might try at this point.

Thanks,
Clark

p.s. Here’s the original stacktrace from the slave that shows the communication problem with the master:

2014-08-27 17:51:34,556 - Error while executing command 'start':
Traceback (most recent call last):

File "/var/lib/ambari-agent/cache/stacks/HDP/2.0.6/services/YARN/package/scripts/nodemanager.py", line 40, in start
...SNIP...
Fail: Execution of 'hadoop fs -mkdir rpm -q hadoop | grep -q "hadoop-1" || echo "-p" /app-logs /mapred /mapred/system /mr-history/tmp /mr-history/done && hadoop fs -chmod -R 777 /app-logs && hadoop fs -chmod 777 /mr-history/tmp && hadoop fs -chmod 1777 /mr-history/done && hadoop fs -chown mapred /mapred && hadoop fs -chown hdfs /mapred/system && hadoop fs -chown yarn:hadoop /app-logs && hadoop fs -chown mapred:hadoop /mr-history/tmp /mr-history/done' returned 1. mkdir: Call From slave1.jhuapl.edu/127.0.1.1 to master.jhuapl.edu:8020 failed on connection exception: java.net.ConnectException: Connection refused; see: http://wiki.apache.org/hadoop/ConnectionRefused

Reply To: Data import from SQL Server

$
0
0

Thanks mejane,

I still get the jdbc error . the error says “Java Runtime Environment (JRE) version 1.6 is not supported by this driver. Use the sqljdbc4.jar class library, which provides support for JDBC 4.0.”

I have downloaded and copied sqljdbc4.jar and sqljdbc.jar files to sqoop lib path.

below is the command and it result :

COMMAND :
C:\Hadoop\sqoop-1.4.3.1.3.0.0-0380\bin>sqoop import –driver com.microsoft.sqlserver.jdbc.SQLServerDriver –connect “jdbc:sqlserver://ch1gfssldev;database=skylabuat;username=XXXXX;password=XXXX” –table sku –target-dir /user/hadoopuser/Sreeni/SKYLABUAT/NEWDATA -m 1 –verbose

RESULT:
File “C:\hadoop\\hcatalog-0.11.0.1.3.0.0-0380\bin\hcat.py”, line 134
print “Would run:”
^
SyntaxError: invalid syntax
14/08/27 13:47:24 DEBUG tool.BaseSqoopTool: Enabled debug logging.
14/08/27 13:47:24 DEBUG sqoop.ConnFactory: Loaded manager factory: com.cloudera.sqoop.manager.DefaultManagerFactory
14/08/27 13:47:24 WARN sqoop.ConnFactory: Parameter –driver is set to an explicit driver however appropriate connection manager is not being set (via –connection-manager). Sqoop is going to fall back to org.apache.sqoop.manager.
GenericJdbcManager. Please specify explicitly which connection manager should be used next time.
14/08/27 13:47:24 INFO manager.SqlManager: Using default fetchSize of 1000
14/08/27 13:47:24 INFO tool.CodeGenTool: Beginning code generation
14/08/27 13:47:25 DEBUG manager.SqlManager: No connection paramenters specified. Using regular API for making connection.
Aug 27, 2014 1:47:25 PM com.microsoft.sqlserver.jdbc.SQLServerConnection <init>
SEVERE: Java Runtime Environment (JRE) version 1.6 is not supported by this driver. Use the sqljdbc4.jar class library, which provides support for JDBC 4.0.
14/08/27 13:47:25 ERROR sqoop.Sqoop: Got exception running Sqoop: java.lang.UnsupportedOperationException: Java Runtime Environment (JRE) version 1.6 is not supported by this driver. Use the sqljdbc4.jar class library, which provi
des support for JDBC 4.0.
java.lang.UnsupportedOperationException: Java Runtime Environment (JRE) version 1.6 is not supported by this driver. Use the sqljdbc4.jar class library, which provides support for JDBC 4.0.
at com.microsoft.sqlserver.jdbc.SQLServerConnection.<init>(SQLServerConnection.java:238)
at com.microsoft.sqlserver.jdbc.SQLServerDriver.connect(SQLServerDriver.java:841)


Reply To: Spark 1.0.1 Tech preview available

$
0
0

Mike,

Try using 1.0.1 version of the spark-core since that’s what the TP is build with. When we revise the TP we will make sure to publish the spark-core jar to the HWRK repo.

Please let me know how you make out.

Thanks,
Vinay

Cannot run PIG from HUE

$
0
0

I have installed a 4 node cluster. The cluster is up and running. I’m able to load data and run queries. However I’m not able to run PIG scripts. I get the error :
ls: cannot access /grid/1/hadoop/yarn/local/usercache/hue/appcache/application_1409126352550_0003/container_1409126352550_0003_01_000002
/hive.tar.gz/hive/lib/slf4j-api-*.jar: No such file or directory
Error: Could not find or load main class hive.metastore.uris=thrift:..localhost:9933
Please suggest ways to resolve this issue.
Thanks,
Component Version information
Hue 2.5.0-632
HDP2.1.4
Hadoop 2.4.0
Pig 0.12.1
Hive-Hcatalog0.13.0
Oozie 4.0.0
Ambari1.6-98
HBase 0.98.0
Knox 0.4.0
Storm 0.9.1
Falcon 0.5.0

Reply To: Importing snapshots from Amazon S3 to HBase

$
0
0

Managed to get s3n:// import to work using this tool but not s3 block import. I’m looking at forking the tool and producing a howto guide.

mysql Password

$
0
0

I deployed the Sandbox VM and noticed MySql comes installed. Does anyone know what the root password is?

Unable to download Hortonworks Connector for Teradata

Viewing all 3435 articles
Browse latest View live


Latest Images

<script src="https://jsc.adskeeper.com/r/s/rssing.com.1596347.js" async> </script>