[hadoop@dmp1 ~]$ hdfs dfs -ls / 16/09/20 15:00:44 WARN ipc.Client: Exception encountered while connecting to the server : javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)] ls: Failed on local exception: java.io.IOException: javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)]; Host Details : local host is: " dmp1.youedata.com/192.168.249.129"; destination host is: "dmp1.youedata.com ":9000; [hadoop@dmp1 ~]$ klist Ticket cache: KEYRING:persistent:1004:1004 Default principal: [email protected]
Valid starting Expires Service principal 09/20/2016 14:57:34 09/21/2016 14:57:31 krbtgt/[email protected] renew until 09/27/2016 14:57:31 [hadoop@dmp1 ~]$ I have run kinit [email protected] before . 2016-09-21 10:14 GMT+08:00 Wei-Chiu Chuang <[email protected]>: > You need to run kinit command to authenticate before running hdfs dfs -ls > command. > > Wei-Chiu Chuang > > On Sep 20, 2016, at 6:59 PM, kevin <[email protected]> wrote: > > Thank you Brahma Reddy Battula. > It's because of my problerm of the hdfs-site config file and https > ca configuration. > now I can startup namenode and I can see the datanodes from the web. > but When I try hdfs dfs -ls /: > > *[hadoop@dmp1 hadoop-2.7.3]$ hdfs dfs -ls /* > *16/09/20 07:56:48 WARN ipc.Client: Exception encountered while connecting > to the server : javax.security.sasl.SaslException: GSS initiate failed > [Caused by GSSException: No valid credentials provided (Mechanism level: > Failed to find any Kerberos tgt)]* > *ls: Failed on local exception: java.io.IOException: > javax.security.sasl.SaslException: GSS initiate failed [Caused by > GSSException: No valid credentials provided (Mechanism level: Failed to > find any Kerberos tgt)]; Host Details : local host is: > "dmp1.example.com/192.168.249.129 > <http://dmp1.example.com/192.168.249.129>"; destination host is: "dmp1.* > *example**.com":9000; * > > current user is hadoop which startup hdfs , and I have add addprinc hadoop > with commond : > kadmin.local -q "addprinc hadoop" > > > 2016-09-20 17:33 GMT+08:00 Brahma Reddy Battula < > [email protected]>: > >> Seems to be property problem.. it should be *principal* ( ālā is missed). >> >> >> >> *<property>* >> >> * <name>dfs.secondary.namenode.kerberos.principa</name>* >> >> * <value>hadoop/[email protected] <[email protected]></value>* >> >> *</property>* >> >> >> >> >> >> For namenode httpserver start fail, please check rakesh comments.. >> >> >> >> This is probably due to some missing configuration. >> >> Could you please re-check the ssl-server.xml, keystore and truststore >> properties: >> >> >> >> ssl.server.keystore.location >> >> ssl.server.keystore.keypassword >> >> ssl.client.truststore.location >> >> ssl.client.truststore.password >> >> >> >> >> >> --Brahma Reddy Battula >> >> >> >> *From:* kevin [mailto:[email protected]] >> *Sent:* 20 September 2016 16:53 >> *To:* Rakesh Radhakrishnan >> *Cc:* user.hadoop >> *Subject:* Re: hdfs2.7.3 kerberos can not startup >> >> >> >> thanks, but my issue is name node could *Login successful,but second >> namenode couldn't. and name node got a HttpServer.start() threw a non Bind >> IOException:* >> >> >> >> hdfs-site.xml: >> >> >> >> *<property>* >> >> * <name>dfs.webhdfs.enabled</name>* >> >> * <value>true</value>* >> >> *</property>* >> >> >> >> *<property>* >> >> * <name>dfs.block.access.token.enable</name>* >> >> * <value>true</value>* >> >> *</property>* >> >> >> >> *<!-- NameNode security config -->* >> >> *<property>* >> >> * <name>dfs.namenode.kerberos.pr >> <http://dfs.namenode.kerberos.pr>incipal</name>* >> >> * <value>hadoop/[email protected] <[email protected]></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.namenode.keytab.file</name>* >> >> * <value>/etc/hadoop/conf/hdfs.keytab</value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.https.port</name>* >> >> * <value>50470</value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.namenode.https-address</name>* >> >> * <value>dmp1.example.com:50470 <http://dmp1.example.com:50470/></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.namenode.kerberos.in >> <http://dfs.namenode.kerberos.in>ternal.spnego.principa</name>* >> >> * <value>HTTP/[email protected] <[email protected]></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.web.authentication.kerberos.keytab</name>* >> >> * <value>/etc/hadoop/conf/hdfs.keytab</value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.http.policy</name>* >> >> * <value>HTTPS_ONLY</value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.https.enable</name>* >> >> * <value>true</value>* >> >> *</property>* >> >> >> >> >> >> *<!-- secondary NameNode security config -->* >> >> *<property>* >> >> * <name>dfs.namenode.secondary.http-address</name>* >> >> * <value>dmp1.example.com:50090 <http://dmp1.example.com:50090/></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.secondary.namenode.keytab.file</name>* >> >> * <value>/etc/hadoop/conf/hdfs.keytab</value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.secondary.namenode.kerberos.principa</name>* >> >> * <value>hadoop/[email protected] <[email protected]></value>* >> >> *</property> * >> >> *<property>* >> >> * <name>dfs.secondary.namenode.kerberos.internal.spnego.principal</name>* >> >> * <value>HTTP/[email protected] <[email protected]></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.namenode.secondary.https-port</name>* >> >> * <value>50470</value>* >> >> *</property>* >> >> >> >> >> >> *<!-- JournalNode security config -->* >> >> >> >> *<property>* >> >> * <name>dfs.journalnode.keytab.file</name>* >> >> * <value>/etc/hadoop/conf/hdfs.keytab</value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.journalnode.kerberos.principa</name>* >> >> * <value>hadoop/[email protected] <[email protected]></value>* >> >> *</property> * >> >> *<property>* >> >> * <name>dfs.journalnode.kerberos.internal.spnego.principa</name>* >> >> * <value>HTTP/[email protected] <[email protected]></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.web.authentication.kerberos.keytab</name>* >> >> * <value>/etc/hadoop/conf/hdfs.keytab</value>* >> >> *</property>* >> >> >> >> >> >> *<!-- DataNode security config -->* >> >> *<property>* >> >> * <name>dfs.datanode.kerberos.pr >> <http://dfs.datanode.kerberos.pr>incipal</name>* >> >> * <value>hadoop/[email protected] <[email protected]></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.datanode.keytab.file</name>* >> >> * <value>/etc/hadoop/conf/hdfs.keytab</value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.datanode.data.dir.pe >> <http://dfs.datanode.data.dir.pe>rm</name>* >> >> * <value>700</value>* >> >> *</property>* >> >> >> >> *<!-- datanode SASL-->* >> >> *<property>* >> >> * <name>dfs.datanode.address</name>* >> >> * <value>0.0.0.0:61004 <http://0.0.0.0:61004/></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.datanode.http.address</name>* >> >> * <value>0.0.0.0:61006 <http://0.0.0.0:61006/></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.datanode.https.address</name>* >> >> * <value>0.0.0.0:50470 <http://0.0.0.0:50470/></value>* >> >> *</property>* >> >> >> >> *<property>* >> >> * <name>dfs.data.transfer.protection</name>* >> >> * <value>integrity</value>* >> >> *</property>* >> >> >> >> *<property>* >> >> * <name>dfs.web.authentication.kerberos.principal</name>* >> >> * <value>HTTP/[email protected] <[email protected]></value>* >> >> *</property>* >> >> *<property>* >> >> * <name>dfs.web.authentication.kerberos.keytab</name>* >> >> * <value>/etc/hadoop/conf/hdfs.keytab</value>* >> >> *</property>* >> >> >> >> *and [hadoop@dmp1 hadoop-2.7.3]$ klist -ket /etc/hadoop/conf/hdfs.keytab* >> >> >> >> >> >> *Keytab name: FILE:/etc/hadoop/conf/hdfs.keytab* >> >> *KVNO Timestamp Principal* >> >> *---- ------------------- >> ------------------------------------------------------* >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 16:00:41 hdfs/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 16:00:41 HTTP/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (aes256-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (aes128-cts-hmac-sha1-96) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (des3-cbc-sha1) * >> >> * 2 09/19/2016 20:21:03 hadoop/[email protected] >> <[email protected]> (arcfour-hmac) * >> >> >> >> 2016-09-20 15:52 GMT+08:00 Rakesh Radhakrishnan <[email protected]>: >> >> >>>>>>Caused by: javax.security.auth.login.LoginException: Unable to >> obtain password from user >> >> >> >> Could you please check kerberos principal name is specified correctly in >> >> "hdfs-site.xml", which is used to authenticate against Kerberos. >> >> >> >> If keytab file defined in "hdfs-site.xml" and doesn't exists or wrong >> path, you will see >> >> this error. So, please verify the path and the keytab filename correctly >> >> configured. >> >> >> >> I hope hadoop discussion thread, https://goo.gl/M6l3vv may help you. >> >> >> >> >> >> >>>>>>>2016-09-20 00:54:06,665 INFO org.apache.hadoop.http.HttpServer2: >> HttpServer.start() threw a non Bind IOException >> >> java.io.IOException: !JsseListener: java.lang.NullPointerException >> >> >> >> This is probably due to some missing configuration. >> >> Could you please re-check the ssl-server.xml, keystore and truststore >> properties: >> >> >> >> ssl.server.keystore.location >> >> ssl.server.keystore.keypassword >> >> ssl.client.truststore.location >> >> ssl.client.truststore.password >> >> >> >> Rakesh >> >> >> >> On Tue, Sep 20, 2016 at 10:53 AM, kevin <[email protected]> wrote: >> >> *hi,all:* >> >> *My environment : Centos7.2 hadoop2.7.3 jdk1.8* >> >> *after I config hdfs with kerberos ,I can't start up with >> sbin/start-dfs.sh* >> >> >> >> *::namenode log as below * >> >> >> >> *STARTUP_MSG: build = Unknown -r Unknown; compiled by 'root' on >> 2016-09-18T09:05Z* >> >> *STARTUP_MSG: java = 1.8.0_102* >> >> *************************************************************/* >> >> *2016-09-20 00:54:05,822 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: registered UNIX signal >> handlers for [TERM, HUP, INT]* >> >> *2016-09-20 00:54:05,825 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: createNameNode []* >> >> *2016-09-20 00:54:06,078 INFO >> org.apache.hadoop.metrics2.impl.MetricsConfig: loaded properties from >> hadoop-metrics2.properties* >> >> *2016-09-20 00:54:06,149 INFO >> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: Scheduled snapshot >> period at 10 second(s).* >> >> *2016-09-20 00:54:06,149 INFO >> org.apache.hadoop.metrics2.impl.MetricsSystemImpl: NameNode metrics system >> started* >> >> *2016-09-20 00:54:06,151 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: fs.defaultFS is >> hdfs://dmp1.example.com:9000 <http://dmp1.example.com:9000/>* >> >> *2016-09-20 00:54:06,152 INFO >> org.apache.hadoop.hdfs.server.namenode.NameNode: Clients are to use >> dmp1.example.com:9000 <http://dmp1.example.com:9000/> to access this >> namenode/service.* >> >> *2016-09-20 00:54:06,446 INFO >> org.apache.hadoop.security.UserGroupInformation: Login successful for user >> hadoop/[email protected] <[email protected]> using >> keytab file /etc/hadoop/conf/hdfs.keytab* >> >> *2016-09-20 00:54:06,472 INFO org.apache.hadoop.hdfs.DFSUtil: Starting >> web server as: HTTP/[email protected] >> <[email protected]>* >> >> *2016-09-20 00:54:06,475 INFO org.apache.hadoop.hdfs.DFSUtil: Starting >> Web-server for hdfs at: https://dmp1.example.com:50470 >> <https://dmp1.example.com:50470/>* >> >> *2016-09-20 00:54:06,517 INFO org.mortbay.log: Logging to >> org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via >> org.mortbay.log.Slf4jLog* >> >> *2016-09-20 00:54:06,533 INFO >> org.apache.hadoop.security.authentication.server.AuthenticationFilter: >> Unable to initialize FileSignerSecretProvider, falling back to use random >> secrets.* >> >> *2016-09-20 00:54:06,542 INFO org.apache.hadoop.http.HttpRequestLog: Http >> request log for http.requests.namenode is not defined* >> >> *2016-09-20 00:54:06,546 INFO org.apache.hadoop.http.HttpServer2: Added >> global filter 'safety' >> (class=org.apache.hadoop.http.HttpServer2$QuotingInputFilter)* >> >> *2016-09-20 00:54:06,548 INFO org.apache.hadoop.http.HttpServer2: Added >> filter static_user_filter >> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to >> context hdfs* >> >> *2016-09-20 00:54:06,548 INFO org.apache.hadoop.http.HttpServer2: Added >> filter static_user_filter >> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to >> context static* >> >> *2016-09-20 00:54:06,548 INFO org.apache.hadoop.http.HttpServer2: Added >> filter static_user_filter >> (class=org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter) to >> context logs* >> >> *2016-09-20 00:54:06,653 INFO org.apache.hadoop.http.HttpServer2: Added >> filter 'org.apache.hadoop.hdfs.web.Au >> <http://org.apache.hadoop.hdfs.web.au/>thFilter' >> (class=org.apache.hadoop.hdfs.web.AuthFilter)* >> >> *2016-09-20 00:54:06,654 INFO org.apache.hadoop.http.HttpServer2: >> addJerseyResourcePackage: >> packageName=org.apache.hadoop.hdfs.server.namenode.web.resources;org.apache.hadoop.hdfs.web.resources, >> pathSpec=/webhdfs/v1/** >> >> *2016-09-20 00:54:06,657 INFO org.apache.hadoop.http.HttpServer2: Adding >> Kerberos (SPNEGO) filter to getDelegationToken* >> >> *2016-09-20 00:54:06,658 INFO org.apache.hadoop.http.HttpServer2: Adding >> Kerberos (SPNEGO) filter to renewDelegationToken* >> >> *2016-09-20 00:54:06,658 INFO org.apache.hadoop.http.HttpServer2: Adding >> Kerberos (SPNEGO) filter to cancelDelegationToken* >> >> *2016-09-20 00:54:06,659 INFO org.apache.hadoop.http.HttpServer2: Adding >> Kerberos (SPNEGO) filter to fsck* >> >> *2016-09-20 00:54:06,659 INFO org.apache.hadoop.http.HttpServer2: Adding >> Kerberos (SPNEGO) filter to imagetransfer* >> >> *2016-09-20 00:54:06,665 WARN org.mortbay.log: >> java.lang.NullPointerException* >> >> *2016-09-20 00:54:06,665 INFO org.apache.hadoop.http.HttpServer2: >> HttpServer.start() threw a non Bind IOException* >> >> *java.io.IOException: !JsseListener: java.lang.NullPointerException* >> >> *at >> org.mortbay.jetty.security.SslSocketConnector.newServerSocket(SslSocketConnector.java:516)* >> >> *at >> org.apache.hadoop.security.ssl.SslSocketConnectorSecure.newServerSocket(SslSocketConnectorSecure.java:47)* >> >> *at org.mortbay.jetty.bio.SocketConnector.open(SocketConnector.java:73)* >> >> *at >> org.apache.hadoop.http.HttpServer2.openListeners(HttpServer2.java:914)* >> >> *at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:856)* >> >> *at org.apache.hadoop.hdfs.server.namenode.NameNodeHttpServer.st >> <http://namenode.namenodehttpserver.st/>art(NameNodeHttpServer.java:142)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.NameNode.startHttpServer(NameNode.java:753)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:639)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:812)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:796)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1493)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNode.java:1559)* >> >> >> >> >> >> *::second namenode log as below * >> >> >> >> *STARTUP_MSG: build = Unknown -r Unknown; compiled by 'root' on >> 2016-09-18T09:05Z* >> >> *STARTUP_MSG: java = 1.8.0_102* >> >> *************************************************************/* >> >> *2016-09-20 00:54:14,885 INFO >> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: registered UNIX >> signal handlers for [TERM, HUP, INT]* >> >> *2016-09-20 00:54:15,263 FATAL >> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode: Failed to start >> secondary namenode* >> >> *java.io.IOException: Login failure for hadoop from keytab >> /etc/hadoop/conf/hdfs.keytab: javax.security.auth.login.LoginException: >> Unable to obtain password from user* >> >> >> >> *at >> org.apache.hadoop.security.UserGroupInformation.loginUserFromKeytab(UserGroupInformation.java:963)* >> >> *at org.apache.hadoop.security.SecurityUtil.login(SecurityUtil.java:246)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.initialize(SecondaryNameNode.java:217)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.<init>(SecondaryNameNode.java:192)* >> >> *at >> org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode.main(SecondaryNameNode.java:671)* >> >> *Caused by: javax.security.auth.login.LoginException: Unable to obtain >> password from user* >> >> >> >> *at >> com.sun.security.auth.module.Krb5LoginModule.promptForPass(Krb5LoginModule.java:897)* >> >> *at >> com.sun.security.auth.module.Krb5LoginModule.attemptAuthentication(Krb5LoginModule.java:760)* >> >> *at >> com.sun.security.auth.module.Krb5LoginModule.login(Krb5LoginModule.java:617)* >> >> *at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)* >> >> *at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)* >> >> *at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)* >> >> *at java.lang.reflect.Method.invoke(Method.java:498)* >> >> *at javax.security.auth.login.LoginContext.invoke(LoginContext.java:755)* >> >> *at >> javax.security.auth.login.LoginContext.access$000(LoginContext.java:195)* >> >> *at javax.security.auth.login.LoginContext$4.run(LoginContext.java:682)* >> >> *at javax.security.auth.login.LoginContext$4.run(LoginContext.java:680)* >> >> *at java.security.AccessController.doPrivileged(Native Method)* >> >> *at >> javax.security.auth.login.LoginContext.invokePriv(LoginContext.java:680)* >> >> *at javax.security.auth.login.LoginContext.login(LoginContext.java:587)* >> >> *at >> org.apache.hadoop.security.UserGroupInformation.loginUserFromKeytab(UserGroupInformation.java:954)* >> >> *... 4 more* >> >> >> >> >> > > >
