Skip to content

Latest commit

 

History

History
99 lines (67 loc) · 4.03 KB

File metadata and controls

99 lines (67 loc) · 4.03 KB
layout title permalink
doc
Site Management
/docs/tutorial/setup.html

Only For Eagle 0.3.0. Since Eagle 0.4.0, a prefix classification. is required for each property and the configuration becomes java properties style

Eagle identifies different Hadoop environments as different sites, such as sandbox, datacenter1, datacenter2. In each site, a user can add different data sources as the monitoring targets. For each data source, a connection configuration is required.

Step 1: Add Site

The following is an example which creates a new site "Demo", and add two data sources as its monitoring targets. setup a site

Step 2: Add Configuration

After creating a new site, we need to edit the configuration to connect the cluster. hdfs setup

  • HDFS

    • Base case

      You may configure the default path for Hadoop clients to connect remote hdfs namenode.

        {"fs.defaultFS":"hdfs://sandbox.hortonworks.com:8020"}
      
    • HA case

      Basically, you point your fs.defaultFS at your nameservice and let the client know how its configured (the backing namenodes) and how to fail over between them under the HA mode

        {"fs.defaultFS":"hdfs://nameservice1",
         "dfs.nameservices": "nameservice1",
         "dfs.ha.namenodes.nameservice1":"namenode1,namenode2",
         "dfs.namenode.rpc-address.nameservice1.namenode1": "hadoopnamenode01:8020",
         "dfs.namenode.rpc-address.nameservice1.namenode2": "hadoopnamenode02:8020",
         "dfs.client.failover.proxy.provider.apollo-phx-nn-ha": "org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider"
        }
      
    • Kerberos-secured cluster

      For Kerberos-secured cluster, you need to get a keytab file and the principal from your admin, and configure "eagle.keytab.file" and "eagle.kerberos.principal" to authenticate its access.

        { "eagle.keytab.file":"/EAGLE-HOME/.keytab/b_eagle.keytab_apd",
          "eagle.kerberos.principal":"eagle@APD.EBAY.COM"
        }
      

      If there is an exception about "invalid server principal name", you may need to check the DNS resolver, or the data transfer , such as "dfs.encrypt.data.transfer", "dfs.encrypt.data.transfer.algorithm", "dfs.trustedchannel.resolver.class", "dfs.datatransfer.client.encrypt".

  • Hive

    • Basic

        {
          "accessType": "metastoredb_jdbc",
          "password": "hive",
          "user": "hive",
          "jdbcDriverClassName": "com.mysql.jdbc.Driver",
          "jdbcUrl": "jdbc:mysql://sandbox.hortonworks.com/hive?createDatabaseIfNotExist=true"
        }
      
  • HBase

    • Basic case

      You need to sett "hbase.zookeeper.quorum":"localhost" property and "hbase.zookeeper.property.clientPort" property.

        {
            "hbase.zookeeper.property.clientPort":"2181",
            "hbase.zookeeper.quorum":"localhost"
        }
      
    • Kerberos-secured cluster

      According to your environment, you can add or remove some of the following properties. Here is the reference.

        {
            "hbase.zookeeper.property.clientPort":"2181",
            "hbase.zookeeper.quorum":"localhost",
            "hbase.security.authentication":"kerberos",
            "hbase.master.kerberos.principal":"hadoop/_HOST@EXAMPLE.COM",
            "zookeeper.znode.parent":"/hbase",
            "eagle.keytab.file":"/EAGLE-HOME/.keytab/eagle.keytab",
            "eagle.kerberos.principal":"eagle@EXAMPLE.COM"
        }
      
  • UserProfile

      {
        "features": "getfileinfo,open,listStatus,setTimes,setPermission,rename,mkdirs,create,setReplication,contentSummary,delete,setOwner,fsck"
      }
    

Step 3: Checking the connection

After the configuration is ready, you can go to classification page and browse the data. If the configuration is correct, data will returned immediately.

Any questions on Kerberos configuration, please first check FAQ