Tuning cloudera navigator - cloudera

I'm trying to tune my Cloudera Navigator, the official documentation only gives recommendations about Cloudera Navigator metadata server memory sizing. Is there any complete documentation / post on Cloudera Navigator tuning?
thanks :)

You're probably referring to this.
https://www.cloudera.com/documentation/enterprise/5-9-x/topics/admin_cn_admin.html
Navigator Audit Server - https://www.cloudera.com/documentation/enterprise/5-9-x/topics/cn_iu_audit_server.html
Navigator Metadata Server -
https://www.cloudera.com/documentation/enterprise/5-9-x/topics/cn_iu_metadata_server.html
The Audit Server is controlled by the database you configure with it which is why there are no specific directions for tuning. It all depends on which database you're using on the backend.

Related

Is Kibana Watcher part of AWS

Is Kibana Watcher part of AWS Elasticearch open distro?
It should have been part of Stack management on AWS Kibana UI.
I donot see it. Would like to see if there is a way to enable it.
Watcher is neither available on the AWS service nor Open Distro. It's an Elastic product (see the documentation) and you can either get it on Elastic Cloud or by running Elasticsearch yourself (download).
PS: Watcher is a commercial product, so you will either need a license or use Elastic Cloud.
As mentioned, you can't use Watcher. But ODFE does provide Alerting of its own.
Plugin installation and configuration are documented here

How to configure target url for BPM 8.5.6 Standard?

I'm trying to install IBM BPM 8.5.6 in a linux environment with Oracle database.
Steps I followed to install was
Installed the IBM Installation
Manager using BPM PFS
Installed WAS
and BPM Process Center using The
installation manager.
Created 3 oracle schema for shred db, process
server and performance server
Configured the installation using
sample single cluster process center
file provided by IBM. : using
BPMConfig –create option
The installation was successful and I could see all tables being created. Then I started started it using BPMConfig –start option. That too completed successfully.
I didn't change any ports so it should be using all default ports. Afterwards when I try to access the console like http://servername:9080/ProcessAdmin or http://servername:9080/ProcessCenter or anything i'm getting a 404 error message Error 404: com.ibm.ws.webcontainer.servlet.exception.NoTargetForURIException: No target servlet configured for uri: /ProcessAdmin
Do I have to do anything else? Or what is the starting point or default url to get to process portal or admin console. The WAS admin console is working fine.
Any help is appreciated. Thanks.
Since you probably used custom installation, you have to properly initialize data calling following command:
bootstrapProcessServerData.bat -clusterName cluster_name

Can apache drill work with cloudera hadoop?

I am trying to setup apache drill in distributed mode. I already have cloudera hadoop cluster with a master and 2 slaves. From documentation given on apache drill, its not pretty clear if it can be set up with typical cloudera cluster. I could not find any relevant articles. Any kind of help will be appreciated.
Drill can be installed along with Cloudera on the nodes of the cluster independently - and would be able to query the files on HDFS.
Refer the link for installation details -
https://cwiki.apache.org/confluence/display/DRILL/Deploying+Apache+Drill+in+a+Clustered+Environment
I got this working with cloudera hadoop distribution. I already had cloudera cluster installed with all services running.
perform following steps:
Install apache drill on all nodes of the cluster.
Run drill/bin/drillbit.sh on each node.
Configure storage plugin for dfs using apache drill webinterface at host:8047. Update HDFS configurations here.
Run Sqlline : ./sqlline -u jdbc:drill:zk=host1:2181,host2:2181,host3:2181
(2181 is the port number used by zookeeper.)
It may only work with a rudimentary insecure cluster as Drill currently isn't tested / documented to integrate with HDFS + Kerberos for secure Hadoop clusters. Vote and check back on this ticket for Drill secure HDFS support:
https://issues.apache.org/jira/browse/DRILL-3584

how to use cloudera manager for monitor the components of CDH4

I have already install CDH4 without using cloudera manager. I wanted to use cloudera manager so that i can monitor the different components of CDH4. Please suggest me how to use the manager now.
I have recently had to undertake the same task of importing already installed and running clusters into new Cloudera Manager instances.
I would firstly suggest taking your time to read through as much documentation as possible to fully understand the processes and key components.
As a short answer, you need to manually import all your cluster configurations and assignments into Cloudera Manager so that they can be managed. A rough outline of the plan I used is below:
Setup MySQL instance on NEW hardware (can use postgresql)
Create Cloudera Manager user on all servers (must be sudo enabled)
Setup ssh key access between cloudera-manager server and all other hosts
Useful Docs below:
- http://www.cloudera.com/content/cloudera-content/cloudera-docs/CM4Ent/latest/Cloudera-Manager-Installation-Guide/cmig_install_mysql.html
- http://www.cloudera.com/content/cloudera-content/cloudera-docs/CM4Ent/latest/Cloudera-Manager-Installation-Guide/cmig_install_path_B.html
Install Cloudera Manager and agent/daemon packages on Cloudera Manager server
Shutdown all services using cluster and cluster services
Save the namespace
Backup Meta Data and Configuration files to MULTIPLE LOCATIONS
Ensure the backup can be loaded by starting a single instance NN
Install Cloudera Manager agent and daemon on all production servers
Start the services on the Cloudera Manager server
Access the Cloudera Manager interface
Skip Setup Wizard
Add all hosts to Cloudera Manager
Create HDFS service - DO NOT start the service
Check hosts assignments are correct
Input all configuration file parameters and verify (this means each servers conf files need to be input manually)
Run host inspector and configuration check
Perform the above process for remaining services
I hope this provides a some assistance for you. If you have any other questions I will be happy to try and assist you as much as I can.
Regards,
James
I just recorded a webinar titled "Installing Cloudera Manager in < 30 mins" for Global Knowledge. Available at: http://www.globalknowledge.com/training/coursewebsem.asp?pageid=9&courseid=20221&catid=248&country=United+States (register in the upper right of page). In the video, I install CM on Ubuntu, set up the core components (Hadoop only), and then browse through some of the graphs for monitoring.

Prevent Transport Package being deployed

We are using Tridion 2011 SP1. Content Delivery is in .Net.
We want to make only Transport Package available to Search engine. This Search engine will extract required content/metadata from transport package and will index it. Search engine is installed on different domain/Server.
To achieve this we want to configure a Publication Target which will publish the content to search engine’s server but will not deploy the content. Only Transport Package will be made available in some folder (incoming\success). Can we do it using HttpUpload.aspx and by disabling some settings in cd_storage_config or cd_deployer_conf.
The standard approach for this would be to extend the CD storage process. You can find a good explanation of how to do this at http://www.sdltridionworld.com/articles/sdltridion2011/tutorials/extending-content-delivery-storage-sdltridion-2011-1.aspx
I would argue that #Jeremy's approach is the correct answer, but if writing a storage mechanism in Java sounds too taxing, you could just allow the items to get published to the file system, and schedule a simple script to delete all the files daily/hourly to save disk space.
This would require no integration effort with SDL Tridion.

Resources