Problem INS-35423 on installing Oracle 11g RAC (empty cluster nodes) - oracle11g

I am trying to install Oracle 11g RAC for training purposes on a CentOS 6.9 machine.
I have succesfully installed the grid and clusterware services and have two nodes (rac01, rac02)
The following does not report any serious problem
./cluvfy stage -pre dbinst -n rac01,rac02
As a matter of fact the only problem reported is a missing pdksh package (which is not a real problem) and the fact the pool of NTP servers used by the nodes return different IP addresses for each node (to be expected since the pool does not always return the same IP address).
Similary the following reports that clusterware services are up and running
[root#rac01 bin]# ./crsctl check cluster -all
**************************************************************
rac01:
CRS-4537: Cluster Ready Services is online
CRS-4529: Cluster Synchronization Services is online
CRS-4533: Event Manager is online
**************************************************************
rac02:
CRS-4537: Cluster Ready Services is online
CRS-4529: Cluster Synchronization Services is online
CRS-4533: Event Manager is online
**************************************************************
I am trying to install the database as the oracle user but when the time comes to select a RAC installation no nodes are reported.
Does anybody have any clue what other possible problems may exist and how/where to look?

I have no idea why the following worked (someone else may explain it) but I re-run the grid installer from each of the nodes as follows
[oracle#rac01] rac01$ /u01/app/11.2.0/grid/oui/bin/runInstaller -ignoreSysPrereqs -updateNodeList ORACLE_HOME=/u01/app/11.2.0/grid "CLUSTER_NODES={rac01,rac02}" CRS=true LOCAL_NODE=rac01
[oracle#rac02] rac02$ /u01/app/11.2.0/grid/oui/bin/runInstaller -ignoreSysPrereqs -updateNodeList ORACLE_HOME=/u01/app/11.2.0/grid "CLUSTER_NODES={rac01,rac02}" CRS=true LOCAL_NODE=rac02
and afterwards I re-run the db installer and the rac nodes appeared in the list

Related

Is it possible to have 2 BizTalk Server nodes running off of a single MessageBox?

Myself and my colleagues are currently in the process of upgrading our BizTalk environment to BT 2020 from BT 2013R2 and as part of this we are intending on setting up two BizTalk servers so that we can have host instances running across both of them. We do not, however, need more than one MessageBox DB based on the load we see, and after looking online, there doesn't seem to be a breadth of information.
Is it possible to have our BT setup to have 2 servers running off of a single MessageBox and is it complicated to configure?
It’s the basic feature of BizTalk to have multi servers group connected to same message box. On your second computer, when you configure BizTalk using BizTalk Configuration Wizard, you choose option to join existing group and you should select your existing dbs to join.
Microsoft Docs Install BizTalk Server in a Multi-Computer Environment

Unit cinder-api.service could not be found - Openstack

I installed openstack devstack on ubuntu 18.04
when I login to horizon. it logs in.
when I try to check status of all services. Services are not found.
So when I execute service cinder-api status it gives me Unit cinder-api.service could not be found. (Same for all services. Installation was successful. although it was interrupted due to network. But is that issue with this (But horizon dashboard opens))
What is wrong with this. As because of this I am not able to create volumes (I guess).
Reason for the message:
The services of the projects are down. For example - n-api, c-api, c-sch..... To rectify it, manually trigger the services using command prompt or Run unstack.sh and again clean.sh and then stack.sh, this will create a new openstack.
Note: While using devstack one important thing need to be taken care is "Never shutdown the machine. If it is shutdown all the services goes down."

Kaa cluster architecture

I am not really understand with Kaa cluster architecture. First is i need to install and configure Kaa components on a single Linux node by using this link: http://kaaproject.github.io/kaa/docs/v0.10.0/Administration-guide/System-installation/Single-node-installation/
I need to install SQL, NOSQL and Zookeeper in it. Does it means this single node is actually a cluster? i want to implement scalability and high availability. Do i need to clone the single node to implement fail over process?
The Kaa cluster architecture is:
http://kaaproject.github.io/kaa/docs/v0.10.0/Architecture-overview/
To setup and configure Kaa cluster you should follow the instructions on the Kaa Cluster setup documentation page. The Single Node Installation page describes what Kaa dependencies should be installed and how they should be configured, but as it for a single node installation, they all are placed at single node and configured respectively.
The cluster setup is most like single-node installation, but requires more nodes and configuration for their correct cooperation.
Thus, the difference between Kaa cluster and single-node operation is generally in configuration of the components rather than in the components themself.
Therefore, you can clone a single-node Kaa server as a basis for a cluster node, but you will need to change its configuration accordingly before it can operate as a cluter node correctly.

how to get instances back after reboot in openstack

After successful installation of devstack and launching instances,but once reboot machine, need to start all over again and lose all the instances which were launched back then.I tried rejoin-stack but did not worked,How can i get the instances back after reboot ?
You might set resume_guests_state_on_host_boot = True in nova.conf. The file should be located at /etc/nova/nova.conf
I've found some old discussion http://www.gossamer-threads.com/lists/openstack/dev/8772
AFAIK at the present time OpenStack (Icehouse) still not completely aware about environments inside it, so it can't restore completely after reboot. The instances will be there (virsh domains), but even if you start them manually or using nova flags I'm not sure whether other facilities will handle this correctly (e.g. neutron will correctly configure all L3 rules according to DB records, etc.) Honestly I'm pretty sure they won't...
The answer depends of what you need to achieve:
If you need a template environment (e.g. similar set of instances and networks each time after reboot) you may just script everything. In other words just make a bash script creating everything you need and run it each time after stack.sh. Make sure you're starting with clean environment since OpenStack DB state remains between ./unstack - ./stack.sh or ./rejoin-stack.sh (you might try to just clean DB, or delete it. stack.sh will build it back).
If you need a persistent environment (e.g. you don't want to loose VM's and whole infrastructure state after reboot) I'm not aware how to do this using OpenStack. F.e. neutron agents (they configure iptables, dhcp etc) do not save state and driven by events from Neutron service. They will not restore after reboot, so the network will be dead. I'll be very glad if someone will share a method to do such recovery.
In general I think OpenStack is not focusing on this and will not focus during the nearest release cycles. Common approach is to have multi-node environment where each node is replaceable.
See http://docs.openstack.org/high-availability-guide/content/ch-intro.html for reference
devstack is an ephemeral environment. it is not supposed to survive a reboot. this is not a supported behavior.
that being said you might find success in re-initializing the environment by running
./unstack.sh
follower by
./stack.sh
again.
Again, devstack is an ephemeral environment. It's primary purpose for existing is to run gate testing for openstack's CI infrastructure.
or try ./rejoin-stack.sh to re-join previous screens.

Machine's uptime in OpenStack

I would like to know (and retrieve via REST API) the uptime of individual VMs running in OpenStack.
I was quite surprised that OpenStack web UI has a colon called "Uptime" but it actually show time since the VM was created. If i stop the VM, the UI shows Status=Shutoff, Power State=Shutdown, but the Uptime is still being incremented...
Is there a "real" uptime (I mean for a machine that is UP)?
Can I retrieve it somehow via the OpenStack's REST API?
I saw the comment at How can I get VM instance running time in openstack via python API? but the page with the extension mentioned there does not exists and it looks to me that this extension will not be available in all OpenStack environment. I would like to have some standard way to retrieve the uptime.
Thanks.
(Version Havana)
I haven't seen any documentation saying this is the reason, but the nova-scheduler doesn't differentiate between a running and powered off instance. So your cloud can't be over-allocated or leave an instance in a position that would be unable to be powered on. I would like to see a metric of actual system runtime as well, but at the moment the only way to gather that would be through ceilometer or via Rackspaces StackTach

Resources