BMC Control-M - can it manage remote offsite servers - control-m

Our organisation uses Control-M to manage all the scripts within our network. However, we have recently entered a contract to have a new application hosted for us. It is an ecommerce application and has scripts run through Cron jobs.
What I'd like to know is that whether Control-M has any offsite agent functionality so that it can be installed on this remote site and then in some way keep in communication with the Control-M server we have within our infrastructure so that we can monitor the scripts along with the rest of our applications.
Thank you.

Control-M supports Agents anywhere as long as you have network connectivity. The default Server/Agent ports are 7005-7012 inclusive, so those will need opening on any firewall.
If you don't want the remote site to use the full Agent (which will need a local install) then you can use the Agentless option (works via WMI on Windows or SSH for Unix) and only needs defining on the Control-M Server side.

Not sure about you question. Control-M server works with tcp/ip to contact its agents and does not care if it is connecting to a server in the same datacenter or on another continent.
Install a VPN between you and your hosted application and install a control-m agent or setup an agentless scheduling using ssh and you are good to go.

Related

Web interface for SFTP server

An SFTP client such as CuteFTP or Filezilla provides a rick user interface for an SFTP server. These are clients that are installed locally on the user's PC. Instead of a client installed at the user's side, is it possible to set up a web-based user interface on the SFTP server, so that a user with only a browser is able to access the files on the server? Are such open source or commerical products available that can be deployed on the SFTP server for enhancing the file transfer experience?
Note: The base server needs to be SFTP as there will be scripts that clients will be using to transfer files in a non-interactive manner. For interactive usage, I am looking for a web interface that be serve as an add-on.
I suggest you check out: https://filebrowser.org/features
There is another good one called "droppy" but its no longer active and apparently has too many git forks at the moment to know where it will go.
The question you need to answer is the application value/scope.
You can always go with Dropbox or something too depending on the user's intended application.

Azure DevOps Pipeline Task to connect to Unix Server and execute commands

I am seeking to set up a Release Pipeline in Azure DevOps Services that will deploy
an application to a Unix server, where it then executes some unix commands as part
of the deployment.
Would appreciate some guidance on which pipeline Task(s) I can set up to therefore
achieve the following objectives:
Connect to the Unix server.
Execute the required Unix commands.
By the way, the Agents are currently installed on Windows hosts but we are looking to
extend that to Unix servers in due course, so a solution that fits both setups would
be ideal, even though the former is the priority.
You can check out task SSH Deployment task.
Use this task to run shell commands or a script on a remote machine using SSH. This task enables you to connect to a remote machine using SSH and run commands or a script.
If you need to copy files to the remote linux server. You can check out Copy Files Over SSH task.
You probably need to create a SSH service connection. See steps here to create as service connection.
In the end, due to concerns raised about the install of private keys on the target server which is part of the SSH Deployment setup, we opted for the use of Deployment Groups which has enabled us to set up a persistent connection to our Linux server.

Can I arrange all my apps on Heroku to run on the same LAN?

I'm looking to host a Python webapp on Heroku. The backend is written in different Java programs (the servers) and ZeroMQ is used for communication. If I deploy everything on Heroku (the servers just need to be running listening to requests), would they all run on the same LAN?
I'm asking because I don't want to run ZeroMQ over the internet, cause then I'll need something like SSH tunneling which I want to avoid.
Thanks!
Dynos do not share the same network. See this article on Heroku's website. This is by design for resiliance but you can choose the (Amazon) region that they all reside in.

Error in communicating LabView Client VI with WinCC OPC Server over the network (LAN)

I'm trying to automate some system using Siemens PLC; for that we have developed a WinCC Flexible RT application for control and monitoring, and have also declared the PC based HMI panel as the OPC server. Which is working fine.
Now, when i try to create a OPC Client application either in WinCC or Labview, on the same PC, am able to access all the tags of the server on the client, am able to do so.
When i try to create the client application on some other computer connected over the local area network [perfectly fine and well configured network working], am unable to access the OPC server/tags in that case.
What can be done in this case.?
The easiest way to establish connection between OPC client and server applications is to perform the following:
turn off the firewalls on both machines
create the same windows account on both machines using the same login name and password
for windows account (pointed out in 2) allow all "access permissions" and "Launch and Activation Permissions" by running "dcomcnfg" command and navigating to "console root->component services->computers->my computer->com security"
The more detailed example of how to do steps described above is given in this article http://opc-da.com/Winsetup.html
Also, there is a powerful OPC client API implemented on LabVIEW using OPC .NET API with detailed documentation on how to use: http://opc-da.com
It would be helpful if you mentioned the OS on the server and the client computers. Anyway, I'm guessing that this is a DCOM issue, as #Ken mentioned.
It's been a while since I worked with this sort of thing, but you can try this article: http://www.opcactivex.com/Support/Tutorials/DCOM_Tutorial_-_Configuring_th/DCOM_XP/dcom_xp.html and see if it helps.
Make sure the firewall isn't blocking the connection. Disable it on both sides.
Take a look at this NI article: http://zone.ni.com/reference/en-XX/help/371361G-01/lvhowto/psp_firewall/
keep in mind, the tcp ports used by OPC/shared variables changed somewhere between LV2009 and LV2011, so depending on what version you have, the ports may be different.
In addition to the ports and the firewall...
Make sure that the account in which the the OPC client is running has access to the OPC server, either through the DCOM default settings or the OPC Server-specific settings (also part of DCOM configuration) on the server machine.

azure connect between azure myWebRole and nonazure server not working

My webapplication hosted on windows azure, needs to communicate with TFS Server. When any one login to my web app using live id, I want the logged in user to use my Team foundation server(TFS) credentials -username,password and domain to programatically authenticate and connect to our TFS server and create some work items.
I configured my azure connect for the communication to happen between azure WebRole and TFS server (our TFS is non-azure ).I added both the WebRole and the TFS Server into single Connection Group
In my azureportal ,I can see mywebrole and my TFSServer as connected the machine endpoint is active, and that it refreshes since the last connected updates
.But when I try to run my web application from azure and when it tries to communicate with our TFS server ,its throwing error message saying Error message : Team Foundation services are not available from server eg.,http://xyz-abcxyx-01:8080/tfs/eas/. Technical information (for administrator): The remote name could not be resolved: 'xyz-abcxyx-01'
Any suggestions to resolve this issue ?
You should enable remote desktop on your WebRole and connect to one of your instances. Then, try to ping the IP of your TFS server (not the hostname xyz-abcxyx-01). Maybe this is simply a DNS issue (even though using hostnames works with Windows Azure Connect).
If pinging the IP works, but pinging the hostname doesn't work you have a few options left:
Use the IP instead of the hostname. This won't work if you configured your TFS to use host headers.
Create an elevated startup task to modify the hosts file and map the IP to the hostname. In your code you can keep working with the hostname.
Try to modify the DNS server configured in your WebRole to use the default DNS server + your internal DNS server. But to me this doesn't look like a clean solution.
Anyways, in each solution you'll want to store the IP/hostname in the ServiceConfiguration and make sure your code supports changes to the ServiceConfiguration. This will allow you to change the IP/hostname without having to redeploy.
You should check if TFS server is listening on all network interfaces, include the one created by Azure Connect (start with 2a01). Next try to connect to TFS from a machine on the local LAN, just to make sure it is configured correctly. You don't need to use IP for referring to TFS, DNS name is definitely supported out of box.

Resources