I am trying to install openstack on a single node server.
I need to access Instances from internet.
I am new to openstack, so I spent some time trying to get it work correctly but without success. I tried devstack but it is not persistent after reboot.
For microstack, it is not configurable.
I need to assign Public IPs to instances. I have 2 physical networks. I tried with external network, but I don't found an option how to do that.
Did anyone successfully installed openstack on a single machine, and is there a way to expose instances to bring them puclic IPs from a pool.
Thanks in advance.
Related
I installed openstack using packstack. But I don't know how to configure network.
I want to access my instance using ssh like this.
Is it possible to do this?
By default, Packstack creates an "external" network that is entirely contained inside the host. This is great for creating a cloud without any knowledge of the network environment, but prevents you from accessing instances from outside.
Configure your Packstack so that its external network is your network. This is documented at https://www.rdoproject.org/networking/neutron-with-existing-external-network/.
Actually, without external network its not possible.Your instance should have external(public) ip.I'm using this way but you have other option like proxy.
if your Centos7 machine can connect your instance;
Use HaProxy at Centos machine and proxy instance ssh port to specific centos7 port and connect using this port.I'm using this way on microstack.
I'm trying to set up an Openstack environment with two Kubernetes clusters, one production and one testing. My idea was to separate them with two networks in Openstack and then have a VPN in front, to limit the exposure through floating ip:s (for this I would have a proxy that routes requests into the correct internal addresses).
However, issues arise when trying to tunnel requests to both networks when connected to the VPN. Either I choose to run the VPN in its own network or in one of the two, but I don't seem to be able to make requests across network boundaries.
Is there a better way to configure the networking in Openstack or OpenVPN, so that I can keep the clusters separated and still have access to all resources through one installation of OpenVPN?
Is it better to run everything in the same Openstack network and separate them with subnets? Can I still have the production and test cluster expose different IP-addresses externally? Are they still separated enough to limit the risk of them accessing each other?
Sidenote: I use Terraform to deploy the infrastructure and Ansible to install resources, if someone has suggestion in the line of already prepared scripts.
Thanks,
The solution I went for was to separate the environments with their own networks and cidr and then attach them to the VPN instance to let it get access to them. From there I just tunnel everything.
I'm trying to install PacketFence on a virtual machine, and have dedicated 8 IPs to the server.
They all are static on eth1:0 -> eth1:7, so not eth1, eth2, etc.
When trying to configure vlans for "normal", "registration" etc. in Packetfence I get an error saying they need to be on different networks.
Can't I install Packetfence without physically separate networks?
The static routes use the same primary IP assigned to the device, so it might be a good idea to provision a portable subnet for your purpose.
Nevertheless, you could submit a ticket to SoftLayer, so that they might provide you further assistance.
I have a couple of virtual machines in one Cloud Service. They are assigned to the same VNET and have received private IP addresses in the same subnet.
I noticed that I was unable to PING from one server to another and when I started to look into it there is no connectivity whatsoever between the servers. I have disabled windows firewall on both servers but that didn't do the trick.
Just now I tried on one of the vm's to ping the internal ip address assigned to itself but it fails.
Can anyone shed some light into this? Is this expected behavior?
The reason I am looking into this right now is because we are adding a third VM to do some performance monitoring and since the other two VM's are part of a Cloud Service we cannot open endpoints to both of them using the same port and need to go directly to the internal IP's.
Thanks in advance
I had a similar issue not too long ago. I had three servers in the same vnet that were able to communicate via site-to-site VPN to my HQ but could not communicate with one another. After several hours of banging my head against the desk, I ended up just re-building the vnet and connectivity to one another was restored successfully. The vnet router feature had become corrupt and could no longer send traffic internally.
To rebuild the vnet, you'll need to delete the VM's. Keep the disks though, and you can re-build them quickly after the new vnet is back online.
Here i the scenario:
I am using MacbookPro as host
Using VBox for creating / using VM's
Using ubuntu as VM , with the installed software which needs access to my work (over VPN, of course)
Have VPN Connectivity from Mac, and able to ssh to hosts # work
I am unable to access the same hosts when logged in to the VM on my Mac.
The company network is 10.x.x. (as the case with most companies)
I am using comcast at home, which defaults to a 192.168.x.x type of address
When using the same setup # work, i am able to have access to work computers , but not when i am at home. Any clues to what i need to do to get access from within a VM?.
Thanks,
Since the VM instance is an entirely separate system I don't think you are going to be able to do this, as the Mac has no knowledge of the VPN that you are connected to.
You might ask this over on SuperUser.com as well, it might get more answers.