Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed last month.
Improve this question
I have set up a VPN on a home server (I am using Wireguard in a docker container running on open media vault). My goal is to connect to my network shares from my windows Laptop when I'm on the go.
I set up the VPN and I was able to connect to OMV, both via PuTTy and the web interface, from my Laptop (while not connected directly to my home wifi).
However for some reason windows cannot detect any of the network shares I have set up on OMV
I tried to modify all the windows setting that came to my mind and googled a fair bit but I'm out of ideas.
Looking forward to any possible solutions.
Edit:
I can access the share without a problem when I am directly connected to my home network.
I just tried to access the folder, while connected through the VPN, by typing the full path into the "adress bar" in file explorer, By doing this I managed to connect to the folder and use the files.
But I still can't map the networkdrive... So it really seems to be a settings problem, but I still have no idea.
All of my the other devices that are on the network aren't showing up either while connected through the VPN.
One possible issue that I once ran into, is that the latest versions of windows stopped supporting SMB 1.0. Fortunately, you can explicitly re-enable them.
In the start menu, open "Turn windows features on/off".
A pop-up will allow you to enable also version 1.0 of SMB. I would suggest to enable the "Client" and "Automatic Removal", but you won't need the "Server".
That could possibly solve the problem, assuming that your other devices use older SMB versions.
You may also want to check these settings ...
The Solution was btw. pretty simple:
I was able to map the drive after manually entering the ip adress to file explorer. (in the format \192.168.x.xxx)
Related
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 7 years ago.
Improve this question
I want to build an IT setup. At the very basic I want to setup few things
I want to have a central remote server which would have enough processing power to host everything. The local PCs should be on the economical end.
I want that all code and documents remain secure. None of my employees should be ableto take away with code or confidential documents.
The applications I would be working on would be mostly based on Java. I am thinking about the following setup. But I do am not sure about the feasibility or practicality of implementing it.
Buy a windows dedicated server from a hosting website. [Probably a Windows Server 2008 R2 machine with around 16GB or 32 GB RAM]
Use economical windows desktops in the organization. [Probably Dual Core or Core I3 PCs.]
Host all code and applications/programs on the dedicated remote server.
Employees access the remote machine from their windows desktop using Remote Desktop (RDC). Do all coding and run applications on the remote machine only.
All internet access should be disabled on the remote server. So that there is no option to upload the code anywhere onto the internet.
Internet access should only be available on local machines.
Copy and Paste from the remote machine to the local machine should be restricted for all users other than administrator.
Copy and Paste from the local machine to the remote machine should be allowed.
Need suggestion if the above would work fine. I am actually in doubt about few things:
I have come to know that the RDC to remote server would require purchase and installation of windows terminal services TS licenses, so that multiple concurrent user login is possible. Is this correct?
If license cost could be deterrent for a startup would it be better to go for a linux setup, where I can access desktop and programs using SSH. I suppose the latter would not incur much costs as is in the case of buying TS licenses.
Is it possible to setup such copy paste restrictions, which can only be managed and changed by an administrator.
Do I need to setup VPN also? I think this might be needed so that the remote machine would only be accessible from my local office environment and not any other place. I mean my employees or any other people should not be able to access the same remote machine from any other network.
These is the typical setup of any large Software company. All these requirements are realistic. Get an experienced person/company to configure this environment. Any small mistakes in configuration might do a lot of damage to your Intellectual property.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I have upwards of 30 Dell GX2xx models doing nothing. I've decided on using them to build a cluster, but I am lost as far as getting started. I've used ClusterKnoppix, and even straight OpenMosix in the past, but those projects are, very sadly, dead now.
I've checked out Xen, to an extent. I don't know if Xen is the solution I need. I'd like to have the ability to spin up a few VMs (when I need) in a server pool, with the VMs having the ability to run just off of resources in the pool, leaving me without the care of which node they run on.
I need some insight here... Thanks all!
Xen is not itself going to manage the whole cluster.
It will act on every single machine to instantiate/manage/delete the VMs.
You can have a look at Eucalyptus if you want to build that kind of private cloud solution with open-source software.
I would also recommend that you use OpenStack that tends to be the successor of Eucalyptus.
Have you checked out XCP, I find that it's really easy to start up a virtual cluster with this software.
Have you looked at such projects as OpenAIS, Corosync, DRBD and Pacemaker? They are all apart of the Linux High Availability project (http://www.linux-ha.org). They offer many different configuration options for numerous types of servers (IE. MySQL, Apache, Xen, etc.).
They have custom scripts (LSB and OCF) that are ran in place of your standard init scripts and assume the roles of these init scripts. I have included a detailed guide for setting up a Xen HA cluster on OpenSuse 11.1 below for your reference. The configuration of the Linux HA components should be the same from distro to distro, except that the package names to be installed will vary, as well as the location of the specific configuration files. The command line tools should be the same as well as functionality. Hope this helps.
http://www.howtoforge.com/installation-and-setup-guide-for-drbd-openais-pacemaker-xen-on-opensuse-11.1
As a heads up, I have some experience programming on Windows and other devices, but I have almost zero experience doing web anything, so I'm sure this is simple/easy to find online and I just don't know what I'm doing.
I have a computer on my home network that is connected to the internet(I can VNC into it from online if that helps) and I want to set it up so that I can connect to this PC online and access a web page stored on the PC.
This will be used by two people tops, so I'm not concerned about the number of connections or that kind of thing, I just want to be able to look at this page from online. I'm not sure what to call this, but I guess I want to have own webpage on my home PC that I can access to do stuff remotely.
Basically, I want to be able to go to http://{my-pc-ip}/webpage.html and see it online. If it helps, this is largely a learning exercise for me, I want to experiment and play around with what I can do through a webpage on my home network, through an online web page interface. Like maybe start a program on my home machine using a button the page, but do this from a computer connected over the Internet(this stuff I'll figure out myself, I just don't know how to set up the online stuff).
this is quite simple. You can use either IIS which is shipped with your XP Pro or use free and open source solutions:
XAMPP - http://www.apachefriends.org/en/xampp-windows.html. I've been using this for years. Very simple. The last step of installation is just to secure MySQL and FTP (if needed at all). PHP is shipped so you can start doing web development without extra hassle.
Lighttpd - http://www.lighttpd.net/. Another FOSS webserver, which is very light-weight.
There are a few more but Apache is the most popular so you can just go ahead with XAMPP which is very mature and has large community of users.
Lastly, remember to relax your firewall allow access to your IP address from LAN & VNC (whatever you needed). Some materials to get things done:
http://www.microsoft.com/windowsxp/using/networking/security/winfirewall.mspx
http://support.microsoft.com/kb/875357
http://technet.microsoft.com/en-us/library/bb877979.aspx
http://members.shaw.ca/nicholas.fong/vnc/
Enjoy doing web!
Not really programming related, but none-the-less, WAMP is by far the easiest solution out there.
http://www.wampserver.com/en/
To expose your PC as a web server you need to do couple of steps:
Have external ip (static).
Configure the firewall to allow incomming connection to your PC to port 80 (if you need SSL then port 443 too).
Set up a web server: you can use IIS (if WinXP is not Home edition), Apache server or smallest possible HTTP server like ihttpd.
Put the pages into your root directory.
That is the basic explanation of the steps to do.
For basic house-holding tasks like temporarily sharing static files on a lan i have used HFS which is only 600K, has a gui, adds itself to the explorer context menu, and yes, it is free with source code available.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 8 years ago.
Improve this question
I'm operating a neighbourhood WIFI network in a rural environment.
Now I'm looking fo a monitoring tool to run on a server (Windows or Linux) which would track bandwidth, uptime (clients as well as internet connection), etc...
Most of this information is exposed via SNMP by my routers and access points, so SNMP support is required.
Additional features should be:
Graphical data representation
free license
So what's the best choice for me?
Edit These are the tools mentioned so far:
MRTG
Munin
Nagios
Zenoss Core
ntop
cacti
ZABBIX
MRTG is probably the easiest to setup. If your router has SNMP (as you mention), to setup it's a single command:
cfgmaker --output=mrtg_myrouter.cfg public#1.2.3.4
MRTG is good for high-bandwidth routers and the likes. It's not great for other data (it can be coerced into graphing most things, but it's a little unintuitive to setup)
For monitoring other stuff I like Munin. I would describe it again, but I posted an answer a while ago here (about graphing disc-usage).
Munin can of course graph network usage, and easily pull data via SNMP (in fact it's the recommended setup for grabbing data from Windows-based servers - run a SNMP daemon on the Windows machine, and have Munin connect to this). The graphs are also prettier than MRG, I would say (clearly the most important factor..)
There's an example installation of MRTG here, and Munin here
IMHO, Cacti is easiest to install and use.
Zabbix is interesting, but harder to use.
And here is a very comprehensive list of all network monitoring tools.
Not sure if this fits your usage but a lot of web hosting provider uses Nagios for network monitoring
Zenoss Core is free and open source. It keeps RRD graphs (like other monitoring tools mentioned here). To monitor parameters other than basic network bandwidth (and up state), the switch or router SNMP definitions and MIBs should be available as a ZenPack. Runs on a Linux (virtual?) server. Uses Google Maps to display link status.
I have been using ntop it is free on linux and for purchase if you want a windows binary and worked pretty well for us
I had the same question last week and tried several options.
For basic snmp graphing needs, cacti is great, but graphing apache, mysql, etc. is a bit too hard I think.
ntop is also a nice tool, but has a different usecase than the other ones in your list.
You should look at Zenoss. The Core version is FOSS, userfriendly, and very powerful. I had no need for the Enterprise version, but your needs may differ.
It does graphing, monitoring and alerting of all the basic stats, but download some ZenPacks and you can easily add Apache, MySQL or many other stats. All configuration can be done via the GUI. The interface is clear and responsive and allows for easy management of very large networks.
In short, I'm glad I never spent much time on Nagios, because I believe Zenoss is the best option available.
Also consider CactiEZ on a VM or small server, it is a baremetal CentOS 6 based system.
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 6 years ago.
Improve this question
At my job we make & sell websites. Usually we install our .NET C# based site on a customer's server and maintain and support it remotely. However, every once in a while, for bigger development works and just to make things simpler (and faster!), we will copy the site to a local server.
This is great, but has one pain - moving the site back to the customer. Now, If nothing was change on the customer's copy - no problem. However, it is the sad truth that sometime (read more often than I would like) some fixes were needed to be applied on the production server. Either because the customer needed it NOW or simply because it was major bug.
I know that you can easily apply those bug fixes to the local copy as well, but this is an error prone process. So I'm setting my hopes on a distributed version control to help synchronize the two copies.
Here is what I need:
Easy to install - nothing else needed except the installer and admin rights.
Can integrated in an existing website as a virtual directory and works on port 80 - no hassle with new DNS required.
Excellent software
That's it. Any ideas?
Some comments on the answers
First, thanks! much appreciated.
I've looked at Mercurial and Bazaar and both look very good. The only caveat is the installation as a virtual directory on IIS. Mercurial, as far as I understand, use a special protocol (wire) and Bazaar needs and addition of python extensions. Is there another system which is easier to integrate with IIS? I'm willing to take a performance hit for that.
I'd look at either Mercurial or Bazaar. I'm told Git also works on windows, but I suspect the windows port is still a second class port at best.
You'll probably need to be able to run python scripts on your webserver to host either of them.
Maybe not exactly what you request but checkout DeltaCopy which is a windows version of rsync. You can also read about another rsync solution here
I can also vouch for Mercurial. Simple to use and powerful to boot!