My work has strict data policies that don't allow data to be stored locally. I am partly doing a project in SAS but the last step would be much much better in R.
Normally I use Windows Remote Desktop Connection to connect to a remote server and then use WinSCP to connect to the shared drive from that server where the data is stored, and do SAS from that server, which has access natively to the remote server's shared drive.
I have R Studio locally installed, and installed on the server, but no idea how to point it to the shared drive folder I need.
I have the folder address, the computer name, and my credentials.
I have tried mapping a network drive to the desired folder from the remote server.
I have included both the full server path and the \computer_name\full\file\path
(The computer name is of the form blahblah.vsp.sas.com)
I have tried searching stack exchange and Google.
Related
I have now a working Website that I have deployed to the IIS on my local machine using Visual Studio 2013. I can access the website successfully from other machines in my office and in neighboring offices. I have even had others check the access from geographically remote locations, and still been successful. The access method is to put my machine's IP Address and add the name of the application, ###.###.###.###\name, into the address bar of a browser.
The next step is to promote it to production. We are not using Azure, so the tutorials for promotion to production I have found aren't useful, nor are we using 3rd party providers.
The method I have conceived is as follows:
Have an admin directly login into a company server
install the database software
install Visual Studio
install IIS
copy the local machine's db to the server db
copy the locally deployed files to the server
admin login into Visual Studio
deploy the site on the server in the same way it was deployed on the local machine
use the server IP or update the host file on all networked computers to map the IP to an appropriate name (or the local network has a local DNS we can update)
This would allow me a DEV environment (the VS IDE), a TEST env (locally deployed version), and a PROD env (the version deployed on the server).
I don't see any reasons why this wouldn't work, maybe a bit tedious, but workable.
Is this method ok? Am I missing anything critical?
No, this is an altogether inappropriate way to push a build to a production environment.
Your source code should be stored in a source code repository. You should have an automated, continuous build server pull from the repository and complete the build in a dedicated environment that is itself under change control. The build should include the generation of installation files, e.g. a click-once deployment package. From there you can have an admin run the deployment package, or, ideally, you'd push it automatically with a tool like Octopus.
The above, honestly, is the bare minimum for a commercial web site. There is much, much more you can do to make things even more robust, e.g. blue-green deployment.
Note that none of this involves installing development tools like Visual Studio on your server. The server should stay as clean as possible, running the fewest applications that you need, to minimize any sort of attack surface and to keep the machine running efficiently.
Dont Put Yours Machine IP, Decompilers Can Take It And BruteForce Your IP To Remote Acess Your Machine.
I have installed R studio on my local laptop and trying to access files located in AWS server (Windows).
I do not want to use FTP protocol.
What are other possible ways to remotely access the files located on a remote server?
How to use SCP/SSH protocol to access files on AWS server?
My objective is to work and analyze log files generated by a remote server.
I am looking to find a way to get through to access the files first. Once my first hurdle is achieved, later I'll import the files into spark environment using "sparklyr".
Kindly help me get access to these files
maybe the question title is a little bit weird... I will explain more here...
Basically, suppose that I have a local laptop and a remote machine. I can of course run R tasks on the remote machine itself. Now I wonder if I can connect the remote machine from my local R and run the tasks on the remote machine while I still call these things in my local machine... that is I don't need to open R or type/source R codes on the remote machine.
One related thing is that I am considering to develop a small app with R-shiny. As far as I understand, though I just started looking into it, one can have R installed the local machine and use the app - the calculation will be done in R on the same machine. This question is essentially that if it is possible to add a feature of “connection to a calculation back end” in the app, i.e. by connecting to the remote machine, the task is done on the remote machine when one uses the app on his local machine.
You can do this with RStudio Server. It runs RStudio inside a web server on the remote machine. You connect to it from the web browser on your local laptop, so you can still enter your R code on the local machine.
RStudio Server can be downloaded here:
https://www.rstudio.com/products/rstudio/download-server/
In my ASP.NET application I have ftpdata folder, there is an MS Access 2010 file e.g. somename.accdb.
On my local computer (Win XP), there is no problem to connect with OleDb to this file and read write datas into and from tables.
But on server (Windows Server 2008), I cannot connect to the same file like on localhost.
Can you help me, please, where problem could be?
Finally, I found solution of problem.
There was not installed 2010 Data Connectivity Components on the server.
The windows user that your application pool is running in does probably not have read access to the directory/file
You need either direct access through the file system or SMB networking access (via a share defined on the server). If the FTP folder has an SMB share name, use that to get to it, but it's likely not shared (since it's purpose is for FTP access, not SMB access).
Installing the 2010 Data Connectivity Drivers does not seem to work, what worked for me was to set the corresponding website application pool in IIS to enable 32-Bit applications.
Official Microsoft Reference:
A forthcoming project requires that a Flash projector (.exe) be kept on a server and accessed by multiple users over network (LAN). There will be a central access database as a back-end. I am using a third party SWF2EXE (SWF Studio) product to have database functionality. I would like to know if sharing an .exe file this way is practical. I know it's not a typical client-server methodology but I can't use Server side platform like ASP due to lack of resources at the client's place. Is it likely to cause any crashes if the same .exe is launched simultaneously by network users doing a simultaneous read/write to database (mdb)?
Each computer should get a copy of the exe file, and it should be only the database which is shared. If you are on a LAN and you are using MS Access, placing the MDB file in a shared location should be fine.
You may also want to check what the concurrent user limit is for an ms access database.
Andrew