I have a new computer in the same network as the old one. In the new one, I first installed RStudio, then the latest version of R (I hope the order does not matter). Now, the new R installation gets "unable to access index for repository" error for every library. Simple task becomes hours of googling, during which I tried the following:
tick or untick use secure download method for http in global options (was ticked in the old computer)
try different, close my location repositories in global settings
copy the folder of an existing package (like ggplot2) from the old computer to the new one's library folder (does not become visible in packages pane)
uninstall and reinstall R
compare installations (the only difference is that in the old computer, folder RTools is under c:/program files while in the new one I accepted the standard location and folder (c:/Rtools40)
Please note that both computers are in the same corporate network, so the firewall and other network settings are the same.
Finally I found the reason: our company uses an internal repository that mirrors an external one. After inserting a custom repository, everything worked perfectly.
I hope this helps if someone runs to a similar problem.
Thanks everyone for your comments!
Related
Here is a example situation: I wish to test if a new user of my R package will be able to install. If it will demand any additional configuration that I have in my developer environment and the user do not have.
I would like to see possibilities. I have already tried Docker and Virtual Machines. But I guess that exists easier ways to accomplish this!
Edit:
What I want is possible to do in Virtual Machines/Docker!
I made a brand-new system to check what errors can I get when installing the package in this new environment.
What I really want here is to know how to do a new configuration of a new environment in my own PC, and not in a virtual machine like system.
if(!require(somepackage)){
install.packages("somepackage")
library(somepackage)
}
Maybe this is some of your solution in the packages question
Link of the info:
Elegant way to check for missing packages and install them?
I have been using R on commandline (BASH). I am unable to access the internet (download any packages). I have tried proxy system wide, and tested it with wget, which works. The "install.packages()" command however does not.
Per some user's advice, I also tried setting the proxy in .Rprofiles file. That didn't help either. Please advice.
I recently ran into the same issue on my work machine. Our Firm uses Cylance as its antivirus software. Cylance was quarantining the file "internet.dll" that R uses to access the Internet. Fortunately, however, it only does so in the 32-bit version of R. For me, there were two solutions:
First, I was able to download packages directly from the 32-bit version of R (outside of RStudio). This works fine. The downloaded packages will run in 64-bit RStudio.
The longer-term solution was to submit an IT service request to release this file from quarantine (that is, to "whitelist a blocked entity"). At my Firm this was promptly done, as there is (obviously) nothing unsafe about this R file.
I have recently experienced a serious problem with Rstudio when developing a package. Whenever, I open an existing project with Rstudio where versions are controlled with Git, it takes so long for it to respond to any command. It is also impossible to type something in the console (e.g. 1+1) and obtain the result. Even quitting the Rstudio, should be done with task manager. There is no problem when I create a new project / package or when I open directly a R script.
This problem appears both when the project is saved on a dropbox or on a local repository.
To overcome this issue everytime I need to modify my code, I create a new project, and then I move toward the new repository all my current R scripts and the folder ".git".
I would appreciate if anybody could help me with this issue.
I had a similar problem to yours. Changing the attribute of my .git folder into hidden solved my problem.
We recently discovered an issue where projects using git for version control could become laggy / unusable on Windows if the .git folder within the project had become a non-hidden directory.
https://github.com/rstudio/rstudio/issues/1918
With the risk of me misunderstanding something, I can't get my packages to sync. I went through the following scenario:
I install new packages on Machine 1 and upload the settings through the "Sync Backup" command in atom. I can see that the new packages are listed in the packages.json file in the gist.
On Machine
I restore the settings and can indeed see settings being restored, like keymaps. However I don't get my new packages. I have restarted and reloaded Atom without luck.
Are there any extra steps I need to take to get the new packages on Machine 2?
You can try atom-package-sync. It is a package that I created a couple weeks ago. It works a little bit like the synchronization of Google Chrome, you just login and it syncs your packages and settings automatically across all your Atom instances. It is based on package-sync but I find it easier to use.
Recently I have been attempting to install a software package automatically over a moderately sized computer network for a university lab. The package is aspenOne v8.4. I have tried creating response files, but every time I finish the install I am unable to locate the response file. I included the f1 flag option to specify file location and name, and I have also checked the %systemroot% directory as well, to no avail. I do not believe one is being created. I have also tried creating a response file for the msi package included on the disc with msiexec, but I was unsuccessful in that path as well. The install requires too much time to install manually across the network, so I was hoping there might be another option to automatically deploy the application package throughout the network. Is there a way to do this?
Thank you in advance.