R version: 4.0.3
RStudio version: 1.3
Windows 10, 1909
At my work I have to use OneDrive to store all my files and R scripts. This has worked fine up until recently. Now I get an error message when trying to save or modify an R script. I can create a new script and save it once, but if I try to make changes and save it again the error prevents me from doing so.
The error message in RStudio is:
"The create operation failed because the name contained at least one mount point which resolves to a volume to which the specified device object is not attached"
Not the most helpful error message.
The same behavior happens in R itself, but the error message just says it cannot save.
I have seen some talk online about this being related to that latest Windows 10 version (1909), but I started to experience the issue about 2 weeks before my computer updated to 1909. Besides, my work will not allow a roll back to the previous version.
Microsoft's OneDrive team failed to provide any help on this because they see it as an error on the R/RStudio side.
Anyone else experiencing this? Any solution?
Related
I used to not having issues on this particular subject, but bigrquery::bq_auth is not working on my installed RStudio. Funnything is that within an R terminal session it works perfectly.
All started when I updated the R to R version 4.2.2 (2022-10-31).
This impacts me because I'm not being able to run a shiny application due to this problem.
Does any one knows what am I missing or has any clue where to look for? At least for start, because I am lost.
The code I am running is quite simple:
json="The_file.json"
bigrquery::bq_auth(path = jason)
The message:
*Error: Can't get Google credentials.
Are you running bigrquery in a non-interactive session? Consider:
Call bq_auth() directly with all necessary specifics.
*
As a matter of fact, I was just hoping to have the same answer in the R terminal session:
The bigrquery package is requesting access to your Google account.
Select a pre-authorised account or enter '0' to obtain a new token.
Press Esc/Ctrl + C to cancel.
Folks, I removed all packages and installed everything possible again. It worked. Sorry. I had tryed a lot and this last one ocurred me later. Regards all.
I'm trying to use the profvis package to do memory profiling of a large job in R (64 bit), run under RStudio, run under windows 7. profvis keeps crashing and I get an error message saying that Pandoc is out of memory. The message is copied below.
My understanding, and please correct me if this is wrong, is that the problem is likely to go away if I can set the /LARGEADDRESSAWARE switch on Pandoc. And to do that, I need to install a linker, etc., do my own build, after learning how to do all those things. Or, there is a shortcut, involving installing MS Visual Studio, running the editbin utility, and set the switch that way. However a new install of Visual Studio is unhappy on my machine, and demands that I fix some unspecified problem with Windows Management Instrumentation before it will go forward.
So my question is this: Is there a way to set the /LARGEADDRESSAWARE switch on Pandoc from inside R?
I had a similar problem and was able to resolve it by following the advice at https://www.techpowerup.com/forums/threads/large-address-aware.112556/. See in the post where it has an Attached File called laa_2_0_4.zip. I downloaded it and ran the executable it contains. Basic mode was sufficient; I simply navigated to C:/Program Files/RStudio/bin/pandoc/pandoc and turned on the checkbox for Large Address Aware Flag (step 2), then did Commit Changes (step 3). After this, the profvis-invoked pandoc command eventually ran to success. I was able to watch pandoc's memory consumption in Task Manager rise up to a peak of about 2.7 GB.
I am unable to load an R project and I am getting an error: Error occurred while checking for updates. Unable to establish a connection with R session with R studio. I am using large data sets inside that projects which might be causing this issue. Is there a way I could log in from the terminal and remove some datasets from the workspace. Can someone help how I can log in?
It sounds like you're auto-saving the workspace to .RData. It's usually not a problem (though still not advised) to do so with small-ish data sets/objects in the workspace but it's almost deadly to use this setting if you work with large data sets/objects often.
Disable the Restore .RData into workspace at startup setting in RStudio preferences and also set Save workspace to .RData on exit to Never.
Hunt down all the .RData files in your various working directories and delete them, too.
I had the same problem. And the above proposed strategy did not work well. However, having anaconda installed in my PC I downloaded Rstudio on that platform. Contrarily to the other Rstudio standalone version, Rstudio in anaconda works well. I do not know why, but in case of need you can smash out your Rstudio without deleting your files and use the Rstudio on the anaconda platform. Just to have another point of view!
I am getting strange and annoying behavior in one (yes, only one) of my RStudio projects. Whenever I restart the R session within this project, I can no longer save any of the files, and when I try to (or RStudio tries to automatically) I get a popup error saying "The system cannot find the file specified" and I have to reopen the project and all of the R scripts.
I looked at the log file, and I see a lot of the following error:
07 Oct 2015 00:52:05 [rsession-Mark] ERROR system error 5 (Access is denied) [path=C:/Users/Mark/GoogleDrive/Research/CEQUAL_event/.Rproj.user/5E5F98D9/sdb/s-26E44539/lock_file];
OCCURRED AT: bool rstudio::core::FilePath::exists() const C:\Users\Administrator\rstudio\src\cpp\core\FilePath.cpp:308;
LOGGED FROM: bool rstudio::core::FilePath::exists() const C:\Users\Administrator\rstudio\src\cpp\core\FilePath.cpp:308
It looks like the problem may be in the .Rproj.user folder. So I deleted that and let RStudio regenerate it, but the behavior started right back up again.
I am using Windows 7, RStudio version 0.99.441, and R version 3.2.2.
I think #Jack Wasey is right: Google drive locks files during sync. This causes many issues with other programs as well. There are alternatives from companies that do file syncing for a living, such as dropbox, whose clients don't lock files. With google drive you only have the option to use selective sync and exclude every .Rproj.user folder. Once the RStudio's database is corrupted, moving the project out of the Google drive folder won't be enough - you will need to delete/rename it to have RStudio fix its database.
Summary:
Copy contents from all unsaved files to a temp file outside of RStudio (while clicking away the error message every other second)
Quit Rstudio
Pause Google drive syncing
Rename (or move/delete) the hidden .Rproj.user folder in your project
Reopen the project in Rstudio
Exclude the new .Rproj.user folder from syncing with Google drive (Preferences -> sync only these folders)
Remember that Google drive is free / Become an awesome google developer and fix the issue once and for all
May not be related but might be helpful to someone else. I was getting the same error (over and over) and finally realized that I had an R session open from days ago, that had eventually gotten stuck. It was running in the background with a "Fatal Error" message waiting for me to click OK. When I closed that session everything started working again.
This was happening to me as well, but curiously only when I was using Project files (.Rproj - not stand alone R scripts). It was driving me crazy.
I found that buried in the settings (Tools -> Global Options -> Git/SVN) there was a setting "Enable Version Control Interface for RStudio Projects). I guess I had turned that option on when I was messing with my settings and then forgot about it since I don't normally use project files.
What worked for me was I noticed DescTools package I installed caused it, though it served the purpose I installed it - fetched me mode value without going through mode functions. I then deleted the line I got the install.package, re-saved the project (ctrl S) and I could knit it.
Bottom line, read the error message soon as the box pops up, it's going to give you a hint of what line is causing the issue.
I hope this works. Enjoy
I have just updated my R to version 3.1.1 on my Mac machine, and since then have started getting this strange message when running analyses.
Persistent UI failed to open file ..../org.R-project.R.savedState/window_2.data: No such file or directory (2)
Where I have written the .... there is the rest of the file directory, but I didn't think it necessary to add in.
There does not appear to be anything specific that sets off this error. I am usually running a number of R windows at once - which are running difference analyses (around 6) and the error appears across all windows at roughly the same time I think, I am not usually watching the analyses tick over.
It doesn't kill the analyses. And I haven't found it causing any problems.
I am just curious what it means, and how I should fix it?
Thanks,
Sam
This is pretty old, but I just ran in to the same problem.
You see this message because R does not have permission to change your ~/Library/Shared Application State directory.
If you want to change it, 1 just navigate to the folder, 2 right click it, and 3 uncheck the lock checkbox or change the permissions to that folder.
I don't mind the R error message because I hate OSX writing all the current application states to that folder.