Excel messes up some dots (".") in a number [closed] - r

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Closed 6 years ago.
This question was caused by a typo or a problem that can no longer be reproduced. While similar questions may be on-topic here, this one was resolved in a way less likely to help future readers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Improve this question
I have a tab delimited file:
I created this file in R as a data.frame and wrote it to the above file using write.table(dataFrame,"filepath",row.names=FALSE). However after I opened this in excel I got some ##### in my excel file:
The only difference between the tab del file and the excel file is that in the excel file the . is omitted, but I don't have any idea how this is possible because most of the other numbers are just fine. Any suggestion to fix this problem is welcome.
Update
I can fit the data in the column:
However there should be a . after the 1

Probably your import settings are wrong regarding the seperation for thousands and decimals. Notice that the problem arises when the first number is >1. Excel interprets a number as a thousand if the first number is > 1 , because it woudln't make sense for excel to convert a number which begins with a 0 to a thousand. So you have to fix this:
You have to do this while importing the file in the last step, you have to click on Advanced and then set the Decimal seperator to: . and the Thousands seperator to: , (or visa versa, it's what you prefer offcourse but in your case it has to be this)

Related

Remove re-occuring text strings [closed]

Closed. This question needs debugging details. It is not currently accepting answers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Closed 4 years ago.
Improve this question
I am new to R and have searched the forum for almost 2 hours now without getting it to work for me.
My problem: I have a long text string scraped from internet. As I scraped code for images were included. The are coded in a way that they start with "Embed from Getty Images" and ends with "false })});\n". I would like to remove everything in between those strings. I have tried gsub() as per:
AmericanTexts3 <- gsub("Embed.*})});\n", "", AmericanTexts)
But what happens then is that they remove everything between the first picture and the last picture. Do anyone know how to solve this?
You need to use a non-greedy regular expression.
Try
AmericanTexts3<-gsub("Embed.*?})});\n","",AmericanTexts)
The ? matches the first occurence of the second part of the regex, so that only the part between the matches should be removed.

R cannot export all rows to csv [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Closed 4 years ago.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
This question was caused by a typo or a problem that can no longer be reproduced. While similar questions may be on-topic here, this one was resolved in a way less likely to help future readers.
Improve this question
Fundamental stuff but i couldn't seem to get around this. I performed the following process:
d1<-read.csv('hourly.csv',sep=",",header=F)
names(d1)<-c("date","rain","q","qa","qb")
d2<-read.csv('event.csv',sep=",",header=F)
names(d2)<-c("enum","st","et","rain2","qtot")
for(k in 1:206){
st<-d2[k,2]
et<-d2[k,3]
Datetime<-d1[st,]
print(Datetime)
write.csv(Datetime, file="DatesA3.csv")
}
In the end, i exported the results to a csv file. There are 206 rows altogether and they display fine in R. But when exporting, only the last row is exported in the csv file. I tried multiple things such at write.table, append, etc. but nothing seems to work.
How do i export every row into one file?
Please advise and thank you!
Datetime[k, ] <- d1[st, ] # instead, otherwise you overwrite
# and write the result outside the loop

Find duplicate registers in R [closed]

Closed. This question needs debugging details. It is not currently accepting answers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Closed 5 years ago.
Improve this question
I have an excel file with a list of emails and channels that collected it. How can I know how many emails per channel are duplicated using R and automate it (every time I import a different file just have to run it and get the results ) ?
Thank you!!
Assuming the "df" dataframe has the relevant variables under the names "channel" and "email", then:
To get the number of unique channel-email pairs:
dim(unique(df[c("channel", "email")]))[1]
To get the sum of all channel-email observations:
sum(table(df$channel, df$email))
To get the number of duplicates, simply subtract the former from the later:
sum(table(df$channel, df$email)) - dim(unique(df[c("channel", "email")]))[1]

Reading all observations from a csv file [closed]

Closed. This question needs debugging details. It is not currently accepting answers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Closed 5 years ago.
Improve this question
i have imported this file into R only problem is there is 380 observations and it only reads first 100 observations. How can i get the rest of it, here it is
BPL16_17 <- read.csv("BPL16:17.csv")
BPL16_17
Thanks
Personally I always recommend using readr::read_csv over read.csv.
While I am unsure why read.csv is limited to 100 columns (This has not been true for many years now, my mistake) read_csv is not and handles data_frames much better especially dates, times and doesn't include factors by default.
https://github.com/tidyverse/readr
Also a great resource is this chapter from the R for data science book which is available online always for free.
http://r4ds.had.co.nz/data-import.html

working directory with spaces in R [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Closed 5 years ago.
This question was caused by a typo or a problem that can no longer be reproduced. While similar questions may be on-topic here, this one was resolved in a way less likely to help future readers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Improve this question
I am having the problem that when I use setwd() with a path that includes spaces (e.g. setwd("C:/Users/Name/My Documents/") I get the error message
"cannot change working directory"
I am a bit suprised that I did not find much about this here or on google - so it must either be a rare error or everyone knows about it I reckon. Either way, is there a why to work around it?
I am using Windows 7 and R version 3.0.2.
R cannot setwd into a directory that it doesn't have 'x' (execute) permission for.
This should work, but if really needed, you can use the function shortPathName.
> shortPathName("C:/Program Files (x86)/Adobe/")
[1] "C:\\PROGRA~2\\Adobe\\"
I can replicate your error if I just copy and paste
setwd("C:/Users/Name/My Documents/")
to console as is. The problem is that R cannot find the specified path. I believe that you should replace "Name" with your username...
If I replace "Name" with my username, it works as expected.

Resources