Best method to work in R program [closed] - r

Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 7 years ago.
Improve this question
I have planned to start my R program carrier now. I have understand the usage R but i have some doubts on the procedure to be implement in R. As a startup i am planing to retrieve data from google analytics and show it to frontend using R shiny. So which will be the good flow to fetch the data. Is it good to fetch the google analytic data using R and process it or it will be good to fetch the data using other languages like php or Java. Also R is good to fetch the data from database or its a good practice to fetch the data using other commonly used languages and process using R.
Sorry the question is little bit descriptive but i am expecting a help from you guys
Thanks,

Organize it in a package(http://r-pkgs.had.co.nz/).
See https://developers.google.com/analytics/solutions/r-google-analytics?hl=en .
The database question depends on exactly what you are using, but R has a lot of resources.

Related

Is Firestore a good choice for time series data? [closed]

Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 3 years ago.
Improve this question
I have a small project involving some simple financial time-series data with some real-time components on the front end. I was hoping to use the Firebase infrastructure since it offers a lot of things without having to set up much infrastructure, but upon investigating it doesn't seem to be a good choice for storing time series data.
Admittedly, I have more experience with relational databases so it's possible I am asking an extremely basic question. If I were to use Firestore to store time-series data, could someone provide an example of how one might structure it for efficient querying?
Am I better served using something like Postgres?
Probably best bet would be to use a time-series database. Looks like Warp 10 has already been mentioned (https://www.warp10.io).
The benefit of something like warp is the ability to query on the time component of your database. I believe firebase only has simple greater/lesser than queries available for time.

Communication between Different programs in Different Computers (C# or Python) [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
I'm building a program that I want to be able to exchange information with other programs running in another computer. I started using C# and a library called SimpleTCP. The main issue is that is too simple and only send and receive messages.
I'm looking for something that I can predetermine functions that one or the other can call from each other.
I looked on google and stack overflow but I was unable to find an appropriated subject to study, what should I be looking for to learn this? Thank you
The most complete protocol for what you want is gRPC. There is a learning curve but worth it in my opinion. https://github.com/grpc/grpc
There is a way but it's little bit different
Such programs like this written in tow different languages
You can make a center database between the both programs
In this situation it's very easy to communicate and receive ,send data
You can mysql ,oracl, mariadb or any Database you prefer

How to automate the process of generating daily report from firebase? [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
For example, to automate the process of producing a daily report for selected events and the duration of time that unique users spend on some specific event. And it is even better if I can customize the reporting information and have it generated in excel sheet automatically. Any suggestions or ideas would be much appreciated.
You could use a Cloud Function to generate you excel report, for example by using excel4node (https://www.npmjs.com/package/excel4node)
And to call this Cloud Function regularly, you have to trigger it via http through a cron-job.
Have a look at:
https://firebase.google.com/docs/functions/http-events).
and
https://www.youtube.com/watch?v=CbE2PzvAMxA
Note: What works quite well too is to generate some PDFs via the Cloud Function, using pdfmake.

Data confidentiality in R [closed]

Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 6 years ago.
Improve this question
I'm considering using R in my work. Primarily I am interested in ggplot2, dplyr and tidyr, probably using the RStudio environment.
I deal with a lot of sensitive educational data, such as names, dates of birth etc and could only only use R at work if I knew the data were safe.
Will my data be kept private/confidential or is it shared in any way?
By default, data or code is not shared to the outside world. This may not hold for cases including but not confined to:
when you are using (online) version control
if you are pushing a report to RPubs repository
using R on a server, which may be accessible to third parties

Proper Coding Techniques in R [closed]

Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 7 years ago.
Improve this question
I am writing an R code that has many different functions that eventually I will want to use all together on different data sets.
As I keep building functions it seems to be getting harder to keep track of everything in my script.
My question is, is it proper R coding to break functions into separate R Scripts or should it all be in one massive script?
Thank you for your help. This is my first time trying to code something this large!
-B
Yes, you can store your functions in multiple R scripts.
If you need to call them, you can use source().
For eg:
Say you have func1 , func2 saved in myfunc.R.
To call them,
source('myfunc.R')
#other codes
func1()
func2()
As to whether this approach is recommended, depends on your project requirements.
Alternatively, you can consider packaging them as recommended by Richard.

Resources