How to make sure that a process on Gunicorn3 runs continuously? [closed] - nginx

Closed. This question is not about programming or software development. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 3 months ago.
Improve this question
I have a flask app deployed on a Ubuntu server. I use Nginx, and Gunicorn3. I know from this StackOverflow post that one of the correct ways to have the app running continuously on the server is to use something like:
gunicorn3 app:application --workers 3 --bind=127.0.0.1:1001 --daemon
But to be completely safe, since there are many other processes running on that server, I would like to find a way to check whether this process IS indeed running, and if it is NOT running (for whatever reasons) then to start it again.
In addition to that question, to make the app working at reboot, I use the following cronjob:
#reboot bash ~/restart_processes.sh
Where the .sh file executes the command line given above for starting Gunicorn3. Is this good practice or is there a better way to achieve the same result?
Thank you!
Kind regards,

I always use to deploy it in production with supervisorctl + nginx. Check this tutorial . You can simply start, restart or stop with a command.

Related

Tunnel ---------ngrok.io not found [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 1 year ago.
Improve this question
I am trying out Hidden Eye from GitHub as a test with my friends.
Hidden Eye works like this:
1. It makes a fake login page.
2. Uses ngrok to forward this to the people outside your network.
When my friend does it, it works but when I ask him to run for example 2107f2ac.ngrok.io it says Tunnel 79cd0b59.ngrok.io not found.
I have restarted, reinstalled and done everything under the moon.
I thought that ngrok passes firewalls but my only guess is maybe my firewall is blocking it so I added a firewall exception of the ngrok port but it still does not work.
Please help
The generated link is online as long as your terminal/shell is open once you generate link and turn terminal off then it won't work. So if you close your terminal you need to regenerate new link.
I had the same problem and here is the solution: The ngrok system fails many times when the user is not authenticated, I recommend that you log in to the official ngrok web, create an account, follow all the steps indicated in the page and it should work fine for you, I did that and solved the problem
I might be really late, but I came up with the same problem a few days ago, and solved it, the problem I had was that first my tunnel didn't work at all and after that only in local host, but I figured it is because I was using a virtual machine, so Use a dual boot and this should solve the problems, I have encountered a lot of people that says Virtual machines have a lot of bugs with different tools.
Mauricio

AWS EC2 Rstudio Server Error Occured During Transmission [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
Closed 7 years ago.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Improve this question
After over a month, I have managed to piece together how to setup an AWS EC2 server. It has been very hard to upload files as there are very conservative (size) limits when done via the upload button in Rstudio Server. The error message when this is attempted is "Unexpected empty response from server".
I am not unique in this respect e.g. Trouble Uploading Large Files to RStudio using Louis Aslett's AMI on EC2
I have managed to use the following commands through putty and this has allowed me to upload files via either filezilla or winscp.
sudo chown -R ubuntu /home/rstudio
sudo chmod -R 755 /home/rstudio
Once I use these commands and log out, I can no longer access rstudio on the instances in future logins. I can relogin to my instances via my browser, but I get the error message:
Error Occurred During Transmission
Everything is fine other than once I use Putty I lose browser access to my instances.
I think this is because the command is change of ownership or similar. Should I be using a different command?
If I don't use a command I cannot connect between filezilla/winscp and the instance.
If anyone is thinking of posting a comment that this should be closed as it is a hardware issue, I don't have a problem with hardware. I am interested in the correct coded commands.
Thank you :)
Ok so eventually I realised what was going on here. The default home directory size for AWS is less than 8-10GB regardless of the size of your instance. As this as trying to upload to home then there was not enough room. An experienced linux user would not have fallen into this trap, but hopefully any other windows users new to this who come across this problem will see this. If you upload into a different drive on the instance then this can be solved. As the Louis Aslett Rstudio AMI is based in this 8-10GB space then you will have to set your working directory outside this, the home directory. Not intuitively apparent from Rstudio server interface. Whilst this is an advanced forum and this is a rookie error I am hoping no one deletes this question as I spent months on this and I think someone else will too.
Don't change the rights of /home/rstudio unless you know what you are doing, this may cause unexpected issues (and it actually does cause issues in your case). Instead, copy the files with filezilla or winscp to a temporary file (let say /tmp), then ssh to your instance with putty and move the file to the rstudio directory with sudo (e.g sudo mv /tmp/myfile /home/rstudio).

Commands to start and stop Nginx and how many nginx processes will be created? [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 8 years ago.
Improve this question
I am new in using Nginx. I have two questions:
(a) I noticed that are two sets of commands to start and stop Nginx:
$ sudo nginx
$ sudo nginx -s stop
and
$ sudo service nginx start
$ sudo service nginx stop
What are the differences between them?
(b) Once Nginx is started, there are a number of its processes running. So, how many copies will be created and how does the system determine the number of processes to create?
A - In the first way you are using nginx executable to restart it, the second one uses built-in (I typically use it on Centos) operating system util which handles service scripts.
B - It is configured in nginx.conf, by default it is (as far as I remember) set to 2 working processes. Nginx always creates one master process as well
http://nginx.org/en/docs/beginners_guide.html

Debian Wheezy Networking Spontaneously Shuts Down [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 8 years ago.
Improve this question
I have just recently upgraded to Wheezy. Since I have updated my server will spontaneously kill networking. What logs could I look in to see if I can find the issue? I have looked in /var/log and there are no logs that look relevant that have been updated in the past few days. This server runs headless so re-enabling networking means turning the server off and on again as I can't ssh to it.
Any suggestions would be welcome.
Thanks
var/log/syslog should have something. You can run dmesg which may pick it up if it's a kernel module problem or something - to find the module name use lspci -v | grep -i ethernet and look for the module name a few lines later (it could be e1000 or something). Use the module name when grepping the dmesg output.

How do I access data from my personal computer on my AMI instance running RStudio server [closed]

Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
This question does not appear to be about a specific programming problem, a software algorithm, or software tools primarily used by programmers. If you believe the question would be on-topic on another Stack Exchange site, you can leave a comment to explain where the question may be able to be answered.
Closed 9 years ago.
Improve this question
I have recently set up RStudio on an AMI ec2 instance using the process generously laid out by Louis Aslet from his website. But in an embarrassing turn of events I can't access the data I need because it resides on my personal computer. I am new to cloud computing and have 0 functional knowledge of Linux, but I do know SQL, and R well. Any help or suggestions would be greatly appreciated.
Have you tried the "Upload" button in the "Files" window of Rstudio?
use scp in terminal.
To put files from your remote server
Example: if the files are located locally in ~/mylocalfolder and you want to put them in /home/rstudio/mydata you would execute in terminal:
scp ~/mylocalfolder/*.csv ubuntu#<your address>:/home/rstudio/myData/
Note that if you want to access them under a different user, eg, rstudio, you need to change owners on the files. Use chown
To grab data from your remote server
Example: if the files are located on /home/rstudio/mydata and you want to put them locally in ~/mylocalfolder you would use
scp ubuntu#<your address>:/home/rstudio/myData*.Rda ~/mylocalfolder
I use the RStudio AMI all the time and what works for me is to use Dropbox. I can't remember exactly how I did it but I think I may have started the shell from within RStudio and installed Dropbox from the command line.
This link has a little more info:
http://www.louisaslett.com/RStudio_AMI/#comment-1041983219

Resources