I can see kibana dashbord but it does not work after blnavigating the port on browser - kibana

I am new to open distro. I installed elastic search and kibana on ubuntu 20 and trying to reach kibana localhost.After navigating port 5601 i can see the dashboard but kibana does not work. I used the sample docker-compose.yml file in this link:logz.io/blog/open-distro-for-elasticsearch/amp. After, reaching the localhost with "curl -XGET localhost:9200 -u admin:admin --insecure". I searched in the browser and could enter the kibana dashbord. The kibana status is YELLOW but when I click on the dashbord options on the left (e.g alerting). The process does not proceed and again it shows me the home page of kibana dashbord. Should I do something else to be able to use these options? I would be so thankful if you help me.

Related

Not showing pages of WordPress installed site on localhost

I have installed three WordPress sites on a WAMP localhost:
localhost/pacific_ocean
localhost/surfing_new
localhost/construction
Everything was working yesterday. Today, when I go to the above url's, the screen is blank white, and the WordPress installed site is not showing.
1.) I have checked options.php and it is pointing the site url is fine: http://localhost/pacific_ocean/
2.) I have run 'netstat -abno' on an administrator command prompt and it shows that [httpd.exe] is running on port 80
Then I ran httpd.exe -e warn and here's what I got:
(OS 10048)Only one usage of each socket address (protocol/network address/port) is normally permitted. : AH00072: make_sock: could not bind to address 0.0.0.0:80
AH00451: no listening sockets available, shutting down
AH00015: Unable to open logs
Is httpd.exe getting in the way of Apache running or is this another issue?
One of the sites shows /index of folders, but that's it.
Can someone help me?
Thanks!

kibana service not running

I am trying to install Elastic search, Nginx, Kibana and Sense.
I am following this guide:
https://www.digitalocean.com/community/tutorials/how-to-install-elasticsearch-logstash-and-kibana-elk-stack-on-ubuntu-14-04
I successfully installed Elastic search.
However I am stuck at Kibana.
I successfully followed all steps however:
root#dev:~# service kibana start
kibana started
root#dev:~# service kibana status
kibana is not running
When I run kibana service it says it started, and after that when I want to check if kibana is running, it says it is not running.
If more details are needed for this question to be answered, comment and I will provide it.
Use curl localhost:5601 to test if kibana is really working.
If not working , go to etc/kibana/ to modify the config to check if host is 0.0.0.0 and port is 5601
And the other problem is that your server'memories are not enough for kibana starting.
Hope you can provider the kibana log.
try use:
journalctl -u kibana.service
to show the log.
Even I was facing the same issue.
I have installed the latest version of kibana ( Version: 5.0.2) which is dowloaded from this site:
https://www.elastic.co/downloads/kibana
And I have installed the Elasticsearch of version ( 5.0.2)
I used the below site:
https://www.elastic.co/guide/en/elasticsearch/reference/current/zip-targz.html
It worked for me.
Thanks.
If you have installed older version of kibana or using nginx to wrap to kibana then the following command will not work
service kibana start
To start or stop service, instead of the above, use the following command:
systemctl start kibana.service
It seems like the Kibana process failed to start or execute. (I would reckon it is an issue with the kibana.yml file.) You can see why in the Kibana log output. However, Kibana doesn't have a log file by default. You can set it up using the log_file Kibana server property.
Have a look here: https://www.elastic.co/guide/en/kibana/current/kibana-server-properties.html

How to generate fixed url with ngrok

I am working on facebook messenger.
Facebook app only accept one url for webhook but ngrock is generating new URL every time. Now I am unable to test my app because of webhook URL changed.
UPDATE May 2020
Serveo is up and running again! No installation, no signup!
All you need to do is to run this:
ssh -R <unique subdomain>:80:<your local host>:<your local port> serveo.net
like
ssh -R youruniquesubdomain:80:localhost:8000 serveo.net
UPDATE January 2020
Since there are some issues with Serveo and localtunnel, I want to share with you another free ssh-based self-hosting service: Localhost.run
Unfortunately, it does not provide unique subdomains but it is ssh-based so you do not have to install additional applications. Still waiting for Serveo coming back.
UPDATE April 2018
I've found Serveo just now! And it is totally incredible!
UPDATE November 2017
Probably, it is not the best option for you but I started using localtunnel instead of ngrok.
An installation and run flow is very simple:
npm install -g localtunnel
lt --port <your localhost port> --subdomain youruniquesubdomain
Then I can go to my http://youruniquesubdomain.localtunnel.me
That's it!
No more free subdomain support from ngrok.....pls have a error as below
Tunnel session failed: Only paid plans may bind custom subdomains.
Failed to bind the custom subdomain 'arvindpattartestfb.ngrok.io' for the account 'arvccccc'.
This account is on the 'Free' plan.
Upgrade to a paid plan at: https://dashboard.ngrok.com/billing/plan
ERR_NGROK_313
You need to set up auth token. You can find it here https://dashboard.ngrok.com/auth. (W̶o̶r̶k̶s̶ ̶w̶i̶t̶h̶ ̶f̶r̶e̶e̶ ̶v̶e̶r̶s̶i̶o̶n̶,̶ ̶n̶o̶ ̶n̶e̶e̶d̶ ̶t̶o̶ ̶p̶a̶y̶ it's now paid feature, see ngrok pricing).
Then you can use it like this:
ngrok http 80 --subdomain yoursubdomain
Neither localtunnel.me or Serveo are working for me right now so I created a temporary solution that works for some use-cases including mine (react-native local development): using the ngrok npm package one can save the generated ngrok url into a json file, and that file can be read for any other app.
First make sure to install ngrok using npm install ngrok then use this node script:
const ngrok = require('ngrok');
const fs = require('fs').promises;
(async function() {
const url = await ngrok.connect(3000);
const api = ngrok.getApi();
let data = await api.get('api/tunnels');
data = JSON.parse(data);
let dict = {'domain': data.tunnels[0].public_url}
await fs.writeFile("config.json", JSON.stringify(dict));
console.log("saved " + data.tunnels[0].public_url);
})();
Then from your app you may read it using code similar or equal to:
const backend = require('./config.json').domain;
For DHIS2 local installation, I did this on the terminal on ubuntu server.
Make sure your web is running on the specified port. Mine was on 8080.
ssh -R dani.serveo.net:80:localhost:8080 serveo.net
The beauty of this serveo.net is you can reuse the same hostname prefix url before serveo.net as many time as you want, even if power goes off or internet diconnection.
Staqlab tunnel is providing domain for free. Its works great but need a binary to be downloaded from there website. Using this service for month without any hassle
In 2022 (almost 2023) pagekite.me works for me.
It is very similar to ngrok, and requires the installation of pagekite.py (and, obviously Python).
After installation click on pagekite.py opens pagekite shell.
Run command: 8080 subdomain.pagekite.me
I noticed that no one mentioned how to have static ngrok urls, which was the main question about.
A way to do it is to edit the ngrok.yml file, which is located at
Linux: "~/.config/ngrok/ngrok.yml"
MacOS (Darwin): "~/Library/Application Support/ngrok/ngrok.yml"
Windows: "%HOMEPATH%\AppData\Local\ngrok\ngrok.yml"
You can have content such as:
version: "2"
authtoken: valid_auth_token
tunnels:
first-app:
addr: 3000
proto: http
hostname: yourfixedngrok_id1.ngrok.io
second-app:
addr: 8000
proto: http
hostname: yourfixedngrok_id2.ngrok.io
This will help you expose multiple ports, and have a persisted url for each of which based on the value you set for hostnames.
After that, you run your ngrok using this command:
ngrok start --all
Though its not a solution but take it as workaround, I had the same problem while testing. What i did is keep the ngrok running with my http port, so my ngrok url is not changing. but I frequently changing and restarting my server for testing and debugging.

MAMP not connecting to localhost, any solutions?

I have downloaded the newest MAMP version (3.0.5) and I am unable to connect to the localhost. All I get in Google is "Oops! Google Chrome could not find localhost:8888". I have tried all these things...
Re-downloaded MAMP several times and restarted comp
Changed the ports to 80 & 3306
Turned Firewall on and off and added "MAMP" as an incoming connection
Turned Web Sharing on and off
Checked and unchecked options in MAMP Preferences and hit ok/restart
Is there any solutions out there that have worked for you? I know it all varies, but anything to get this going would be a miracle at this point.
Thank you!
I was also experiencing the same issue on my OS X 10.9.4 with MAMP Pro 3.0.5 installed. The fix was very easy: missing info in my hosts file.
Solution found on: http://forum.mamp.info/viewtopic.php?f=2&t=82253
Note - I am using Mountain Lion
1) check your host file
go to folder /etc/
File path is: Your-hard-drive:private:etc:hosts
this is an invisible folder so you will need something that allows you to view invisible files folders on your Mac. I have Pathfinder installed which allows me to do this.
there should be an entry that makes localhost work
For me there was nothing in the hosts file. It was blank
2) Drag it to the desktop, that'll copy it
3) Paste the following into it and save it:
##
# Host Database
#
# localhost is used to configure the loopback interface
# when the system is booting. Do not change this entry.
##
127.0.0.1 localhost
255.255.255.255 broadcasthost
::1 localhost
fe80::1%lo0 localhost
4) Then delete the one in the etc folder (it'll ask for your password)
5) Then move the copy into the etc folder (password again)
6) Then open terminal and paste this in:
sudo killall -HUP mDNSResponder
(It will require your password)
7) I needed to restart my computer and then MAMP started working again!
MampGuy

WordPress FTP update on EC2 server

I have an AWS EC2 instance running an Ubuntu 12.04 web server that I host Wordpress on. For Wordpress to update, it's asking me to supply FTP credentials. I have set up FTP according to this post: http://stephen-white.blogspot.co.uk/2012/05/how-to-set-up-wordpress-on-amazon-ec2_31.html
But the FTP user I created (ftpuser) can't log in. WP only gives very vague errors, but I tried using FTP in the OSX terminal, which gives 'Login incorrect', but the password is definitely correct. I can FTP in using my normal username and password.
This is the content of my vsftpd.conf file (I've removed all commented out lines):
listen=YES
anonymous_enable=NO
local_enable=YES
write_enable=YES
local_umask=022
dirmessage_enable=YES
use_localtime=YES
xferlog_enable=YES
connect_from_port_20=YES
secure_chroot_dir=/var/run/vsftpd/empty
pasv_enable=YES
pasv_min_port=14000
pasv_max_port=14050
port_enable=YES
pasv_address=54.241.13.224
pasv_addr_resolve=NO
This is an nmap of the servers ports:
PORT STATE SERVICE
20/tcp closed ftp-data
21/tcp open ftp
22/tcp open ssh
80/tcp open http
443/tcp closed https
14000/tcp closed unknown
The /var/www (where I have Wordpress installed) folder is owned by ftpuser and this is the entry for the ftpuser in the file /etc/passwd:
ftpuser:x:1001:1001::/var/www:/sbin/nologin
I'm only an amateur server admin, so haven't a full clue of what I'm doing. Anyone have any ideas why this is happening and what needs to be done?
If you are receiving the following error message "Login incorrect" on AWS EC2:
331 Please specify the password.
Password:
530 Login incorrect.
ftp: Login failed
There is a problem with login using Shell. To overcome this there is one further step missing after the following in that Blog Post:
Add an FTP user, giving access only to the WordPress files and for additional security >ensuring the user can not open a shell:
useradd ftpuser -d /var/www/html -s /sbin/nologin
Add the following:
Add /usr/sbin/nologin on to the last line of /etc/shells file:
$ vi /etc/shells
/usr/sbin/nologin
Try logging in again using your FTP client. That's how I got it working on my instances.

Resources