I am trying to push Piwigo CMS(http://piwigo.org/) to Cloudcontrol I tried the method as same as Drupal, which is described in https://www.cloudcontrol.com/dev-center/Guides/PHP/Drupal%207 But I am getting this error
This is indeed, as pst pointed out, a pubkey issue.
You might want to check your ssh config, it helped to define the following in my ~/.ssh/config:
Host cloudcontrolled.com
user USERNAME
IdentityFile ~/PATH_TO_YOUR/rsa_key
I dunno if this works as well in cygwin or similar: it would help if you could specify your environment a little bit more, as there are many elements (read binaries, configs, etc) in play here.
Related
I want ask the question.
I want to deploy from github to nginx server using jenkins. Maybe you can help me with this? When i try to add in Jenkins setting Publish over SSH i have error
jenkins.plugins.publish_over.BapPublisherException: Failed to add SSH key. Message [invalid privatekey: [B#2a8ec46d]
I just started to study these technologies and most likely I am doing something wrong. Maybe i need configure my nginx files?
Also please tell me something links maybe with Guid how to deploy from GitHub using Jenkins and Nginx.
check this question from StackOverflow, someone faced the same problem and solution provided.
Check first (in addition of my old answer already mentioned) if you need to add to your SSH daemon config file (/etc/ssh/sshd_config), as seen in jenkinsci/publish-over-ssh-plugin issue 94:
PubkeyAuthentication yes
PubkeyAcceptedKeyTypes=+ssh-rsa
I cannot ssh into my Google Compute Engine (GCE) Wordpress instance anymore.
It was working one month ago when I tried last.
I use the Google built-in SSH client in a Chrome browser window.
Yesterday I tried an got the following message:
The VM guest environment is outdated and only supports the deprecated
'sshKeys' metadata item. Please follow the steps here to update.
The "Steps here" link navigates to https://cloud.google.com/compute/docs/images/configuring-imported-images#install_guest_environment which does not seem to help me much.
I am not aware of any changes that I may have made.
How can I fix this?
It looks like your instance's disk is full, and so the SSH keys can't be created in the temp directory. You can do the following:
Stop your instance and wait for it to shut down
Click on the disk your instance is using, and choose "edit" at the top
Enter a larger disk size, and save
Go back to your instance and start it up again
You should now be able to connect via SSH. While you're in there, check to see what filled up your hard disk so you can prevent this from happening again (maybe a rogue program is printing out too many logs, etc).
If you're seeing this on Debian 8 or 9, the most likely reason for this is that the google-compute-engine.* packages that allow SSH access to the instance have been removed by apt-get autoremove.
If you have an open SSH connection to the machine or can use a tool like gcloud, running apt-get update && sudo apt-get install gce-compute-image-packages should fix this.
If you no longer have any SSH access, there is a procedure available on the GCP docs site that can be used to restore it.
I've created a bug report here for this.
Might be a bit late, but you can
1) Stop the VM
2) Edit and enable serial console
3) Use the serial connection to login and update the VM
recent days, I meet similar problem, later I find the permission rights of my home directory fools me, as a lazy-bone, I chmod 777 ~
After did that, I cannot ssh via my terminal, even cannot ssh via browser, only get 'The VM guest environment is outdated and only supports the deprecated 'sshKeys' metadata item, Plese follow the steps here to update'. Sounds like you must set 755 to your home dir, not just care your 700 .ssh or 600 authorized_keys.
I met the similar issue after I created a FreeBSD VM, gcloud ssh not works, but I am lucky that I can use the browser window ssh to my VM. Then I manually add the google_compute-engine public key to the .ssh/authorized_keys, now it work, I can use the gcloud ssh to connect. But not sure if this is a better/security way.
I've researched this feature extensively and need help. PAM is a difficult authentication program for me to thoroughly understand although I'm learning.
We have the option of running either ProFTPD or pure-ftpd on a VPS hosted by hostgator. The OS is CentOS release 5.11. I have complete access to the server. Our goal is to send an email upon successful FTP login by a user. I've successfully used a script which is run by PAM for notification emails dispatched upon successful SSH logins and need to modify it for successful FTP logins. That script which is amazing works fine with ssh (but we need ftp as I've mentioned) and can be viewed here:
http://blog.stalkr.net/2010/11/login-notifications-pamexec-scripting.html
Before I even try to wrap my head around why this script does not work using FTP, despite PAM working on our VPS (for at least ssh), can someone chime in with some theory as to why this is not working at all. I.E., maybe I'm way off base with this idea and it can't work with FTP and PAM in which case I'll just drop this approach. However, when I read configuration settings, etc., on using PAM with these two FTP programs, they seem to indicate this should work.
Additionally, running the following code which is supposed to test for PAM-enabled applications does not return FTP.
ldd /{,usr/}{bin,sbin}/* | grep -B 5 libpam | grep '^/'
Any suggestions or help from gurus would be greatly appreciated.
Found solutions. The answer was to include session_support=YES in the /etc/pam.d/vsftd configuration file. So that's the answer to anyone who may have a similar issue.
In the pursuit of a more professional WordPress dev and deployment environment I am trying to use Capistrano to deploy from my local dev environment to staging and production servers but I can't quite get it to work.
I am on Site5 shared hosting and am deploying to one server with two different domains - staging.example.com and example.com.
I have used https://github.com/markjaquith/WP-Stack as a basis and have added
set :user, "myserveruser" to my config.rb file as the connection didn't work without that.
Running cap deploy:checktells me "You appear to have all necessary dependencies installed" and I can run cap deploy:setup which successfully connects to staging and production creating releases and shared directories etc.
The problem comes when I run cap deploy. Everything seams to work fine until I am prompted for a sudo password. This is not a normal login prompt though and does not allow me to type a password. As I am on a shared host I don't have sudo access anyway.
Its similar to this problem Capistrano is hanging when prompting for SUDO password to an Ubuntu box but the solutions didn't fix all my issues.
I have set default_run_options[:pty] = true which I seam to need but still get the sudo prompt.
I am using passwordless ssh so Capistrano is able to connect and do everything it needs to do without prompting for a password and I am also using an SSH config file to handel agent forwarding so the server can also connect to my git repository. I have confirmed this is working.
I have found other people having similar problems - Capistrano using sudo even with "set :use_sudo, false" - etc but none of the solutions have worked.
I am using default_run_options[:pty] = true and have tried using set :use_sudo, false and default_run_options[:shell] = false but I still have the same issue.
You can view my Terminal output here - http://pastebin.com/5xZmCnyA
I am seriously going crazy! Any help would be greatly appreciated!!!
Cheers
You can't run setup without sudo. That's the only part that REQUIRES sudo, because it makes a directory public_html/staging.exposecreative.org, in your case...
That step shouldn't be required, you can make that directory yourself (assuming you have permission)..
The problem you will face however is that the shared host will expect your index.php to be directly in public_html, for that to work you'd need write permission to whatever directory public_html is in, which your shared host won't allow (almost certainly)
The line default_run_options[:pty] = true has to do with whether Capistrano pretends to be an interactive shell or not. Many commands (sudo included) use this to determine if they should bother asking the user for input, or if they are part of an automated process (
in which case there's no way to ask for input)
My advice (as Capistrano maintainer) - don't try this on a shared host, it almost certainly won't work. (Sorry, their limitations, not ours)
I wrote an extensive post on how to deploy WordPress with Capistrano on a shared host (Bluehost). I use the Roots/Bedrock stack and it only took me about 20 minutes to get up and running.
I made a WordPress development stack using Shipit JS instead of Capistrano because I'm not comfortable with Ruby. Maybe that can help. Feel free to use it and/or post some issues if needed. Regards.
WP-Jungle / Bonzai
I am trying to login in to my works Solaris box from cygwin on windows.
SSH is not supported. I have tried different combination in the .rhosts with no luck
I do not have root privileges.
I found the answer. When I put in the password the way it percieved my login was on top of the screen(whic was not what you would expect) . I put that in the .rhosts and it worked.
While you found the answer to your question with rlogin/rsh, my recommendation is to discontinue use of these insecure programs. Both are known to have buffer overflow exploits, and they both send passwords in clear text.
There's really no excuse for not using SSH these days. Tell whoever owns/manages the system to enable it. That's really in their best interest anyway. It comes installed by default on most Unix varieties, and is available from most vendors as a package.
Anyone who is running a Unix/Linux system and doesn't use SSH is just being irresponsible.