Migrated JFrog from On-premise to AWS Cloud but my artifacts are showing only 40 where as we used to have 200k+ artifacts - jfrog-cli

I am trying to migrate our on-premise JFrog artifactory to AWS cloud. I have followed the "Method-1" instructions.
Even I have mounted a file system and moved the filestore content to that location and set it as Home directory.
But, still I am only able to see 40 artifacts showing. Not sure what is causing the issue.

Related

Artifactory to Artifactory remote repos

i have 2 Artifactory server and have configured ubuntu repos.
One off the artifactorys goes to internet to ubuntu and the other
artifactory connects to the artifactory with internet access.
I have the following problem. From my local artifactory i get always the error 404.
I cant fetch the metafile (Packages) from the ubuntu repo.
But if i reconfigure my remote repo and set --> store artifacts locally all seems fine.
I want store the artifacts locally. My local artifactory should ask the artifactory with internet access and should get all files from the remote artifactory.
Have anyone a idea how to solve my problem?
best regards
I assume you are trying to set up a smart remote repository with Artifactory. Refer to this wiki and set up the smart remote repository, basically, you should be adding the URL in your local Artifactory's remote repository as http://ARTIFACTORY_URL/ubuntu-remote/ and make sure the "store artifacts locally"is checked so that this remote repository can able to index the artifacts.

I am trying to connect a local artifactory running on a laptop in docker to a remote artifactory on AWS behind an ELB

I have set up Artifactory OSS Version 6.9.1 on an AWS instance behind an ELB and have been successfully deploying builds to it from GitLab CI/CD. I am now trying to set up a local Artifactory OSS Version 6.10.0 on my laptop so that I can develop builds locally before sharing with the team.
My local artifactory connects perfectly to JCentre and I can browse that repository.
My gradle build will happily connect to the AWS hosted artifactory at http://{URL}/artifactory and resolve my dependencies.
When I connect a remote repository with http://"{URL}/artifactory I get a 500 Internal Error message on Test. If I take off the /artifactory it says it has connected successfully but when I try to browse the remote repo it is empty.
I read Connect one Artifactory to another Artifactory and followed the instructions to edit the json configuration and make the remote repository a smart repository here https://www.jfrog.com/confluence/display/RTF/Smart+Remote+Repositories.
It now has the smart repository image but still cannot be browsed.
Gradle still cannot resolve dependencies with the local artifactory using the remote-repo name.
As stated in the Smart Remote Repository documentation, you should configure the remote repository URL with the following structure:
http://ARTIFACTORY_URL/api/package-type/repository-key
So if you have a Gradle repository named "gradle-test", the URL should be:
http://ARTIFACTORY_URL/api/gradle/gradle-test
Hope this helps.
In the end it turned out to be ridiculously simple. When setting up the remote repository the key needs to be in the url as well as the key field.
So for a repo with a repository key 'fractal' the connection URL is counter-intuitively http:///artifactory/fractal

Does artifactory home directory need to be world-readable?

We operate Artifactory on a RHEL 7 Linux server. It runs as the local user artifactory. His home directory is /var/opt/jfrog/artifactory and its mode is 755. Our security folks would prefer it to be 750. Does this directory need to be world-readable or can we restrict it?
I installed Artifactory and changed the home directory and all its subdirectories to be "chmod 750". After some testing, I did not encounter any issues with this change. I successfully started Artifactory and performed some basic tests such as creating a few repositories, changing some configuration, uploading and downloading artifacts, creating users and permissions, etc. All of my tests were successful and it seems that you can change the mode of the Artifactory home to 750 without any issues.

How to clone JFrog Artifactory into a local installation of JFrog Artifactory?

I have JFrog Artifactory http://organization.com/artifactory/ with read access. I can download individual artifacts let's say gulp-git-2.8.0.tgz using native browser.
Now, I have created local JFrog Artifactory http://localhost.com:8081/artifactory/ in machine and it does not have any artifacts.
I am trying to fetch my organizations artifacts lets say of npm package http://organization.com/artifactory/npm/ into my local artifactory http://localhost:8081/artifactory/npm/
Since it has many artifacts and I cannot do it manually of download from organization server and deploying it to my local server.
Is there anyway that I can fetch only the required package let's say npm to my local server? Or is there anyway to replicate my organization's package using read access into my local server?
Thanks in advance.
If you have Admin permissions to the source Artifactory you can setup Push replication or to perform a Repository Export.
Without Admin permissions, you can:
Setup Pull Replication.
Use the JFrog CLI in order to download entire repository content, and then upload it in the same structure.

Bonobo Git server on Azure Websites and Azure Storage

I'm using Bonobo Git Server. I want to host this application on Azure Website, but there is a disk limitation to ~10gb (Basic plan). This is not enough to host git repositories. I'm thinking if there is any way to use Azure Storage connected to my website to host those repositories?
Currently bonobo application uses local git .exe to perform appropriate processes with repositories. I have no idea how I can make this working with Azure Storage or if it is possible at all.
You won't be able to get git.exe to work with azure storage without mounting it as a normal file system like in Azure Files which won't work in azure Web Apps anyway. You can upgrade to premium and get 500 GB though, so depending on your scenario, you may wanna look into that.

Resources