My team is currently thinking about uploading files to our repository (BitBucket) to minimize storage cost on our on-prem DC. We'd like to build a web application (ASP.NET) that will upload files to the application's repository on a separate folder. Basically using the repo as a cloud storage.
I have little knowledge on repositories and I think repos should only be used for code/versioning. I'd like to explain why is this feasible or not to my team.
Is it possible to use BitBucket as a cloud storage why or why not?
Related
i Already hosted a Web page Project to Filebase console using firebase deploy comment , How can i Download the Project files Now ?
firebase deploy
You can not download the project files...
That you can download is the data/files you have stored in your realtime database or firestore database.
You can switch between deploys as well, you can roll back to older deploy or to the newer depending of your needs, but you can not download any of it.
There is no blanket "code pull" from Firebase, so you will need to take a different approach for each product.
Product
How to get the code/data
Hosting
There is no API to get your hosted files back from Firebase, but you can usually get pretty close by crawling the site. See Pull lost code from Firebase Hosting deployment
Realtime Database
You can either use one of the SDKs, use the REST API (for small databases this may be as simple as adding .json to the end of the database URL), or you can set up automatic backups and get the data from there.
Firestore
You can either use one of the SDKs or REST API here too, or you get export through the console or gcloud CLI as explained here
Authentication
You can export all your users through the Admin SDK or the Firebase CLI
Cloud Functions
The closest you can get here is to copy/paste the code from the Google Cloud console as explained here: Get code from firebase console which I deployed earlier
Is there a way to upload a folder with some files into different accounts on firebase hosting simultaneously?
I need this to be able to update the same app on different firebase accounts without the need to do it one by one.
There is no built-in command to deploy to multiple projects on multiple accounts. You will have to call firebase deploy separately for each project.
You may be able to automate the process though, by using the CLI as it'd be done in CI systems.
Is it safe to keep the firebase config file (with the APIkey)when deploying my mobile app (front-end react native) to the google play store and app store?
How can I make it safer?
As Doug commented, the google-services.json (or google-services.plist file for iOS) does not contain any secret credentials. It merely contains the configuration data that your app needs to find its Firebase project on the servers. So sharing it with other developers on your app is not only safe, it's required for them to build an app that communicates with the same Firebase project.
You may want to consider keeping it out of version control though, and instead only deploy it onto your build server. The reason for this is not as much that that data is secret, but more that each developer should typically set up their own Firebase project for their development work. That way they won't be stepping onto each other's toes during feature development work.
Also see:
Is it safe to expose Firebase apiKey to the public?
Should I add the google-services.json (from Firebase) to my repository?
I need to develop an web based app in .net and upload image, Files and videos to AWS using asp.net and also make these files downloadable using link to aws source.
Looking at Amazon AWS sort of confused me. with so many products.
I would appreciate how can we develop .net based app which will upload different kinds of large files to amazon aws .
We want to host this app on different Server and upload file to aws server.
Not sure where to start from of what products to look for on amazon.
Confused with cloud front, amzon buckets etc..
After re-reading your question and noticing that you want to host your app externally to AWS, all you will need at this point is:
S3 (Simple Storage Storage)
This will host your static files - images, videos, files. Files stored in S3 can be made publically accessible via URL.
Your .NET application will make use of the AWS S3 SDK for interacting with your S3 bucket. This abstracts the technical details and makes it simple to CRUD files in S3.
Here's an example of the code you'll need.
At a high level, what you'll need to do:
Create Amazon AWS account
Create S3 bucket
Include AWS S3 SDK to your application
Deploy and test
Authentication
If you want to make some or all of the files in your S3 bucket public, you will create a bucket policy.
As your application is not going to be hosted in AWS, you will be authenticating using access keys, ensure these are protected and also ensure these are not generated from your root AWS account.
I have an ASP.NET application running on shared hosting and want to use Windows Azure blob storage to store images (my shared hosting has limited storage). My application will upload and download these images from/to the client PC.
I created a new cloud project to test this out and got it working successfully with the offline storage emulator, but I noticed that the cloud project has a web worker role which, I believe, needs a compute instance when deployed to Azure. Azure storage is very cheap, but if I need a compute instance for my application to talk to to access it then it becomes much more expensive, even with the smallest instance.
My question is - do I need a compute instance in order to upload and download files to blob storage or can my application talk directly to the blob storage?
Short answer: You do not need compute instances to talk to Azure storage. You can talk to Azure storage directly from client apps or from apps hosted by other providers.
Nor do you need to create a new Cloud project within Visual Studio to take advantage of Azure blob storage. Simply, reference appropriate .DLLs (Azure Storage Client) and you're good to go.