Instaloader Scrape Results Lost On Computer? - web-scraping

>>> import instaloader
>>>
>>> insta=instaloader.Instaloader()
>>>
>>> insta.login('myusername', 'mypassword')
>>>
>>> acc="targetprofile"
>>>
>>> insta.download_profile(acc,profile_pic_only=False)
I have been downloading full Instagram profiles from the command line no problem. The results are in the App data folder in folders names for the target profile. I used the above code from python, which looks like it worked, but I have no idea where the data went. Searching turns up nothing.

Not sure why the search wouldn't pull them up but it didn't.
Buried way deep in multiple layers in App Data in the Python folder itself.
Previously, results were in app data. This scrape result was placed directly in the python folder.

Related

Connect airflow to google fusion

I'd like to write python script which manages my google data fusion pipelines and instances (creates new, deletes, starts, etc). For that purpose I use airflow installed as library. I've read some tutorials and documentations but I still can't make that script connect with data fusion instance. I've tried to use next string:
export AIRFLOW_CONN_GOOGLE_CLOUD_DEFAULT='google-cloud-platform://?extra__google_cloud_platform__key_path=%2Fkeys%2Fkey.json&extra__google_cloud_platform__scope=https%3A%2F%2Fwww.googleapis.com%2Fauth%2Fcloud-platform&extra__google_cloud_platform__project=airflow&extra__google_cloud_platform__num_retries=5'
with my data json key file and Project id but it still doesn't work. Can you give me an example of creating that connection?
You can find an example python script here:
https://airflow.readthedocs.io/en/latest/_modules/airflow/providers/google/cloud/example_dags/example_datafusion.html
This page provides a breakdown for each Data Fusion Operator if you would like to learn more about them:
https://airflow.readthedocs.io/en/latest/howto/operator/gcp/datafusion.html

Kibana 6.5.4 does not import all exported saved objects and no reason is provided

Here is what happens
I export all the saved objects from one kibana instance with the function: Management/Saved Objects/Export X objects
I try to import the exported objects with `Management/Saved Objects/Import
When selecting the right file, and clicking on Import, I get a similar message to
Import successful
Successfully imported Y objects.
X is a lot more than Y.
I have no logs in browser console, and also no logs that can give any clue in kibana's server log.
I also have run into this problem. Form me it happened because in the Kibana that I exported the object from, there was a plugin installed (enhanced-data-table), and in the Kibana where I tired to import, the plugin was not yet installed.
After installing enhanced-data-table on the new Kibana, it imported all visualisations.
At least a warning message would be nice while importing with missing plugin.

Is it possible to copy the monkey runner results to a excel file

I want the results to be in a excel sheet.
The results should be made as a report.
I am testing a android device and want the result.
monkeyrunner
If you want to automate the whole process you can try xlwt podule to write Excel files from python.
If monkeyrunner gives you some problem importing python modules give AndroidViewClient/culebra a try.
Additionally, ask yourself whether you need Excel at all when you can generate nice reports from python using Sphinx or pod.

Installing Google Cloud Datastore gcd

I'm trying to wrap my head around working with Google Cloud Datastore but am having trouble getting started. I've downloaded the zip of the gcd tool (v1beta2) as described here, which, when unpacked, is comprised of three files: gcd.sh, gcd.cmd, and CloudDatastore.jar. Unfortunately, there are no further instructions on what to do next - where to install it, what path variables or permissions to set, etc. Can someone fill me in?
TIA - Joe
Typical usage looks something like:
# create a dataset
gcd.sh create my-project
# start the local datastore
gcd.sh start my-project
Then, if you're using the Java or Python protocol buffers library, you set a couple of environment variables to instruct the client to use the local datastore:
export DATASTORE_HOST=http://localhost:8080
export DATASTORE_DATASET=my-project
You can find more details about the gcd tool (including instructions for managing indexes) here.

Is there a way to back up my App Services / Usergrid data

App Services is a great place to store data but now that I have a lot of critial info in there I realized there isn't a way to create a backup or roll back to an earlier state (in case I did something stupid like -X DELETE /users)
Any way to back up this data either online or offline?
Apart from API access to fetch records x by x and storing locally, there is no solution at the moment. Team is planning an S3 integration (export data to S3) but no completion date is defined for that yet.
Looks like the only way is to query the data using e.g. CURL and save the results to a local file. I dont believe there is a way to export natively.
http://apigee.com/docs/app-services/content/working-queries
From 2014/2015 Usergrid versions it is possible to make exports and imports using "Usergrid tools"
On this page it is explained how to install them :
https://github.com/apache/incubator-usergrid/tree/master/stack/tools
Basically once you run
$ java -jar usergrid-tools.jar export
and this will export your data as json files in an export directory.
There are several export and import tools avaible, the best way to see them is to visit this page :
https://github.com/apache/incubator-usergrid/tree/6d962b7fe1cd5b47896ca16c0d0b9a297df45a54/stack/tools/src/main/java/org/apache/usergrid/tools

Resources