I am looking to extract WorldClim climate data for (current data and future projections) and convert the .geotif to .asc in order to run this through MaxEnt and create future climate change projections.
Problem 1: Worldclim gives me 1 .geotif from which I need to extract 19 separate variables, each as their own .geotif file.
Problem 2: Converting these .geotif files into .asc to run using MaxEnt.
I have access to free GIS software (QGIS/DIVA-GIS) and R, although I am fairly new to R. Any solutions would be really helpful, thank you.
Related
I am trying to create a time series object from extracted climate data (NEX-GDDP) using the Google Earth Engine (GEE). The data is daily metrological data, and in the attached file, the data for January, 2005, is collected over an area of interest. The images from GEE are stored in the geotiff as bands (numbered 1-31), and now I am struggling to get these individual bands into a dataset, and add a time dimension to the file. GEE will not export for more than ten years, so my idea is to create yearly files, which, when saved locally, will be merged (concatenated) on the lat/lon and time dimensions.
I am using python in a windows environment, so I am a bit limited (for example, I can't use cdo as this is a Linux based library), and I think that what I would like to do is possible with xarray, but I am missing the (learning) resources to solve this problem with code. Any help and suggestions are more than welcome to help me with this problem.
The image shows the xarray view of the metadata:
xarray metadata view
Showing one band: Test data
Missing the time dimension.
This got resolved on another thread in a special group:
https://gis.stackexchange.com/questions/449759/convert-a-multiband-geotiff-to-a-timeseries-netcdf-file-in-xarray
Is there a possibility to read a subset of a NetCDF file in R?
In Matlab, I can use
ncread(filename,'WS',[i_timeseries,j_timeseries,1,1],[1,1,1,48])
I am trying to extract a time series from the New European Wind Atlas using free software (preferably R) on Windows.
Many thanks indeed.
I'm working in R trying to use the data found here (https://datadryad.org/resource/doi:10.5061/dryad.dk1j0; two top files) to create a table similar to this: [administrative_name, GDP2010, GDP2011....., GDP 2015]
As far as i can see i need to extract the name of the administrative units from the "admin_areas_GDP_HDI.nc" file and combine them with the annual data in the GDP_per_capita_PPP_1990_2015.nc file.
With the ncdf4 package i've managed to open the archives, and to get all the attributes and variables, however I don't know how to access the data and extract it.
I've been trying to access the data all day, but i have limited experience with NetCDF archives, and have not managed to extract the data. Any pointers would help me out!
I like to use the raster package for dealing with NetCDF files. It uses the ncdf4 package to read in the files, but offers some additional tools for processing rasters. You did not mention what data you want to extract, so the example below shows the mean GDP for each administrative unit.
library(raster)
#Read in NetCDF files
ad -> brick('admin_areas_GDP_HDI.nc')
gdp -> brick('GDP_per_capita_PPP_1990_2015_v2.nc')
#Calculate mean GDP using admin zones
zoneMean -> zonal(gdp, ad[[1]], fun='mean', na.rm=T)
I am new to R and have just started to use it. I am currently experimenting with the quantmod, rugarch and rmgarch packages.
In particular, I'm implementing the last package to make a multivariate portfolio analysis for the case of the european markets. In this sense, I need to download the 3-month german treasury bills, in order to use them as risk free rate. However, as far as I known, I canĀ“t download the the mentioned data serie from Yahoo, Google or FDRA databases, so I have already downloaded them from investing.com and I want to load them in R.
The fact here is, my data is different from the ones downloaded by the getsymbols () function of yahoo, because in this case I only have 2 columns, the date column and the closing price column. To sump up, the question arises here is, is there any way to load this type of data in R for rmgarch purposes??
thanks in advance
Not sure if this is the issue, but this is how you might go about getting the data from a csv file.
data <- read.csv(file="file/path/data.csv")
head(data) # Take a look at your data
# Do this if you want the data only replacing ColumnName with the proper name
data_only <- data$ColumnName
It looks like the input data for rugarch needs to be an xts vector. So, you might want to take a look at this. You might also want to take a look at ?read.csv.
I am still a beginner to R and I have the following problem. Any tips will be highly appreciated - it will be a big help to start off.
I have a road shapefile for one country in ArcGIS, that I would like to use as a network variable in R to apply network K-function (look for point randomness on the network using spatstat package). I have never worked with this type of problem before in R and I can't figure out how to convert the vector line dataset into network variable in R. I tried to used the shapefile package, but after reading the .shp, .shx and .dbf files, I don't know how it can be used any further.
The comments above are not correct.
In the spatstat package, first convert your shapefile to an object of class "psp" (planar segment pattern) as explained in the accompanying vignette on shapefiles.
Then use the function as.linnet to convert the "psp" object to a linear network object of class "linnet".
For more explanation, see chapter 17 of the spatstat book.