New to MapBox.
I have read the tuto on how to create a layer from OSM which relies on a PostGIS connection. Tried it. So far, so good.
Nonetheless, given the very limited magnitude of the project i'm working on (single user...), i'd really rather avoid having to have a pgsql instance running just for that. sqlite comes across as an option of course!
Can someone help with the following questions:
Any reason why sqlite would not do the job ? The data set i'm pulling from OSM is about 30MB
Any recommandable and tried script to convert OSM .xml or .pbf to .sqlite ? There is a bunch of osm2sqlite out there on github and where-else, but can't see any reference to them that let me think they will still be maintained in a foreseeable future...
How straightforward will it be to link such sqlite output to a tile mill layer ? Clearly, I have no idea on the underlying data model and subsequent sql statement to bring it across...
Many thanks
Laurent
Use mbtiles (this is basically sqlite if you didn't know).
Mbtiles will work perfectly for you. In fact Mapbox uses it for their maps
As you've seen from that link, tilemill can export your project to mbtiles . Personally I would import to postgres using imposm3, it's fairly fast and doesn't use up ask your memory. After importing I'd style the map with tilemill then use it to generate mbtiles.
Seems like you're already comfortable with the first 2 steps of this. The docs cover the exporting step
Never used mbtiles but I can't see why you would have difficulty with them.
Nb In the future you'll be better off asking questions these at gis.stackexchange.com/
Related
everyone.
I've been stuck for some days searching some way to get the skeleton of a point cloud data (like OBJ) but not using kinect. Is it possible?
I found the Point Cloud Library which does a lot of tasks related to point cloud data, and in their documentation there is a body keypoints detector, but it also works with kinect grabbers.
In my case, I have a point cloud data like in the picture, which was generated by another depth sensor scanner. Is it possible to find the key points in such data?
I really would appreciate any help. Thanks in advance.
Even if it's not explicitly mentioned in the tutorial you linked, a quick to the code suggests that you can use different data sources (e.g. PCD files), so you're not stuck with the live capture from Kinect.
All the tutorial code really does is the following:
Setup the GPU for the people parts detection.
Pick the appropriate data source.
Load the tree files for the body part detector.
Run the PeopleDetector on a single frame captured from the live grabber stream/PCD file.
I need to create a dependency graph for a software suite that I am working on. In the past the company I work for has always done this manually, but I am guessing that there is a tool somewhere that will do what we need.
The software I am working with is Ada95, and has about 200 code modules/files, with about 40 packages. I need to create a map that will trace every output, individually, back to each input or constant that will have an impact on the output. Does anybody know of a tool that would accomplish this? Or even just partially accomplish it?
AdaCore's GPS (available from http://libre.adacore.com) comes with a command line tool named gnatinspect. You can use this tool to load all cross-reference information generated by the compiler (assuming you are compiling with GNAT). This creates a sqlite database (gnatinspect.db) which contains all information you need. gnatinspect itself provides a number of pre-made queries that might get you at least partially to where you want to go.
You could also look at ASIS, as a way to do this kind of queries directly on the code. I am told this is not so easy to use the first time around though.
There is also an older tool provided with gnat (gnatxref) which does something similar, although it is being superceded by gnatinspect.
Finally, you could look at gnat2xml as an alternative to ASIS if you are more comfortable parsing XML files.
I am trying to use broadleafcommerce and customize it.On study i found it uses Apache Solr . However, i am already handy with
ElasticSearch as i am currently using ElasticSearch only in my workplace. so, i'm curious as if i can replace that customizable code of broadleafcommerce with ElasticSearch. If it is possible, i also want to know how long will it take or what will be its difficulty level ?
Thanks in advance !
The product is open source, you can have a look at the code yourself. Here is the package that would need to be made solr independent. As far as I see there are quite some dependencies on Solr now, but maybe you can give it a shot and contribute it back. In the end that's the power of open source.
I can't tell exactly how much work that would be since I don't know the product and what it does with the data. The solr schema would need to be translated to the related elasticsearch mapping, then the indexer will need to be converted in order to push data to elasticsearch (otherwise if technically doable you could write a river that imports data in elasticsearch from the framework itself). Last step is to convert the search code together with the facets, highlighting etc.
Maybe you (or the people behind the project) might want to have a look at spring data which has now a community driven spring-data-solr project and an unofficial elasticsearch implementation too.
I want to build a small application (on Windows), one of function is read the SQLite database (SELECT query). I just need to select some columns from table, no need to update/insert/delete ...
Of course, SQLite is an open source software, then I can get source code easily. Actually I did it, but it's to big because it support full APIs. All I need are just some APIs: sqlite3_open, sqlite3_close, sqlite3_get_table, sqlite3_free_table.
I tried to porting source code to smaller lib, but the data structure and API too complex ...
Please let me know if you know the other library (smaller and simpler) or some one did it before can help with share some idea.
Thanks.
I am using fusion tables to generate pins over the map. now I want to insert more coordinates data into the table using php but couldn't get any exact idea how to do it?
Can anyone help me for that?
Thanks in advance
I would like to respectfully suggest that answers that simply refer posters to the Fusion Table PHP Library may not be helpful to many and can be outright discouraging to some. The PHP Library is a tremendously impressive, interconnected set of scripts. It is also massive, mostly uncommented, and requiring many nested 'includes' and 'requires.' The whole thing must be over 2000 lines long and it is hard to even decipher which parts you need.
For the poster the starting point needs to be the Developer's Guide and the OAuth2 Guide. These are at:
https://developers.google.com/fusiontables/docs/articles/oauthfusiontables
https://developers.google.com/fusiontables/docs/developers_guide
See the Fusion Table PHP Library