I am trying to find a list of what ISBNs are in use. I guess I could scrape a website like Amazon but that would waste a lot of bandwidth. Is there a better (free) way?
Maybe you could use the remote API for isbndb.com.
Trying to keep an enormous ISBN list up-to-date yourself is quite a huge task if you ask me.
Just for the record: note that if you actually want an ISBN for your publication, you need to go to the official agency in your country. In the US this is http://www.isbn.org/ , but it varies by country. In Australia, for example, it is here.
This might help: What is the most complete (free) ISBN API?
As the accepted answer states there is also an API to search Amazon but it's not actually supposed to be used in the way you wish to.
ended up using partial list from http://my.linkbaton.com/isbn/
Yes, try isbndb.com
Related
Im looking for a way to add an NPC which sells items based on the achievements a character has. Because of my lack of other coding skills, if at all possible i would like to achieve it with SQL commands, hence modifying the db. I was looking through the conditions page on the wiki but have no idea how to use the provided information.
Also i was backtracing the db regarding the NPC Charles Worth who happens to teach tailors recipes based on achievements they have. I intended to copy this toons conditions, but couldnt find what entries to use.
Any help, clarifying db entries, or pointing to the right direction in another way, is much appreciated.
Please follow this link for the documentation:
https://www.azerothcore.org/wiki/conditions
You can use the source type: "SOURCE_TYPE_NPC_VENDOR" and the condition type: "CONDITION_ACHIEVEMENT" for what you need, how to implement this, you can find that in the link above.
Also, one way to make this easier is to use the tool developer by the azerothcore team, Keira3.
This is a very visual Database Editor and can help you understand what each column do as almost each cell is documented and you have links to the full documentation as well.
Keira3 link: https://github.com/azerothcore/keira3
I'm going through crawling wikipedia using website downloader for windows, i was looking through the whole options in this tool to find an option to download wikipedia pages for specific period, for example from 2005 untill now.
Does anyone get any idea about crawling the website in specific period of time ?
Why not download the SQL database containing all of Wikipedia?
You can then query it using SQL.
Give a try to the Wikipedia API and your programming skills.
There should be no need to do web scraping; use the MediaWiki API to directly request the information you want. I'm not sure what you mean by "wikipedia pages for a specific period" - do you mean last edited at a certain time? If so, while skimming, I noticed an API call that lets you get a look at the last n revisions; just ask for the last revision and see what its date is.
It depends if the website in question offers the archive and mostly don't so its not possible in a straightforward way to crawl a sample started from specific date. But you can implement some intelligence in your crawler to read the page created date or something like that.
But you can also look at Wikipedia API at http://en.wikipedia.org/w/api.php
I am in need of a tool to show the size of a word according to how many occurrences it has (similar to twitter trending topics),
is there a software to do this?
if not, is Processing a good option for this?
Thanks a lot
edit,
tnx for the response!, no I know this is called tag cloud and can use the proposed software on the answer and others!, cheers!
I recommend wordle, as it organizes de words in different ways
Here you go: http://tagcrowd.com/
I have been looking for an autosuggest search script and I have finally found one that I like, the only problem is that I cannot find a way to get it to run off our database results.
Is there any way to customize this script so that it runs from our own database, and not off the freebase pre-defined data types?
http://www.freebase.com/docs/suggest
Have you tried overriding service_url and service_path ? There are also the corresponding params for the flyout service. It's documented in the docs that you pointed to.
As masouras says, you can override service_url and service_path, but that's not particularly helpful unless you have another service which provides the same APIs as Freebase.
Dae Park recently posted a recipe to the Freebase mailing list which might help - however, I'm not aware of anyone who's actually managed to get Suggest working with anything other than Freebase.
It seems like such a simple thing, but I can't find any obvious solutions...
I want to be able to take two or three feeds, and then merge then in to a single rss feed, to be published internally on our network.
Is there a simple tool out there that will do this? Free or commercial..
update: Should have mentioned, looking for a windows application that will run as a scheduled service on a server.
There are a whole pile of options here: http://allrss.com/rssremixers.html.
Maybe http://www.planetplanet.org/
will do what you want.
It's for creating blog aggregations like planet lisp.
Google reader, create a group, add your feeds into the folder and then share that as an RSS feed.
:-)
Works while you're asleep!
Yahoo Pipes could be nice. Depends on how much "private" you want the resulting feed to be.
For 100% offline solution investigate Atomisator. It's a Python framework basically for doing offline what Yahoo Pipes does online.
If you're using PHP, the SimplePie library will do this. Here's a tutorial.