I am building a web application in which i will be using another websites services.
That website is asking me to provide it with "the URL of the XML feed". I am unable to understand, what is the URL of the XML feed and how to create a XML feed, as I am new to this.
Any code walkthroughs or articles will be helpfull.
Please help !
It sounds like what is being asked for is an Atom Feed. Building an atom feed in ASP.NET is pretty easy to do yourself, but existing libraries like Atom.NET will make it much easier to get going and will also help you follow the standard as closely as possible.
Related
Please assist:
I'm familiar with VBA and C++, but not with Java. Now wanting to delve into Office Scripts.
However, I want to know if I can achieve the same as in VBA:
I am logging into niche websites and fetching data in tables using VBA Internet Controls (getElementByID()), etc.
As far as I know, these niche websites do not have an API, as the sample scenario of webscraping on the Microsoft website does:
https://learn.microsoft.com/en-us/office/dev/scripts/resources/scenarios/noaa-data-fetch
I would like to know if I can log onto these websites, and then fetch information using HTML (getElementByID()) or similar?
I am just unsure if I can use Office Scripts directly, or if I require to include some library or something.
Any guidance would be appreciated.
Currently, there is no way to do this through Office Scripts alone. The fetch command and REST APIs are the only ways to get data in a script directly from webservices. If you'd like to request the addition of a specific library, please use the Send feedback button in the Office Scripts Code Editor.
The discussion in the comments about using Power Automate is a reasonable path to pursue. The linked video (https://www.youtube.com/watch?v=_O9eEotCT0U) is a good place to start.
Does anyone know a web crawler tool for collecting contact details from a website? Say I have a www.website/contact.. I want to pull out the address, phone number, etc.. There are 2 tools I've been looking at: cralwer4j opensource jar for java and Scrapy opensource in Python. But I am finding it a bit hard to use for my scenario.
Any suggestions would be great. Thanks
You might google for "simple web crawler" to find a solution that fits you best. In the net there are plenty "pure python" based web crawlers. Based on sceleton code you add db wrap up. I think the most problem would be db setting and saving data in it.
What if there are 1000000s of websites to crawl.. Is there a way to crawl all websites in my are?
No problem for scripting. Just put millions addresses in a file (or files), open it for reading in python or other script. Then get link by link from it and crawl/scrape to your pleasure. Result you might also want to save in file (csv, json).
I'd also recommend you a ready simple python crawler.
I am developing a system that has a database for news headlines from various sources. I have not worked with RSS before so I am confused about a lot of things. Can anyone please point a good tutorial for how to develop such a thing? Thanks
In my mind, I have questions like:
1) How will I get the latest news feed? do I have to check the rss feed link every few minutes and see If it's different than the previous one?
2) Is it a good practice to parse the feed xml myself or use a feed reader kind of thing?
3) Will I have any control over the feed sent to me. e.g I only need news feed for Google or Intel.
RSS is a very standard format you can start learning at w3c school.
About your questions.
If you can talk with the RSS provider, maybe they can notify each time
something new comes. They can use, for example,
XML-RPC notification.
You can also ask the RSS provider how often should you check the feed
(in case they cannot provide any kind of notification).
I think it's better to develop your own bot. There is lot of
frameworks that can deal with rss format. In case you are working with C# you can try with SyndicationFeed Class
I'm not sure if I'm undestanding your problem, but if the provider
puts a RSS link at your disposal, you must actively navigate that
feed. When you have that feed, you can work with the metadata in
order to see what's interesting for you. For example checking "category" or "channel" node.
I have been playing with this for a day and cannot seem to get it to work.
elFinder has many options for connections, such as PHP and Python among others. Someone made a third party ASP.NET connector for it and it was reported that it worked.
Here is the page for the DLL for it:
http://elfinderconnectornet.codeplex.com/releases/view/55319
I am having tremendous trouble figuring out how to hook the existing elFinder to ASP.NET.
I hope to find someone who has had some experience with this and can assist me. I'm not sure what other information I can give you.
Any help is appreciated.
There's an even better option right now called ElFinder.Net Connector hosted at CodePlex. The coordinator and main developer of the port Evgeny Noskov is a super considerate person and he listens to feedback and acts as you can read here in a discussion I posted:
Add startPath and uploadMaxSize to connector options
I liked elFinder.Net so much that I wrote a post to showcase a simple use case that is a great fit for it:
Manage folders & files in your ASP.NET MVC app with elFinder.Net
Hope it helps.
See here: http://elfinderaspnet.codeplex.com/
For now it provides some basic functionality but seems to be usable.
ElFinder is very easy to configure and use.
I did some research on 'Implementing ElFinder on multiple pages using MVC' and found "ElFinder.Net Connector" NOT AVAILABLE.
Download the simple MVC sample project from here
I'm pretty new with JSON and i'm a little confused...
What I'm trying to do is to use the userless access for foursquare api, to get the venues of a location. Can you please show me a similar example in .net?
you may have more luck using one of the community contributed .net libraries. See https://developer.foursquare.com/resources/libraries for details.