I would like to post an article to all of my groups where I am an administrator. Is there a plug in or some code to do this? Until now I have made it so that:
first --->had established a category e.g. groups
The second---> copy RSS feed address from my blog and paste it in RSS 2.0 feed (grafiti) application included.
Naturally this also runs, but then it gets up there via RSS graffiti and not my application, as I usually do on my fan page or profile.
Thanks for all answers
Sorry I am from Germany and place answer easy!
Related
I administer a mediawiki for a school group. We have a website that students complete projects in for virtual rewards. I want to put counters on a page of my wiki with statistical information (cumulative exp/coins, assignments completed, most productive student, etc.) about each of the seven groups that the students are divided up into. It would be simple enough if the two sites were hosted on the same server, but they are not. I figure that an RSS feed with the statistical information may be a good way to get info from website server to wiki server. How would I reference the information from the RSS feed in the wiki page?
Just to make sure my idea is clear, I would put in the feed something along the lines of:
[ATLAS]
exp=15000
coins=7500
eva=350
ip=150
dmg=500
[CERES]
exp=13000
;and so on
I would like to reference that in the wiki page. Is it doable?
It is that we can add a url to digg.com reader to add them to one's reading list. Once we submit the link they(digg) find several links which points to the feed map(.xml or from feedburner etc). How do they do this ?
Feed discovery is step 1.
Step 2 is probably to have a "reverse" discovery mechanism. Each feed (RSS or Atom) as a link to an HTML representation. So, even if the other discovery part is missing you can still find a feed URL from a site if you have that 2nd relation... Combine that with user data (number of subscribers... etc) and you get a pretty good solution!
Step 3 is "dumb" search: each story in a feed has a link to domain. The feed itself has a domain. If they both match, you likely have found the feed you're looking for!
There are probably other techniques too :)
In Gitlab there is an RSS news feed button for every project.
http://git.domain.name/userName/projectName/commits/master.atom?private_token=xxxxxxxxxxxxx
Is there a way to retrieve a global RSS feed for all projects related to one user?
Or even better, if there is a way to capture all user's feeds across all projects, that would be cool.
thanks
This is currently not supported.
The closest feature request there is would be "Team-/groupwide RSS feeds", to have RSS feeds for all events in a team or a group. (not enough vote for now).
The global "RSS" shown on demo.gitlab.com is actually a html result of the event-items representing the activity in response to http://demo.gitlab.com/?limit=20&offset=0:
That would be for all public projects, but I don't see it exposed as an RSS feed.
The OP mauro reports in the comments
the global "RSS" shown in demo.gitlab.com is also showing on my gitlab dashboard, the only thing is that mine is giving a 500 error instead.
This may be a simple question, but for some reason I don't know this answer. Is it possible to create an RSS feed file that contains contents for an entire year but only publishes the current date and previous date information?
I have a client that wants to do a "this day in history" post. Currently, I am using IFTTT, and created around sixty dated posts for the next two months. Of course, this works -- but it is very labor intensive.
Is it possible to create an RSS feed that you could put all 365 days of data in to, but if someone pulls up the feed it only shows today's item and prior days in the feed?
Or is RSS not the proper technology to do this? The reason I am using RSS is for ease of use, and IFTTT will take those RSS feeds and pump it in to Facebook and Twitter for automatic status updates for my client.
There are various tools that let you define Facebook and Twitter posts in advance, to be published at a specified date and time in the future. Why not use one of those instead of writing your own?
A quick search for "scheduled twitter post" uncovered Later Bro, Twuffer and twAitter but there must be dozens to choose from.
If you're looking for just posting on Facebook and Twitter, and not an RSS feed as well, I'd follow Matthew's suggestion. If you want an RSS feed, there is a feed for each Twitter feed. But if you want actual RSS, you need to add something in between. An RSS feed is just an XML file. it's not a process. I suggest having a file of some type (maybe RSS, or other XML, or a database table, or even a csv file with all the posts and relevant information, including date. Then a small script that runs as a chron job (or IFTTT if it supports date as trigger and running a script as the "then" part) that pulls the day's feed and updates the actual RSS feed. Pretty simple.
Here is what I ended up doing
Using the Drupal backend of my website, I created a content type specifically for these posts.
I created individual articles for each day, and used the schedule module to schedule the publish date to the date I wanted.
I created an RSS feed of these posts through Drupal.
I linked the newly created RSS feed to IFTTT.
Created an IFTTT recipe to post the text from the RSS feed to Facebook/Twitter/etc.
It wasn't the best solution, but it worked. I was really trying to do this without having to rely on a third-party such as IFTTT, but never really figured out a good way to do it.
When creating an RSS reader, you download the XML formatted document pointed to by the RSS feed link, and you can parse it manually or using the functionality in the SyndicationFeed namespace.
So if we take Scott Guthrie's blog as an example, you download the RSS feed document here, and parse it. My problem is that this document only holds 15 items, yet he has been blogging for a number of years.
Is there a standard or established way of getting the older posts not included in the RSS feed document? Or do you have to find the base address for the blog posts and then parse the pages of the site from there to get them? How do you avoid missing posts on high volume blogs?
With RSS/Atom you can't query older articles.
I built a RSS archival service (https://app.pub.center). All of our data is free to use via REST. We charge money for push notifications.
PubCenter daily polls it's catalog of RSS feeds, and caches the articles. Then, you can get these articles back in a chronological order. For example:
Page 1 of The Atlantic https://pub.center/feed/02702624d8a4c825dde21af94e9169773454e0c3/articles?limit=10&page=1
Page 2 of The Atlantic https://pub.center/feed/02702624d8a4c825dde21af94e9169773454e0c3/articles?limit=10&page=2
As the replies to How Do I Fetch All Old Items on an RSS Feed? already mentioned, a feed may not provide archival data but historical items may be available from another source.
Archive.org’s Wayback Machine has an API to access historical content, including RSS feeds (if their bots have downloaded it). I’ve created the web tool Backfeed that uses this API to regenerate a feed containing concatenated historical items. If you'd like to discuss the implementation in detail please get in touch.