I found this site which allow me to scrap the announcement title and date. However, the amount of announcement is limited to 20. See image below, anyone know where I can scrap the title and date for each announcement for the fiscal year 2021
https://www.asx.com.au/asx/1/company/CBA/announcements?count=50&market_sensitive=false
The API is limited to 20, it is an undocumented API, other than spending time working with the API to figure out other ways around it, they don't provide any other method of getting results outside the bounds of the 20 results.
If you want more, you're going to have to find another service, which likely involves paying or writing more code.
Related
Our main challenge in Google Analytics at the time is to measure the success of our magazine articles.
The problem is that views grow over time so in any timeframe we always have the older articles overshadowing the newer ones. Sidenote: The same problem occurs for measuring social media post success.
My idea of a solution is to measure the rate by which views on articles grow. An article that has a higher growth of views is much more successful than an older article with more views, but with a lower growth rate.
Alternatively something like "views within the first week(s) of publishing this individual article" would also be a good metric.
Unfortunately to some extent also the growth rates rely on this publishing period of individual articles if we are interested in an eternal high score of articles. But since we are mainly interested in recent articles, growth rate would still give us the desired result of showing the most successful recent articles.
Has anyone dealt with the same challenges and found any solution to this, in best case with Google Analytics?
These examples may help, of which I have direct experience.
In the data layer we included a date of publication for the article and then used this to determine growth. This was taken from the CRM and was relatively straightforwards for the dev team. This was stored as a custom dimension in Google Analytics.
We had nothing in the data layer but instead a I just used the date on which page views started appearing as a proxy for date of publication. Not entirely reliable, and you may want to filter by views >5, or whatever is appropriate, to avoid any hits from editors or staff before a page is visible in the site navigation.
In both cases I was exporting data either to Google Sheets (using for example the Google Analytics API addon for sheets) or BigQuery, where it was relatively straight forwards to identify the first date and then calculate, for example, views per day. In your case it would be having a function which looks at the date of publication + 7 days. You may also be able to achieve this with Google Data Studio or similar dashboarding platform.
I am currently using the Google Places API on a free trial. I am interested in paying for the API but can't find the exact cost of the two commands that I use: google_places(), and google_place_details(). I have contacted the Google sales team and looked at the places and billing url, but I have not managed to find the answer of how much it would cost exactly to execute these two commands.
For google_places(), this is an example of a command I would execute:
google_places(search_string = "Cafeteria in Madrid, Spain", key=key)
From the places and billing url, it seems like this counts as a text search, so each time the code is executed it would cost 0,032$. Is this the case?
For google_place_details(), here is an example of the command I would execute:
google_place_details(place_id = "ChIJf_XA-F0U04kR1IPYSdTJ4so", key=key)
This command, as well giving basic place details (which cost 0,017$ according to the billing url),
gives information which counts as contact data (an extra 0,003$) and atmosphere data (an extra 0,005$). It also provides photo data (0,007$ according to the billing url), which I am not interested in but is automatically included in the results anyway. Does this mean that the cost of executing this command once is these four prices summed up?
I am interested in knowing exactly how much it would cost to execute the two commands I have listed.
probably this helps:
First of all you are billed monthly after you exceeded the 200 Euro/Dollars, which are given by google for free (as you probably described as "free plan"). So after every month you get a bill on how many requests of each function you send to google. There everything is written quite clearly including the amount and price of each "unit". then you can easily divide it.
Second option would be your Google Api Cockpit.
It tracks your requests quite precisely on different time bases. So sending your wanted commands only once on a day can give you an exact total-price.
The Cockpit is super handy for different things. If you want you can even set limits, which is probably helpful in your case too.
Here is the link to the billing monitor as well: Billing Google API Cockpit
Furthermore the description of how google charges you. Look here
best regards
I work for a non-profit that needs to see how our fundraising efforts are going in 'real-time'.
We look at results in blocks of about a half hour - so we need to report on how we finished the last 24 hours or so and also where we're at in the current half-hour. We're accomplishing this through google analytics, as we have multiple fundraising streams all pointing to a common GA account.
I have tried using datastudio to report against the GA API, but that connector does not seem to refresh at a reliable rate - someitmes it'll pull fresh data within a minute, sometimes it can take twenty minutes to report on recent transactions. I believe the 'real-time' API could be used to get fresher GA data, but as far as I can tell, that will only report 'live' data, and not prior/historical data (say from four hours ago). Does anyone know what API I could use if any to pull all data historical through current datetime?
I apologize if this request is vague, but I'm just looking for a conceptual approach at this point to get the freshest data - preferably in one fell swoop (API call). There is more complexity post-data intake (I have to then compare it to goals we've set for each half-hour, amongst other nuances to the transacitons themselves), so i wanted to start with this fundamental piece/question.
Thanks!
Given the context provided, I believe that the API solution would not be feasible. Among other reasons:
The real time API only offers a limited amount of dimensions and metrics. For example, e-commerce data is not available.
https://ga-dev-tools.appspot.com/dimensions-metrics-explorer/
https://developers.google.com/analytics/devguides/reporting/realtime/dimsmets
The Standard intraday processing SLA for the Core Reporting API is < 24 hours for standard properties. The processing occurs on a best effort basis. Meaning that an hourly availability can occur from time to time but can not be guaranteed.
https://support.google.com/analytics/answer/7084038?hl=en
As an alternative approach to the API solution, you could consider the use of an App + Web property which would allow you to stream event data in real time to BigQuery. However, this solution has some cost implications and would introduce you to a new tracking paradigm.
https://developers.google.com/analytics/devguides/collection/app-web/tag-guide
https://support.google.com/firebase/answer/6318765?hl=en
https://www.simoahava.com/analytics/getting-started-with-google-analytics-app-web/
For example, getting the information that there were 56000 tweets in Los Angeles, CA on May 30th, 2015 at 5:00 PM. Ideally, I would like to access tweets as far back in time as possible. The main issue with this seems to be getting tweets from the past - Twitter's REST API only has tweets for about the past 7 days.
The only way that I am thinking of is manually counting the tweets for a particular location and time from Twitter's website directly, but surely there must be a better way.
Is there some simple way to accomplish this that I'm not thinking of, or any libraries or frameworks that optimize this task?
As of now, I'm trying to use R, but I'm not sure if other languages or frameworks would be better suited for this.
You will never achieve your goal, even if you pay for Gnip. The reason is less than one to two percent of twitter accounts reveal their location.
If that is not a problem for you, and you don't want to pay for Gnip, you can use statuses/filter, which is a streaming endpoint, along with the locations parameter. Then you would simply count the tweets as they come. However, you would not have access to historical information.
Gnip would also give you the historical tweets you desire, and Gnip has some intelligent algorithms that will increase the number of geo-locatable tweets to around 20%, more or less.
The company I work for has just purchased 4 32" LCD screens to be mounted at the front of the office for demonstration purposes. Whilst we are not demonstrating (most of the time), the screens are to be used as development information screens for the whole team.
What information would people recommend displaying to be most useful to the team? Our focus is on hosted business web-apps but I am interested in what other teams doing other types of development find useful too. Pointers on how to gather the displayed information would be useful also.
Information about your continuous integration status.
Major Development Milestones that have been hit in the last week
Releases within the last month (including a short description why this release is awesome)
Use it as motivational board. The achievements of software development are seldom communicated well enough.
Since you're hosting apps for your customers, server and network status information would probably be useful.
Heck, why not create a "chat room" for the dev team to discuss issues and post a streaming version of that as well?
Schedule information, Scrum notes from that morning, a gantt chart...the possibilities abound.
Outstanding bugcount, sorted by priority and severity. You can likely get this from your bugtracking tool programmatically.
Depending on your process management
system, possibly a list of feature
requests and the percentage complete
on each of them. Again, you can probably get this programmatically from your process management / time tracking tool.
Time spent in the current development
cycle, and time remaining. Again, this should be available from your process / management / time tracking tool. You may want to use this data with your bugcounts as well to give a bugs / day fix rate.
If you're a public company with a
profit-sharing plan (i.e. stock or
options), the current price of the
stock (this can be surprisingly
strongly motivating). You can get stock data from several sources online programmatically (although a small delay may be injected unless you're paying for the service).
The movie 'Office Space'
Weather radar from intellicast.com
Latest Checkin.
Number of checkins per day
Number of customers that use software
Metrics on Bugs found/fixed and the ratio.
One screen could be an aggregated RSS feed of development topics pulled from sites such as Stack Overflow (or even Coding Horror). Not sure what your goal for these screens is, but I could see it useful to me if you had a feed with topics specific to your development team headlined. If I were there, I'd glimpse them, maybe catch an interesting thread, and go learn something. Funnel a bunch of keywords and tags through a Yahoo Pipe and dump it to the screen.
That's if they are more "informal and informational."
I think most popular pages from your webapp(s) would be a fun/interesting thing to show on a big monitor up front.
Another would be a live feed of your error reporting.
We have one monitor showing all meetings for the day, with start-end, subject, and room. I find this helpful, not only for my orientation, but also to see what other people do at our company.
xkcd, bunny, dilbert and savage chickens :-)