User Details
- User Since
- Oct 8 2014, 5:48 PM (506 w, 5 d)
- Availability
- Available
- IRC Nick
- Milimetric
- LDAP User
- Milimetric
- MediaWiki User
- Milimetric (WMF) [ Global Accounts ]
Today
This is now done.
Great question, @mforns. This was mostly for performance reasons. I couldn't find a way to get Spark to optimally work on the full day of pageviews without first aggregating it like this to > 250. But the execution plan I ended up with looks pretty wild. Let's talk tomorrow when you have some time. I'm attaching the change here.
Yesterday
I've migrated and shut off the old instances. I will delete them in a couple of days, just in case. But everything's working fine without them. Did not know about the wmflabs -> wmcloud automatic redirect, that made everything very simple.
I grouped a couple of tasks under this so we're less likely to lose them in the fray.
Fri, Jun 21
The simpler way to do this, just two phases as opposed to progressive, gets us fairly similar results, with about 200 fewer rows which are all detailing specific browser versions.
We get a ton more detailed results this way, and the total coverage increases to 99.7%. Still not 99.9%, but I think we may have too much detail at some point. I'm fairly happy with these results, and I'm going to prepare the new browser general query as a gerrit change. It'll be good to get some review.
This might affect some data we sqoop into HDFS and some of how we compute commons impact metrics or similar future metrics. We have to wait until a schema change is proposed to know for sure.
From a discussion with @Krinkle about the data, a preliminary idea of how to roll up is:
Thu, Jun 20
The long and the short of it is that we can get that "other" to about 2% if we simply roll up remaining data by browser family and os family. We could get fancier but let's see what folks think about just this approach.
Ok, so these rows represent more than 0.1% of all views for this day and they're aggregated all the way down, so this is what we were dumping before, followed by a big "Other" bucket:
ok, I have some results for us to peruse, from rolling up in different ways. First of all, my query so we can debate whether or not it's accurate.
Mon, Jun 17
Analysis available in this spreadsheet: https://docs.google.com/spreadsheets/d/1iSlH5XsRXV7mDoku0F5HbLNJmx1CMBm6ECakZMPUbU8/edit?usp=sharing
Made up some slides to help think about this data:
Wed, Jun 12
Tue, Jun 11
This looks like a great system to get started with. I can think of some potential snags that come up, so as we build it let's keep an eye out for these and similar:
Fri, Jun 7
select day, http_status, count(*) count_by_status from pageview_actor where year=2024 and month=4 and day in (19,26) and geocoded_data['country_code'] = 'HK' and normalized_host.project_class = 'wikifunctions' group by day, http_status
day | http_status | count_by_status |
19 | 200 | 389 |
19 | 301 | 4311 |
19 | 302 | 931 |
26 | 200 | 198 |
26 | 301 | 133801 |
26 | 302 | 1028 |
Thu, Jun 6
Wed, Jun 5
Tue, Jun 4
https://mpic.svc.eqiad.wmnet:30443/ is the endpoint we should use to talk to the mpic service making sure we don't take unnecessary hops through DNS.
Fri, May 31
hypothesis so far: maybe some workers are getting MaxMind updates on a staggered schedule from others, so there's always some variation?
(and sub-country it's much worse)
Wed, May 29
The two instances have been moved, docs updated on wiki and in code, and proxies have been moved. The only problem is the new proxies can't use the old wmflabs.org domain. For now, I left the old proxies up and additionally set up the new proxies. So, for example, both https://pingback.wmflabs.org/ and https://pingback.wmcloud.org/ work. Whenever the old instances are deleted, the old URLs will stop working. I guess part of sign-off will be to communicate this and maybe delete the old instances?
Keeping track of how I do this for future reference. (The previous task where I did this was T236586 and I failed to take good notes there)
May 23 2024
https://gitlab.wikimedia.org/repos/data-engineering/mpic/-/merge_requests/31 is including both the Menu Button (sorry I couldn't find a task for this) and the Multi-lookup, as well as using them from one field each. I'm happy to help with this more when I'm back next week.
May 16 2024
quick update: resolved with Eric to work on this as a separate component. Will start on a patch now, keeping it in the Codex sandbox for now with T363432 as the goal.
May 15 2024
May 9 2024
May 1 2024
ok, I didn't do much here, just provided a very short description and detailed out the schemas as Marcel had them in the design doc. Please let me know if anyone was imagining something else.
filling out the readme right now, thanks Ben!
I am not sure this is 100% squashed because the behavior is so weird. Here's what I found, in short:
Apr 18 2024
It would be cool to do a quick spike into Scalar and the customization we'd need there. Abstain as a voter here, I like all the options just fine and I have bad aesthetics when it comes to reading docs because I just start hacking and see what happens :)
+1 for Option 2. For what it's worth, when we initially put up the endpoint docs on wikitech we were just doing so while we waited for a better end user experience than the swagger UI afforded us. I especially like the integration with wikitech described in option 2 (the discovery pages that would lead wiki users to the docs)
Apr 16 2024
+1, SSR is kind of a pain if done in fancier ways, but in this way you get a lot for free and it helps even reduce code. As a bonus the user gets a great experience.
Apr 15 2024
I've broken this down into subtasks but I'm keeping it as something between an epic and an actual task. It's coordinating and has all the acceptance criteria, it was just too big. So I'll leave the other two subtasks on the boards while I'm on vacation and put this in paused. This can be resumed whenever you'd like to continue work on coordinating and deployment.
Apr 11 2024
Approved, welcome back Andy :)
Approved
Apr 4 2024
Apr 3 2024
Apr 1 2024
Mar 29 2024
I found a candidate bug. The script used to ask for the year and month, and after the change it asks for the day. generate_druid_unique_devices_per_domain_daily_aggregated_monthly.hql seems to have been adapted to give the correct result, but evidence to the contrary, the druid output seems to be only one day. Running now to prove or disprove.
Merge request 582 seems to have changed how we do this monthly druid segment aggregation, so the answer must be around here. Again I checked the new source table, now Iceberg (wmf_readership.unique_devices_per_project_family_daily) and again that seems to have data for all of January, for example.
Looked at this a bit today.
Mar 25 2024
@VirginiaPoundstone: Looks like Giuseppe patched varnish to send more requestctls, so maybe that completely or partially solves the problem. I'd have to look through the data to see. I'm going to do a good job focusing and only do that if you put it in the sprint :) (should take no more than an hour, but it's probably not like a few seconds if I want to be more thorough)
Mar 22 2024
I made the puppet change but I need an SRE to merge. This is not well documented indeed, we should talk about a better way to maintain this interface that so many people use.