The 1st of a 3 part series that will describe some of the changes, which included replacing Varnish with Apache Traffic Server (ATS) as the on-disk HTTP cache component of the CDN.
We have been working this past year to better identify and tag the “bot spam” traffic so we can produce top pageview lists that (mostly) do not require manual curation.
Learn about using the Mediawiki History Dataset to explore the every day experience of editors on Wikipedia.
By Srishti Sethi, Senior Developer Advocate Overview Small Wiki Toolkits is an initiative that focuses on building technical capacity in…
This blog post documents how we prioritized debugging an issue on mobile; how we went about implementing a solution; and what we learned from the experience.
Part 3 of 3 posts on Wikimedia’s event data platform.
Wikimedia Release Engineering is considering a move from Gerrit to Gitlab. Learn more about the reasoning and join the discussion.
In the previous post, we talked about why Wikimedia chose JSONSchema instead of Avro for our Event Data Platform. This post will discuss the conventions we adopted and the tooling we built to support an Event Data Platform using JSON and JSONSchema.
Programs like Google Summer of Code, Google Season of Docs, and Outreachy can provide a platform to accelerate your FOSS journey. Learn about Wikimedia’s outreach programs.
The Wikimeda Foundation has been working with event data since 2012. Over time, our event collection systems have transitioned from being used only to collect analytics data to being used to build important user facing features. This 3 part series will focus on how Wikimedia has adapted these ideas for our own unique technical environment.