Answer to comments on uBlock Origin thread: https://github.com/gorhill/uBlock/commit/5733439f629da948cfc3cae74afa519f6cff7b7f as it seems I do not have permission to comment.
Hi,
First of all I'd like to personnaly thank you for all the work you do on uBlock Origin and other extensions, the source code of which have been an inspiration to me personally many times in the past.
I am also really excited that there are multiple people pushing for more accurate measurements of the efficiency of content-blockers and I think sharing methodologies, data and results is a great start!
It is interesting that the results you obtained diverge from the study published yesterday. If I understand correctly you got similar timings for uBlock Origin itself, but the numbers for Adblock Plus do not seem to match (45µs instead of ~19µs). I'd really like to understand where this difference could come from.
The setup we used for the (synthetic) benchmark was the following:
- The version of uBlock Origin we used was commit 29b10d215184aef1a9a12b715b47de9656ecdc3c
- The version of Adblock Plus we used was commit 34c49bbf029e586226220c067c50cec6e8bf8842 of the adblockpluscore repository
- The code used to run the benchmark for Adblock Plus is the following: https://github.com/cliqz-oss/adblocker/blob/master/bench/comparison/adblockplus.js
We initialized an instance of the CombinedMatcher
class using all the network filters (as it seems to be the case in the extension), then used the matchesAny
method of the matcher as an entry-point. Moreover, the parsing of the URLs were performed using tldts
and not included in the measurement. It could be that the parsing and preparation of requests in Adblock Plus is less efficient than in uBlock Origin (which I know is extremely efficient).
The focus of the study was specifically on the network matching engine of the content-blockers and it seems likely that other parts of the extensions are introducing overhead. That's why I really like the in-browser measurement you have setup in uBlock Origin. In the end I guess all of these can be valuable in some way.
I removed the restrictions on the repo, you may comment there if you prefer.
The steps I used to embed the benchmark loop in ABP itself are in this Gist (which I linked to earlier in an answer to you on reddit):
https://gist.github.com/gorhill/67d7b4e75e268fb36658823968427fb5
The main difference I would say is the use of two
new URL()
in order to comply with ABP's entry pointmatchRequest()
-- which computes 3rd-partyness etc. They do use such code in theironBeforeRequest
, so I consider it's representative. Maybe these URL constructors called for every single request is what is causing the difference.On uBO's side, I used
filteringContext
in order to extract hostname, domain, etc. which is also what is used by uBO in itsonBeforeRequest
handler (the goal offilteringContext
was to allow reusing as much as possible already computed values from one request to another).