Electron efficiency for measurements with the ATLAS detector and using the 2012 LHC proton–proton collision data – The European Physical Journal

This paper describes the algorithms for the reconstruction and identification of electrons in the central region of the ATLAS detector at the Large Hadron Collider (LHC). These algorithms were used for all ATLAS results with electrons in the final state that are based on the 2012 pp collision data produced by the LHC at  = 8 TeV. The efficiency of these algorithms, together with the charge misidentification rate, is measured in data and evaluated in simulated samples using electrons from  Z → ee, Z → eeγ and J/ψ → eedecays.

For these efficiency measurements, the full recorded data set, corresponding to an integrated luminosity of 20.3 fb-1, is used. Based on a new reconstruction algorithm used in 2012, the electron reconstruction efficiency is 97% for electrons with ET = 15 GeV and 99% at ET = 50 GeV. Combining this with the efficiency of additional selection criteria to reject electrons from background processes or misidentified hadrons, the efficiency to reconstruct and identify electrons at the ATLAS experiment varies from 65 to 95%, depending on the transverse momentum of the electron and the background rejection.

Electron isolation

In order to further reject hadronic jets misidentified as electrons, most analyses require electrons to pass some isolation requirement in addition to the identification requirements described above. The two main isolation variables are:

Calorimeter-based isolation: The calorimetric isolation variable EconeΔRT is defined as the sum of the transverse energy deposited in the calorimeter cells in a cone of size ΔR around the electron, excluding the contribution within Δη × Δϕ= 0.125 × 0.175 around the electron cluster barycentre. It is corrected for energy leakage from the electron shower into the isolation cone and for the effect of pile-up using a correction parameterized as a function of the number of reconstructed primary vertices.

Track-based isolation: The track isolation variable pconeΔRT is the scalar sum of the transverse momentum of the tracks with pT > 0.4 GeV in a cone of ΔR around the electron, excluding the track of the electron itself. The tracks considered in the sum must originate from the primary vertex associated with the electron track and be of good quality; i.e. they must have at least nine silicon hits, one of which must be in the innermost pixel layer.

Both types of isolation are used in the tag-and-probe measurements, mainly in order to tighten the selection criteria of the tag.

Whenever isolation is applied to the probe electron candidate in this work, this only happens in the J/ψ analysis, the criteria are chosen such that the effect on the measured identification efficiency is estimated to be small.

Max The efficiency εID of the algorithms used by ATLAS to identify photons during the LHC Run 1 has been measured from pp collision data using three independent methods in different photon ET ranges. The three measurements agree within their uncertainties in the overlapping ET ranges, and are combined.
For the data taken in 2011, 4.9 fb-1 at s√=7 TeV, the efficiency of the cut-based identification algorithm increases from 60– 70% at ET = 20 GeV up to 87–95% (90–99%) at ET > 100 GeV for unconverted (converted) photons. With an optimised neural network this efficiency increases from 85–90% at ET = 20 GeV to about 97% (99%) at ET > 100 GeV for unconverted (converted) photon candidates for a similar background rejection. For the data taken in 2012, 20.3 fb-1 at s√=8 TeV, the efficiency of a re-optimised cut-based photon identification algorithm increases from 50–65% (45–55%) for unconverted (converted) photons at ET = 10 GeV to 95–100% at ET > 100 GeV, being larger than ≈ 90% for ET > 40 GeV.
The nominal MC simulation of prompt photons in ATLAS predicts significantly higher identification efficiency values than those measured in some regions of the phase space, particularly at low ET. A simulation with shower shapes corrected forthe average shifts observed with respect to the data describes the values of εID better in the entire ET and η range accessible by the data-driven methods. The residual difference between the efficiencies in data and in the corrected simulation are taken into account by computing data-to-MC efficiency scale factors. These factors differ from one by up to 10% at ET = 10 GeV and by only a few percents above ET = 40 GeV, with an uncertainty decreasing from 1.4–4.5% (1.7– 5.6%) at ET = 10 GeV for unconverted (converted) photons to 0.2–0.8% (0.2–0.5%) at high ET for s√=8 TeV. The uncertainties are slightly larger for s√=7 TeV data due to the smaller size of the control samples.

(source)

How search engines are making us more racist – A brand new book shows how the Google’s search algorithms quietly reinforce racist stereotypes.

Safiya Umoja Noble

I started the book several years ago by doing collective searches on keywords around different community identities. I did searches on “black girls,” “Asian girls,” and “Latina girls” online and found that pornography was the primary way they were represented on the first page of search results.

That doesn’t seem to be a very fair or credible representation of women of color in the United States. It reduces them to sexualized objects.

They suppressed a lot of porn, in part because we’ve been speaking out about this for six or seven years. But if you go to Google today and search for “Asian girls” or “Latina girls,” you’ll still find the hypersexualized content.

I think what you see there is the gaze of people of color looking at white women and girls and naming whiteness as an identity, which is something that you don’t typically see white women doing themselves.

These search algorithms aren’t merely selecting what information we’re exposed to; they’re cementing assumptions about what information is worth knowing in the first place. That might be the most insidious part of this.

Safiya Umoja Noble
There is a dominant male, Western-centric point of view that gets encoded into the organization of information. You have to remember that an algorithm is just an automated decision tree. If these keywords are present, then a variety of assumptions have to be made about what to point to in all the trillions of pages that exist on the web.

And those decisions always correlate to the relationship of advertisers to the platform. Google has a huge empire called AdWords, and people bid in a real-time auction to optimize their content.

That model — of information going to the highest bidder — will always privilege people who have the most resources. And that means that people who don’t have a lot of resources, like children, will never be able to fully control the ways in which they’re represented, given the logic and mechanisms of how search engines work.

In the book, you talk about how racist websites gamed search engines to control the narrative around Martin Luther King Jr. so that if you searched for MLK, you’d find links to white supremacist propaganda.

You also talk about the stakes involved here, and point to Dylann Roof as an example.

Safiya Umoja Noble
In his manifesto, Dylann Roof has a diatribe against people of color, and he says that the first event that truly awakened him was the Trayvon Martin story. He says he went to Google and did a search on “black-on-white crime.” Now, most of us know that black-on-white crime is not an American epidemic — that, in fact, most crime happens within a community. But that’s a separate discussion.

So Roof goes to Google and puts in a white nationalist red herring (“black-on-white crime.”) And of course, it immediately takes him to white supremacist websites, which in turn take him down a racist rabbit hole of conspiracy and misinformation. Often, these racist websites are designed to appear credible and benign, in part because that helps them game the algorithms, but also because it convinces a lot of people that the information is truthful.

This is how Roof gets radicalized. He says he learns about the “true history of America,” and about the “race problem” and the “Jewish problem.” He learns that everything he’s ever been taught in school is a lie. And then he says, in his own words, that this makes him research more and more, which we can only imagine is online, and this leads to his “racial awareness.”

And now we know that shortly thereafter, he steps into the “Mother” Emanuel AME Church in Charleston, South Carolina, and murders nine African-American worshippers in cold blood, in order to start a race war.

So the ideas that people are encountering online really matter. It matters that Dylann Roof didn’t see the FBI statistics that tell the truth about how crime works in America. It matters that he didn’t get any counterpoints. It matters that people like him are pushed in these directions without resistance or context.

People who are a numerical minority in society will never be able to use this kind of “majority rules” logic to their benefit. The majority will always be able to control the notions of what’s important, or what’s important to click on, and that’s not how the information landscape ought to work.

The platform exists because it’s made by people. It didn’t come down from an alien spacecraft. It’s made by human beings, and the people who make it are biased, and they code their biases into search. How can these things not inform their judgment?

So it’s disingenuous to suggest that the platform just exists unto itself, and that the only people who can manipulate it or influence it are the people who use it, when actually, the makers of the platform are the primary source of responsibility. I would say that there are makers, as well as users, of a platform. They have to take responsibility for their creations.