87 Matching Annotations
  1. Jan 2023
  2. Dec 2022
    1. First, theremay be irreducible noise in the measurement process itself

      if we can learn to upscale, does this mean that we can reduce this "irreducible noise"? Of course, at some other point in time, better measurements would have to have been performed, but if we can learn the relationship between poor, cheap measures and good, expensive measures, then we can variationally approximate the mapping from poor to good measurements with a possibly lower increase in cost. In a collective intelligence, it might be that the irreducible noise comes not from the current measurement process, but the best measurement process available in the network.

  3. Oct 2022
    1. In general, when all agents learn simultaneously, the found best response may not beunique (Shoham and Leyton-Brown 2008).

      This links to a manual. Not sure of where in the book to find the answer :(

    2. A special case of the MG is the stateless setting X = � called strategic-form game 2 .Strategic-form games describe one-shot interactions where all agents simultaneouslyexecute an action and receive a reward based on the joint action after which the gameends.

      Climate change as a strategic-form game?

  4. www.explainpaper.com www.explainpaper.com
    1. Moreover, these stepsare not themselves rewarding and they will be separated from each other and any reward bylong periods of time, during which you will engage with other life tasks.

      They are rewarding if you learn during them and they are reusable in other contexts

  5. Aug 2022
    1. This individualization, if it takes place, triggers its own fairness anddiscrimination issues. First, since it would lead to more disparate pricing, individuals foundas “the riskiest” by the algorithm are likely to get unaffordable premiums, hence excludingthem from the insured community.

      missing citation

    2. or Simon(1988) and Horan (2021), the adoption of this individual point of view by the judge in theNorris case has contributed to reinforcing the erasure of the principle of solidarity whichis at the heart of insurance practice (Lehtonen and Liukko (2011)).

      Norris case makes no sense in this context since you left out a key phrase from the original article

    3. Measuring the impact on insuranceis however maybe more difficult than in other areas. On the one hand, like any otherorganization, insurers are pushed to change their practices to incorporate the new sources ofdata that have become available, the increased computing capacity, and the new algorithms.

      also a copy-paste

    4. Recently, the emergence of big data and new algorithmson insurance actually revived and renewed, with obvious points of continuity and rupture,older debates linked to discrimination in insurance. Conceptually however, these changes areshaking up insurance practice: the ’individualization’ of risks, rendered possible by thesenew technologies, is indeed now considered as fairer than their pooling. Finally, the last partdiscusses the ethical issues of big data for insurance, in comparison with various notions ofalgorithmic fairness.

      also a near copy-paste

    5. Since the beginning of their history, insurers have been quantifying human phenomena,collecting and using data to classify and price risks. This practice is not trivial: insuranceplays a major role in industrialized societies in opening or closing life opportunities (Horan(2021) or Baker and Simon (2002)). As such, insurers were confronted early on with theequity issues associated with data. As early as 1909, the Kansas insurance regulator thusdrew the contours of fair rating practice: he defined a rating as "not unfairly discriminatory"if it treats similar people in the same way (Frezal and Barry (2019), Miller (2009)) orBarry and Charpentier (2022).

      is this literally a copy-paste from Barry and Charpentier?

    Annotators

    1. n practice, regional differences in investment costsexist, but since renewable investment costs are empirically shown to be driven down byglobal cumulative installed capacity – in a process of global “learning” or “experience”(Hepburn et al. (2020), Way et al. (2021)) – the global average represents a robust proxy

      reference to experience curves

    2. To calculate investment costs for differenttypes of renewable energy investments needed to replace coal we use data from IRENA(2021b).

      investment costs for different types of renewables

    Annotators

  6. Jun 2022
    1. ntation details. • Data preparation: All our timeseries data that enter to CALIBNN are padded with a mini-mum sequence length of 20 for COVID and 5 for influenza.We also normalize each features with mean 0 and variance1. To inform CALIBNN of the county to predict, we usea one hot encoding for counties. • Architecture details:In CALIBNN, the encoder is a 2-layer bidirectional GRUand decoder is a 1-layer bidirectional GRU, both with hid-den size of 32. Output layer has two linear layers of size32x16xD with ReLU activation function and D is ABMparameter dimensions dimensions: D = 3 for COVID-19and D = 2 for flu. • Hyperparameters: We found a learningrate of 10−3, Xavier initialization, and the Adam optimiza-tion algorithm work best. • ABM parameters (θP , θT ):For COVID-19, we have three parameters: R0, mortalityrate, and initial infections percentage. These are boundedwith θL = [1.0, 0.001, 0.01] and θU = [8.0, 0.02, 1.0]. Forflu, we have two parameters: R0 and initial infections per-centage. These are bounded with θL = [1.05, 0.1] andθU = [2.6, 5.0]. The initial infections percentage is thepercentage of the population that is infected at time stept = 0 of the simulation. • ABM clinical parameters: Toset some of the parameters of our transmission and progres-sion models, we utilize clinical data from reliable sources.Specifically, for COVID-19 we use age-stratified suscep-tibility and parameters of a scaled gamma distribution torepresent infectiousness as a function of time as per ( 35 ; 31 ).For influenza, we set those parameters based on CDC flufacts (1).Target variables. The target variable for COVID-19 isCOVID-associated mortality, while in flu we use influenza-like-illness (ILI) counts, which is collected by the CDC. ILImeasures the percentage of healthcare seekers who exhibitinfluenza-like-illness symptoms, defined as ”fever (tempera-ture of 100°F/37.8°C or greater) and a cough and/or a sorethroat without a known cause other than influenza” ( 32).The ground truth data for the target variables is obtaineduploads/MMWR_Week_overview.pdf4https://predict.cdc.gov/from JHU CSSE COVID-19 data repository and ILI fromCDC influenza dashboard.Details on baseline implementation. Vanilla-ABM: Asnoted in the main paper, R0 and case-fatality rate are ob-tained from authoritative sources. To set the initial infectionspercentage for this baseline, we set it to the mean value ofthe search range. PC-ABM: We present details on each ofthe ODE models we used for this baseline.(COVID-19) SEIRM (63): The SEIRM model consists offive compartments: Susceptible (S), Exposed (E), Infected(I), Recovered (R), and Mortality (M ). It is parameterizedby four variables Ω = {β, α, γ, μ}, where β is the infectiv-ity rate, 1/α is the mean latent period for the disease, 1/γ isthe mean infectious period, and μ is the mortality rate. Thebasic reproductive number R0 = β/(γ + μ).dStdt = −βtStItNdEdt = βtStItN − αtEt (9)dItdt = αtEt − γtIt − μtItdRtdt = γtItdMtdt = μtIt(Flu) SIRS (58): This model consists of three compartments:Susceptible (St), Infected (It), and Recovered (Rt). It isparameterized by three variables Ω = {β, D, L}, where βis the infectivity rate, D is the mean duration of immunity,and L is the mean duration of the immunity period. Thebasic reproductive number R0 = βD.dStdt = N − St − ItLt− βtItStN (10)dItdt = βtItStN − ItDtD. More details on evaluating policyinterventionsWe reproduce the experimental setup in ( 53). For inter-ventions, we simulate standard covid-19 vaccination ver-sus delayed second dose vaccination prioritizing the firstdose. Sensitivity analyses included first dose vaccine effi-cacy of 50%, 60%, 70%, 80%, and 90% after day 12 post-vaccination with a vaccination rate of 0.3% population perday; assuming the vaccine prevents only symptoms but notasymptomatic spread (that is, non-sterilizing vaccine). Wemeasure cumulative covid-19 mortality, cumulative SARS-CoV-2 infections, and cumulative hospital admissions due tocovid-19 over 74 days. We explicitly modeled the confirma-tion of infections with polymerase chain reaction testing andquarantining of known infected agents with imperfect com-pliance over time. To simulate a natural pattern of infectionat the point vaccinations begin, we started our simulationwith 10 agents infected and ran the simulation for 20 daysbefore starting vaccinations, which corresponds to a cumu-lative infection rate of 1%, similar to the one in the US,

      giant hyperlink

    2. his performance efficiencycan be attributed to the design of a sparse tensor-calculusbased implementation of GRADABM.

      an ablation study of the impact of sparsity of the interactions would definitely be a plus

    3. These unseen sce-narios may include predicting future the evolution of multi-ple infections for a county/region, or it may involve makingpredictions for a new county/region for which historical datafor calibration is not available, or it may involve makingpredictions even when the data (inputs to CALIBNN) for acounty/region is noisy. I

      dubious generalization claims. How you can assume transferability of contact graphs across countries?

    4. The Transmission Model is differentiable as the functionscomprising it, which are: λ is a smooth function, ⋃ is apermutation invariant function (∑) and is linear.

      weird syntax

    5. he output of GRADABM is aggregated to compute cumula-tive deaths. This macro value is compared with correspond-ing ground truth values and the following loss is computed:L(ˆyw, y; (θtT , θtP )w) = MSE(ˆyw, y),

      fairly simple loss function

    6. here Update(Xti , Ni, (Xtj )j∈Ni , θtT , θtP ) is={Transmit(Xti , Ni, (Xtj )j∈Ni , θtT ), if dti = S,Progress(Xti , θtP ), if dti ∈ {E, I

      So there's no reinfection in this model? what happens for R and M?

    7. s. The inputto the simulator are the time-dependent parameters whichgovern the transmission and progression models, denotedas θtT = [Rt, S, T ], θtP = [m, τEI , τIR, τIM ] respectively

      This doesn't look like many more parameters than an SIR model

    8. They key difference between our model and the SEIRM likemodels is that we track states of individual agents ratherthan just the aggregate numbers of agents in various dis-eases stages. Furthermore, we model the S → E exposurestep, using interaction networks which take into accountlocal agent interactions, which is different from the standardSEIRM like models.

      How do you calibrate the individual agents' progression? As I understand it, SEIR models are limited by how many parameters can be tuned. So, you model SEIRM for each individual, and therefore tune the progression model for each individual, which gives your progression model more parameters. However, does the data you use for fitting actually allow you to look at individuals and properly tune the model such that the distribution of parameters for the personalized progression model is roughly the same as what can be found in the population? If not, wouldn't this impact the relevance of emerging phenomena from the simulator to the real world?

    9. The probability of transmission (q(.; .)) from the interactionis represented as: q(dti, dtj ) = 1 − e−λ(R,Si,Tj ,∆Etj ), where,λ(R, Si, Tj , ∆Etj ) = RSiTjˆIi∫ ∆Etj∆Etj −1 GΓ(u; μj , σ2j )d

      confusing notation given that S, R and E and I appear in the agent state

    10. we observethat epidemiological models ( 22 ; 36 ; 4; 53 ) also adhere topermutation invariance wherein the order of infectious in-teractions (pair-wise message passing) within a step doesnot matter when estimating the probability of infection fromthose interactions.

      very cool claim, is it backed up?

    11. Efficacy of DNNarchitectures results from overcoming the curse of dimen-sionality by leveraging the pre-defined invariances arisingfrom the underlying low-dimensionality and structure of thephysical world (17 ).

      Weird reference for this statement. The efficacy of what? DNN learning? Efficacy is a fairly broad term, and these invariances are leveraged in certain network structures (like CNNs). Paper points toward geometric deep learning exactly to take advantage of structured neural nets.

    12. n contrast, the differentiable design of GRADABMallows to calibrate parameters with gradient-based optimiza-tion inside the ABM itself. This helps learn personalizedparameters resulting in better (and robust) forecasting andpolicy decision making.

      claim

    13. 6. DiscussionWe introduce GRADABM, a differentiable ABM that cansimulate million-scale populations in few seconds on com-modity hardware and be merged with DNNs for end-to-endlearning. The key idea of GRADABM is a general-purposedifferentiable sparse tensor-calculus based implementationwhich we validate for epidemiology here. We demonstrateutility of GRADABM to learn personalized time-varyingparameters using heterogeneous data sources to better in-form forecasting and policy decision making. Future workcould explore other benefits like the incorporation of mul-tiple hierarchies of data (also known as macro-micro pre-dictions (64 )). In our experiments, we found that even withsparse optimization, GRADABM utilizes a high amount ofCPU/GPU memory (especially in the backward computa-tion) which could be an issue for simulations of 6+ months.Addressing this limitation could be an interesting futuredirection. Our ABM used a a linear deterministic model forthe disease progression. Future work could explore how toincorporate more complex and stochastic disease progres-sion models (e.g., (53))

      discussion does not mention contributions 2 and 3

    14. Our Contributions: (i) We present GRADABM, a differen-tiable ABM that can simulate million-scale populations infew seconds on commodity hardware and be merged withDNNs for end-to-end learning. The key idea of GRADABMis a general-purpose differentiable sparse tensor-calculusbased implementation which we validate for epidemiologyhere. (ii) We demonstrate utility of GRADABM for robustforecasting and analysis of COVID-19 and Influenza. (iii)We show the use of GRADABM in evaluating pharmaceuti-cal interventions for policy decision making

      contributions

    15. agents are represented as tensors,their interaction networks as (sparse) adjacency matricesand a continuous relaxation of the stochastic disease trans-mission model is used to produce gradient estimates withautomatic differentiation.

      breakdown

    16. However, they are conven-tionally slow, difficult to scale to large population sizes ( 45 )and tough to calibrate with real-world data (29). This is achallenge since simulation results (emergent behavior) canbe highly sensitive to the scale of the input population andcalibration of the input parameters. In addition, incorporat-ing novel sources of data that could inform calibration andother downstream tasks (e.g., forecasting) is often laboriousand adds overhead complexity to the ABM (e.g., incorpo-rating digital exposure data to ABMs ( 36 )). In this paper,we introduce GRADABM to alleviate these concerns

      claims

    17. Mechanistic simulators are an indispensable tool for epi-demiology to explore the behavior of complex, dynamicinfections under varying conditions and navigate uncer-tain environments. ODE-based models are the dominantparadigm that enable fast simulations and are tractable togradient-based optimization, but make simplifying assump-tions about population homogeneity. Agent-based models(ABMs) are an increasingly popular alternative paradigmthat can represent the heterogeneity of contact interactionswith granular detail and agency of individual behavior. How-ever, conventional ABM frameworks are not differentiableand present challenges in scalability; due to which it is non-trivial to connect them to auxiliary data sources easily. Inthis paper we introduce GRADABM which is a new scal-able, fast and differentiable design for ABMs. GRADABMruns simulations in few seconds on commodity hardwareand enables fast forward and differentiable inverse simu-lations. This makes it amenable to be merged with deepneural networks and seamlessly integrate heterogeneousdata sources to help with calibration, forecasting and policyevaluation. We demonstrate the efficacy of GRADABM viaextensive experiments with real COVID-19 and influenzadatasets. We are optimistic this work will bring ABM andAI communities closer together.

      repetition of the abstract

    18. GRAD-ABM which is a new scalable, fast and differen-tiable design for ABMs. GRADABM runs sim-ulations in few seconds on commodity hardwareand enables fast forward and differentiable in-verse simulations

      contribution

    19. However, conven-tional ABM frameworks are not differentiable andpresent challenges in scalability; due to which itis non-trivial to connect them to auxiliary datasources easily.

      Context

    1. The content diversity or the variance of all rec-ommended content at each timestep to measure thedegree of homogenization of visible content similar to(Chaney et al., 2018; Lucherini et al., 2021)

      metric : formulas would be nice

    2. generalize the components ofour simulator to different social media platforms to describehow our software serves as a framework for analyzing theimpact of coordinated attacks on end-users of any of theseplatforms

      claim

    3. collect millions ofreal-world interactions from Reddit to estimatethe network for each user in our dataset and utiliseReddit’s self-described content ranking strategiesto compare the impact of coordinated activity oncontent spread by each strateg

      methodology

    4. Wedevelop a multiagent simulation of a popular so-cial network, Reddit, that aligns with the state-action space available to real users based on theplatform’s affordances.

      claim 3: they develop an ABM to imitate Reddit

    5. simulations have emerged as a populartechnique to study the long-term effects of con-tent ranking and recommendation systems

      claim 3: simulations are a popular tool for studying the effects and impacts of content ranking and recommendation systems

    6. it isunclear how effective countermeasures to disinfor-mation are in practice due to the limited view wehave into the operation of such platforms

      claim 2: the effectiveness of countermeasures to disinformation is unclear due to limited data about the operation of such models

    7. he vulnerabil-ity of ranking and recommendation algorithmsto attack from coordinated campaigns spreadingmisleading information has been established boththeoretically and anecdotally

      claim : the vulnerability of ranking and recommendation systems has been established theoretically and empirically

    1. After N attackers contribute private datasets Di to jointdataset D,

      Perhaps a link to model calibration and training? Or to RL agents interacting with an environment and purposely generating poisoned data?

    2. To achieve this attack objective,a common attack vector is through that of outsourced datacollection

      Problem with relevance: this involves a third party that trains the model before returning it, or that provides an untrusted dataset. This is more like ABM4AI than AI4ABM

    3. n this work, we evaluate existing single-agent back-door defenses to find that they cannot increase the accuracyw.r.t. clean labels during a backdoor attack, and furthermorecontribute three multi-agent backdoor defenses motivatedby agent dynamics. In addition, we are among the firstto construct multi-distributional subspaces to tackle jointdistribution shifts.

      no positioning wrt SOTA, but a claim of innovation for multi-distributional subspaces

    4. This yieldsa side-effect of low accuracy w.r.t. clean labels,which motivates

      Claim 1: the backfiring effect is a natural defense against backdoor attacks and yields a side-effect of low accuracy wrt clean labels

    5. this paper’s work on the construc-tion of multi-agent backdoor defenses that max-imize accuracy w.r.t. clean labels and minimizethat of poison labels.

      this paper works to construct multi-agent backdoor defenses that maximize accuracy of clean labels and minimize accuracy of poison labels

    6. the backfiring effect, a natu-ral defense against backdoor attacks where back-doored inputs are randomly classified

      Context : the backfiring effect is a natural defense against the multi-agent backdoor attack which results in low accuracy for clean labels