Towards Scalable and Policy-Compliant Sensor Placement for Large-Scale Air Quality Monitoring


Vinayak Rana

M.Tech in Artificial Intelligence
Indian Institute of Technology Gandhinagar


Advisor: Prof. Nipun Batra

DSC Members: Prof. Manoj Gupta, Prof. Sameer Patel

Thesis Defense

Air Pollution: A Global Health Crisis

center

  • 7 million premature deaths annually (WHO)
  • More than malaria, HIV, and road accidents combined
  • 91% of deaths in low/middle-income countries

Current Air Quality Monitoring in India

Metric Value
CPCB stations ~600
Population 1.4 billion
People per station 2.3 million
Area per station 5,400 km²

Takeaway: Severely under-monitored for a country of this size and pollution levels

Why Not Just Add More Sensors?

Full CPCB CAAQMS (BAM + gases + shelter): ₹2-3 crore+ per station

  • Large-scale deployment → extremely costly

Current CPCB network: ~600 stations


Smart placement is critical — every sensor must maximize value!

The Core Question

Given a limited budget of new sensors:

Where should we place them?

Goal: Maximize information about the entire region

Problem Formulation

— Existing sensors

— Candidate locations ( points)

— Target region to predict

— Budget for new sensors

Goal: Select from to maximize info about

Optimal Sensor Placement (OSP)

1. Surrogate Model — Predicts values + uncertainty; must be differentiable
2. Acquisition Function — Scores candidates; guides placement

Acquisition 1: Random

Pick locations uniformly at random from candidates

In plain English:

"Close your eyes and point at the map times"

  • No optimization — pure luck
  • Ignores data completely
  • Baseline for comparison

Acquisition 2: Maximum Variance (MaxVar)

Greedily pick the location with highest uncertainty

In plain English:

"Where are we most uncertain? Put a sensor there."

Repeat times, each time adding the selected sensor to context.

Pros: Fast , intuitive

Cons: Doesn't maximize information gain with minimum sensors

The MI Objective: What We Want to Maximize

Mutual Information: How much does knowing reduce uncertainty about ?

Reading this: "Given existing sensors , how much does adding new sensors reduce our uncertainty about the target region?"

MI vs MaxVar: What's the Difference?

For Gaussian outputs (GP/Neural Process): → both use variance

  • MaxVar: "where am I most uncertain?"
  • MI: "what reduces uncertainty everywhere?"

Problem: Exact MI requires searching subsets — combinatorially explosive!

Acquisition 3: Greedy MI (Standard Approach)

Since exact MI is intractable, use greedy approximation: select one sensor at a time

Algorithm: For each of rounds, evaluate all candidates → pick best

Complexity: evaluations

(candidates) (budget) Evaluations
1,000 50 50,000
20,000 100 2,000,000

For India (): computationally infeasible!

The Challenge

Greedy MI gives best quality but doesn't scale

MaxVar scales but gives poor quality

Can we get both?

GD-MI: Gradient-Based Mutual Information Maximization

Scalable sensor placement via continuous optimization




Publication (AAAI 2026 — AI for Social Impact)

Scalable Air-Quality Sensor Placement via Gradient-Based Mutual Information Maximization

Zeel B Patel, Vinayak Rana, Nipun Batra

Our Solution: GD-MI

Key insight: Don't search discrete candidates — optimize coordinates directly!

Greedy MI GD-MI (Ours)
Search space Discrete ( candidates) Continuous (coordinates)
Optimization Enumerate all Gradient descent
Complexity iterations
Scalability ❌ ✓ Any


Greedy MI: Search grid


GD-MI: Follow gradient

GD-MI: How It Works

What we optimize: sensor locations, each a (lat, lon) pair

Objective: Minimize average variance over target region

where — predicted values at proposed locations

Algorithm: Initialize → Forward (get ) → Backward () → Update → Repeat

Key: Model is frozen — only the coordinates are optimized

Surrogate Model: Why Neural Processes?

GD-MI needs a model that provides:

Requirement Why?
Predictions To impute values at proposed locations
Calibrated uncertainty To compute information gain
Differentiable To backpropagate through
Fast inference To iterate quickly

TNP-D (Transformer Neural Process) satisfies all!

Surrogate Model: Transformer Neural Process (TNP-D)

  • Predicts PM and uncertainty
  • Fully differentiable
  • Fast parallel inference
Model RMSE ↓ NLL ↓
Gaussian Process 5.16 -0.19
Convolutional GNP 5.31 -0.30
Transformer NP 4.90 -0.44

Experiments

Does GD-MI actually work?

Experiment 1: Regional Validation

Madhya Pradesh — a central Indian state (~size of Germany, =308 candidates)

Results (k=9 sensors):

Method RMSE ↓
Random 7.2
MaxVar 6.3
GD-MI 5.8
Greedy MI 5.6

GD-MI within 0.2 RMSE of Greedy MI (gold standard)

Experiment 2: India-Scale (The Real Test)

Greedy MI infeasible () — GD-MI shines here

Key findings:

  • GD-MI 4% better than MaxVar
  • Gap grows with more sensors
  • Consistent across budgets

Why Does GD-MI Win? Qualitative Analysis

MaxVar (blue dots)

  • Optimizes for single-point uncertainty
  • Doesn't consider target region

GD-MI (red dots)

  • Optimizes for entire target region
  • Maximizes information gain

MI objective guides placement for maximum coverage

Scalability: GD-MI vs Greedy MI

center

  • Greedy MI: runtime grows with pool size
  • GD-MI: constant time regardless of
  • At : GD-MI is ~100× faster

GD-MI: Limitations

  • Optimizes in continuous space
    → cannot enforce real-world deployment constraints

  • May select invalid / impractical locations
    (e.g., inaccessible terrain)

  • Assumes uniform importance
    → ignores population / exposure differences

  • Non-convex optimization
    → solutions depend on initialization

GSM: Gumbel-Softmax for Constrained Sensor Placement

Differentiable discrete selection under regional constraints




Publication (Under Review)

Large-Scale Air-Quality Sensor Placement via Joint Optimization under Regional Constraints

Vinayak Rana, Neerja Kasture, Anura Mantri, Nipun Batra

Regional Budget Constraints

center

Sensors are allocated per state

  • budget per state
  • Select from candidate pool

Constraint:
exactly per state

GD-MI cannot enforce: "exactly sensors in state "

From Constraints to Selection

Each sensor must choose one location from its state's candidate pool


center

From Constraints to Selection

Represent this as a probability distribution over candidates


center

From Constraints to Selection

argmax picks the highest probability location


center

From Constraints to Selection

But argmax is not differentiable — gradients cannot flow through it


center

How do we make sampling differentiable?

Gumbel-Softmax: Making Sampling Differentiable

center

GSM: How It Works


Don't worry about the complexity — let's go term by term

GSM: How It Works


— learnable logits, one row per sensor, one column per candidate location

The learnable logit matrix W

center

GSM: How It Works


— regional mask: sets out-of-state entries to → probability = 0

Enforcing regional budgets via masking

center

GSM: How It Works


— Gumbel noise: constant in backward pass — this is the reparameterisation trick

GSM: How It Works


— temperature: high → explore broadly, low → commit to one location

GSM: How It Works


  • — learnable logits
  • — regional mask
  • — Gumbel noise
  • — temperature

Does It Converge? Training Dynamics

center

(a) Objective + temperature

  • Predictive variance falls steadily
  • High → exploration; low → exploitation

(b) Location update count

  • Sensors move frequently early on
  • Updates taper to zero as

No post-hoc projection needed — the relaxation self-stabilizes into a valid discrete solution

Experiments

Does GSM actually work?

Regional Validation: GSM vs Baselines

center

Method RMSE ↓
Random 6.50
MaxVar 6.41
GSM (Ours) 6.12
Greedy MI 5.74

GSM closes most of the gap to Greedy MI — at far lower cost

Continental Results: Unconstrained Deployment

No state budget constraints — pure joint optimization

  • GSM lowest at every budget
  • Gap widens with more sensors
  • Greedy methods pick redundant locations

Main Result: Constrained Deployment

Strategy All-India RMSE ↓
Existing network 9.25 ± 0.00
Random 7.69 ± 0.15
Greedy MaxVar 7.46 ± 0.00
GSM Indep (ours) 7.42 ± 0.09
GSM Joint (ours) 7.27 ± 0.03
  • 21.4% reduction in national prediction error over existing network

  • 13/32 states where GSM Joint achieves lowest RMSE

  • Joint > Indep — cross-border gradients capture transboundary pollution structure

Why does GSM Joint perform better?

Qualitative Comparison of Deployments

center

GSM: Limitations

  • Uses a continuous relaxation
    → introduces approximation error

  • Sensitive to temperature (τ) scheduling
    → affects convergence stability

  • Non-convex optimization over logits
    → sensitive to initialization; multiple local optima

Future Work

  • Equity-aware placement
    → weight by population, pollution, vulnerability

  • Dynamic sensing
    → mobile sensors, adaptive placement

  • Multi-pollutant optimization
    → PM₂.₅, NO₂, O₃ jointly

  • Improved discrete relaxations
    → Top-K, optimal transport (Sinkhorn)

Publications

Scalable Air-Quality Sensor Placement via Gradient-Based Mutual Information Maximization (AAAI 2026 — AI for Social Impact)
Zeel B Patel, Vinayak Rana, Nipun Batra


Large-Scale Air-Quality Sensor Placement via Joint Optimization under Regional Constraints (Under Review)
Vinayak Rana, Neerja Kasture, Anura Mantri, Nipun Batra

# The Coverage Gap <div class="cols"> <div class="col"> <img src="assets/images/india_urban_only_coverage.png" style="max-height:340px;" /> </div> <div class="col"> **Urban-only coverage** - Sensors in Delhi, Mumbai, Chennai, Bangalore, Kolkata - Rural India remains **invisible** **Hundreds of millions** unmonitored No data → No policy → No protection </div> </div> ---

# Global Comparison: Room for Growth | Country | Stations ↑ | People/Station ↓ | Stations/1000 km² ↑ | |---------|----------:|--------------:|-----------------:| | USA | 4,800 | 69K | 0.49 | | China | 5,000 | 280K | 0.52 | | Germany | 500 | 166K | 1.40 | | UK | 300 | 223K | 1.23 | | **India** | **611** | **2,290K** | **0.19** | India has **33× more people per station** than USA — significant opportunity for expansion ---

# From Prediction to Placement > Not all locations are equally informative <br> - Some regions are **well-understood** - Some regions are **uncertain** <br> > Place sensors where they reduce uncertainty the most ---

| | MaxVar | MI | |:--|:------:|:--:| | Optimizes for | Single point | <span class="blue">Entire target region</span> | | Variance | **at candidate** | **over target region** |

---

![bg contain](assets/images/sustainability_lab_prompt_subtitles_generated_20251217_162815.png)

---

# Main Takeaways 1. **GD-MI** = First gradient-based MI maximization for sensor placement 2. **Scalability breakthrough:** $O(I)$ vs $O(n \cdot k)$ - Enables **continental-scale** optimization 3. **Quality preserved:** Matches Greedy MI where tractable, **4% better** than MaxVar at scale 4. **Real-world ready:** Deployed framework for India air quality monitoring

---

# Thank You! <div style="display:flex; justify-content:center; gap:3rem; margin:1.5rem 0;"> <div class="qr" style="text-align:center;"> <img src="assets/images/qr_paper.png" /> <div style="font-size:0.7em; margin-top:0.3rem;">Paper</div> </div> <div class="qr" style="text-align:center;"> <img src="assets/images/qr_lab.png" /> <div style="font-size:0.7em; margin-top:0.3rem;">Lab</div> </div> </div> **Sustainability Lab @ IIT Gandhinagar** Positions: PhD · Postdoc · RA · Intern {patel_zeel, vinayak.rana, nipun.batra}@iitgn.ac.in

---

_paginate: false # Backup Slides --- # Impact: Same Budget, Better Outcomes **The cascade effect:** Better placement (GD-MI) ↓ Better predictions (lower RMSE) ↓ Better pollution maps (policy-ready) ↓ Better health outcomes (targeted interventions) > **4% RMSE improvement** at national scale = **millions** of people better served --- # Dataset: WUSTL PM₂.₅ - **Source:** Washington University in St. Louis - **Resolution:** 0.1° × 0.1° (~11 km) - **Period:** 1998-2018 (21 years monthly) - **Split:** Train 98-08 | Val 09-10 | Test 11-18 --- # Full Model Benchmark | Model | NLL ↓ | RMSE ↓ | |-------|:-----:|:------:| | CNP | 0.48 | 11.46 | | Random Forest | -0.11 | 6.55 | | GP | -0.19 | 5.16 | | ConvCNP | -0.27 | 5.28 | | ConvGNP | -0.30 | 5.31 | | TabPFN | -0.37 | 5.09 | | **TNP-D** | **-0.44** | **4.90** | --- # Constraint: Keep Sensors on Land $$\mathcal{L}_{\text{OOR}} = \sum_i \exp\Big(\text{dist}(\color{#00a651}{x_i}, \text{land}) - \delta\Big) - 1$$ Soft penalty grows exponentially as <span class="green">sensors</span> drift toward ocean --- # Scalability Analysis ![width:600px center](assets/images/scalability.png) - GD-MI runtime **independent of pool size** - At 20K candidates: GD-MI is **100× faster** than Greedy MI - Enables continental-scale optimization --- # More Results: k = 50 ![width:700px center](assets/images/quality_50_india.png) --- # More Results: k = 100 ![width:700px center](assets/images/quality_100_india.png)

> Same Madhya Pradesh setup ($n=308$, $k=8$)