Srinivasa Narasimhan

Srinivasa Narasimhan

Carnegie Mellon University

H-index: 52

North America-United States

About Srinivasa Narasimhan

Srinivasa Narasimhan, With an exceptional h-index of 52 and a recent h-index of 36 (since 2020), a distinguished researcher at Carnegie Mellon University, specializes in the field of Computer vision, computer graphics, computational photography, sensors, computational imaging.

His recent articles reflect a diverse array of research interests and contributions to the field:

Toward Planet-Wide Traffic Camera Calibration

A single dose for me, A wealth of protection for us: The public health cost of individualism in the rollout of COVID-19 vaccine

TPSeNCE: Towards Artifact-Free Realistic Rain Generation for Deraining and Object Detection in Rain

Robot Safety Monitoring using Programmable Light Curtains

WALT3D: Generating Realistic Training Data from Time-Lapse Imagery for Reconstructing Dynamic Objects under Occlusion

Addressing Source Scale Bias via Image Warping for Domain Adaptation

One-Step Image Translation with Text-to-Image Models

Learned two-plane perspective prior based image resampling for efficient object detection

Srinivasa Narasimhan Information

University

Carnegie Mellon University

Position

Professor of Robotics

Citations(all)

16906

Citations(since 2020)

7300

Cited By

13284

hIndex(all)

52

hIndex(since 2020)

36

i10Index(all)

101

i10Index(since 2020)

79

Email

University Profile Page

Carnegie Mellon University

Srinivasa Narasimhan Skills & Research Interests

Computer vision

computer graphics

computational photography

sensors

computational imaging

Top articles of Srinivasa Narasimhan

Toward Planet-Wide Traffic Camera Calibration

Authors

Khiem Vuong,Robert Tamburo,Srinivasa G Narasimhan

Published Date

2024

Despite the widespread deployment of outdoor cameras, their potential for automated analysis remains largely untapped due, in part, to calibration challenges. The absence of precise camera calibration data, including intrinsic and extrinsic parameters, hinders accurate real-world distance measurements from captured videos. To address this, we present a scalable framework that utilizes street-level imagery to reconstruct a metric 3D model, facilitating precise calibration of in-the-wild traffic cameras. Notably, our framework achieves 3D scene reconstruction and accurate localization of over 100 global traffic cameras and is scalable to any camera with sufficient street-level imagery. For evaluation, we introduce a dataset of 20 fully calibrated traffic cameras, demonstrating our method's significant enhancements over existing automatic calibration techniques. Furthermore, we highlight our approach's utility in traffic analysis by extracting insights via 3D vehicle reconstruction and speed measurement, thereby opening up the potential of using outdoor cameras for automated analysis.

A single dose for me, A wealth of protection for us: The public health cost of individualism in the rollout of COVID-19 vaccine

Authors

Wei Fu,Li-San Wang,Shin-Yi Chou

Journal

Social Science & Medicine

Published Date

2024/5/1

ObjectiveThis study examines whether individualism weakens the effectiveness of the COVID-19 vaccine eligibility expansions in the United States in 2021, and assesses the associated social benefits or costs associated with individualism.MethodsWe construct a county-level composite individualism index as a proxy of culture and the fraction of vaccine eligible population as a proxy of vaccination campaign (mean: 41.34%). We estimate whether the COVID-19 vaccine eligibility policy is less effective in promoting vaccine coverage, reducing in COVID-19 related hospitalization and death using a linear two-way fixed effect model in a sample of 2866 counties for the period between early December 2020 and July 1, 2021. We also test whether individualism shapes people's attitudes towards vaccine using a linear probability model in a sample of 625,308 individuals aged 18–65 (mean age: 43.3; 49% male; 59.1 …

TPSeNCE: Towards Artifact-Free Realistic Rain Generation for Deraining and Object Detection in Rain

Authors

Shen Zheng,Changjie Lu,Srinivasa G Narasimhan

Published Date

2024

Rain generation algorithms have the potential to improve the generalization of deraining methods and scene understanding in rainy conditions. However, in practice, they produce artifacts and distortions and struggle to control the amount of rain generated due to a lack of proper constraints. In this paper, we propose an unpaired image-to-image translation framework for generating realistic rainy images. We first introduce a Triangular Probability Similarity (TPS) constraint to guide the generated images toward clear and rainy images in the discriminator manifold, thereby minimizing artifacts and distortions during rain generation. Unlike conventional contrastive learning approaches, which indiscriminately push negative samples away from the anchors, we propose a Semantic Noise Contrastive Estimation (SeNCE) strategy and reassess the pushing force of negative samples based on the semantic similarity between the clear and the rainy images and the feature similarity between the anchor and the negative samples. Experiments demonstrate realistic rain generation with minimal artifacts and distortions, which benefits image deraining and object detection in rain. Furthermore, the method can be used to generate realistic snowy and night images, underscoring its potential for broader applicability. Code is available at https://github. com/ShenZheng2000/TPSeNCE.

Robot Safety Monitoring using Programmable Light Curtains

Authors

Karnik Ram,Shobhit Aggarwal,Robert Tamburo,Siddharth Ancha,Srinivasa Narasimhan

Journal

arXiv preprint arXiv:2404.03556

Published Date

2024/4/4

As factories continue to evolve into collaborative spaces with multiple robots working together with human supervisors in the loop, ensuring safety for all actors involved becomes critical. Currently, laser-based light curtain sensors are widely used in factories for safety monitoring. While these conventional safety sensors meet high accuracy standards, they are difficult to reconfigure and can only monitor a fixed user-defined region of space. Furthermore, they are typically expensive. Instead, we leverage a controllable depth sensor, programmable light curtains (PLC), to develop an inexpensive and flexible real-time safety monitoring system for collaborative robot workspaces. Our system projects virtual dynamic safety envelopes that tightly envelop the moving robot at all times and detect any objects that intrude the envelope. Furthermore, we develop an instrumentation algorithm that optimally places (multiple) PLCs in a workspace to maximize the visibility coverage of robots. Our work enables fence-less human-robot collaboration, while scaling to monitor multiple robots with few sensors. We analyze our system in a real manufacturing testbed with four robot arms and demonstrate its capabilities as a fast, accurate, and inexpensive safety monitoring solution.

WALT3D: Generating Realistic Training Data from Time-Lapse Imagery for Reconstructing Dynamic Objects under Occlusion

Authors

Khiem Vuong,N Dinesh Reddy,Robert Tamburo,Srinivasa G Narasimhan

Journal

arXiv preprint arXiv:2403.19022

Published Date

2024/3/27

Current methods for 2D and 3D object understanding struggle with severe occlusions in busy urban environments, partly due to the lack of large-scale labeled ground-truth annotations for learning occlusion. In this work, we introduce a novel framework for automatically generating a large, realistic dataset of dynamic objects under occlusions using freely available time-lapse imagery. By leveraging off-the-shelf 2D (bounding box, segmentation, keypoint) and 3D (pose, shape) predictions as pseudo-groundtruth, unoccluded 3D objects are identified automatically and composited into the background in a clip-art style, ensuring realistic appearances and physically accurate occlusion configurations. The resulting clip-art image with pseudo-groundtruth enables efficient training of object reconstruction methods that are robust to occlusions. Our method demonstrates significant improvements in both 2D and 3D reconstruction, particularly in scenarios with heavily occluded objects like vehicles and people in urban scenes.

Addressing Source Scale Bias via Image Warping for Domain Adaptation

Authors

Shen Zheng,Anurag Ghosh,Srinivasa G Narasimhan

Journal

arXiv preprint arXiv:2403.12712

Published Date

2024/3/19

In visual recognition, scale bias is a key challenge due to the imbalance of object and image size distribution inherent in real scene datasets. Conventional solutions involve injecting scale invariance priors, oversampling the dataset at different scales during training, or adjusting scale at inference. While these strategies mitigate scale bias to some extent, their ability to adapt across diverse datasets is limited. Besides, they increase computational load during training and latency during inference. In this work, we use adaptive attentional processing -- oversampling salient object regions by warping images in-place during training. Discovering that shifting the source scale distribution improves backbone features, we developed a instance-level warping guidance aimed at object region sampling to mitigate source scale bias in domain adaptation. Our approach improves adaptation across geographies, lighting and weather conditions, is agnostic to the task, domain adaptation algorithm, saliency guidance, and underlying model architecture. Highlights include +6.1 mAP50 for BDD100K Clear DENSE Foggy, +3.7 mAP50 for BDD100K Day Night, +3.0 mAP50 for BDD100K Clear Rainy, and +6.3 mIoU for Cityscapes ACDC. Our approach adds minimal memory during training and has no additional latency at inference time. Please see Appendix for more results and analysis.

One-Step Image Translation with Text-to-Image Models

Authors

Gaurav Parmar,Taesung Park,Srinivasa Narasimhan,Jun-Yan Zhu

Journal

arXiv preprint arXiv:2403.12036

Published Date

2024/3/18

In this work, we address two limitations of existing conditional diffusion models: their slow inference speed due to the iterative denoising process and their reliance on paired data for model fine-tuning. To tackle these issues, we introduce a general method for adapting a single-step diffusion model to new tasks and domains through adversarial learning objectives. Specifically, we consolidate various modules of the vanilla latent diffusion model into a single end-to-end generator network with small trainable weights, enhancing its ability to preserve the input image structure while reducing overfitting. We demonstrate that, for unpaired settings, our model CycleGAN-Turbo outperforms existing GAN-based and diffusion-based methods for various scene translation tasks, such as day-to-night conversion and adding/removing weather effects like fog, snow, and rain. We extend our method to paired settings, where our model pix2pix-Turbo is on par with recent works like Control-Net for Sketch2Photo and Edge2Image, but with a single-step inference. This work suggests that single-step diffusion models can serve as strong backbones for a range of GAN learning objectives. Our code and models are available at https://github.com/GaParmar/img2img-turbo.

Learned two-plane perspective prior based image resampling for efficient object detection

Authors

Anurag Ghosh,N Dinesh Reddy,Christoph Mertz,Srinivasa G Narasimhan

Published Date

2023

Real-time efficient perception is critical for autonomous navigation and city scale sensing. Orthogonal to architectural improvements, streaming perception approaches have exploited adaptive sampling improving real-time detection performance. In this work, we propose a learnable geometry-guided prior that incorporates rough geometry of the 3D scene (a ground plane and a plane above) to resample images for efficient object detection. This significantly improves small and far-away object detection performance while also being more efficient both in terms of latency and memory. For autonomous navigation, using the same detector and scale, our approach improves detection rate by+ 4.1 AP_S or+ 39% and in real-time performance by+ 5.3 sAP_S or+ 63% for small objects over state-of-the-art (SOTA). For fixed traffic cameras, our approach detects small objects at image scales other methods cannot. At the same scale, our approach improves detection of small objects by 195%(+ 12.5 AP_S) over naive-downsampling and 63%(+ 4.2 AP_S) over SOTA.

Systems and methods for diffraction line imaging

Published Date

2023/7/20

A novel class of imaging systems that combines diffractive optics with 1D line sensing is disclosed. When light passes through a diffraction grating or prism, it disperses as a function of wavelength. This property is exploited to recover 2D and 3D positions from line images. A detailed image formation model and a learning-based algorithm for 2D position estimation are disclosed. The disclosure includes several extensions of the imaging system to improve the accuracy of the 2D position estimates and to expand the effective field-of-view. The invention is useful for fast passive imaging of sparse light sources, such as streetlamps, headlights at night and LED-based motion capture, and structured light 3D scanning with line illumination and line sensing.

Active Velocity Estimation using Light Curtains via Self-Supervised Multi-Armed Bandits

Authors

Siddharth Ancha,Gaurav Pathak,Ji Zhang,Srinivasa Narasimhan,David Held

Journal

arXiv preprint arXiv:2302.12597

Published Date

2023/2/24

To navigate in an environment safely and autonomously, robots must accurately estimate where obstacles are and how they move. Instead of using expensive traditional 3D sensors, we explore the use of a much cheaper, faster, and higher resolution alternative: programmable light curtains. Light curtains are a controllable depth sensor that sense only along a surface that the user selects. We adapt a probabilistic method based on particle filters and occupancy grids to explicitly estimate the position and velocity of 3D points in the scene using partial measurements made by light curtains. The central challenge is to decide where to place the light curtain to accurately perform this task. We propose multiple curtain placement strategies guided by maximizing information gain and verifying predicted object locations. Then, we combine these strategies using an online learning framework. We propose a novel self-supervised reward function that evaluates the accuracy of current velocity estimates using future light curtain placements. We use a multi-armed bandit framework to intelligently switch between placement policies in real time, outperforming fixed policies. We develop a full-stack navigation system that uses position and velocity estimates from light curtains for downstream tasks such as localization, mapping, path-planning, and obstacle avoidance. This work paves the way for controllable light curtains to accurately, efficiently, and purposefully perceive and navigate complex and dynamic environments. Project website: https://siddancha.github.io/projects/active-velocity-estimation/

High resolution diffuse optical tomography using short range indirect imaging

Published Date

2023/2/16

the anatomical structures include a complex heterogeneous distribution of tissue, vasculature, tumors (benign or malignant) that vary in optical properties (density, scattering, absorption) and depths below the skin. This makes the modeling of light propagation below skin challenging.the photons can be assumed to be traveling diffusely in the medium and can be described as a random walk. This has enabled accurate forward models under diffuse photon propagation.

Virtual Home Staging: Inverse Rendering and Editing an Indoor Panorama under Natural Illumination

Authors

Guanzhou Ji,Azadeh O Sawyer,Srinivasa G Narasimhan

Published Date

2023/10/16

We propose a novel inverse rendering method that enables the transformation of existing indoor panoramas with new indoor furniture layouts under natural illumination. To achieve this, we captured indoor HDR panoramas along with real-time outdoor hemispherical HDR photographs. Indoor and outdoor HDR images were linearly calibrated with measured absolute luminance values for accurate scene relighting. Our method consists of three key components: (1) panoramic furniture detection and removal, (2) automatic floor layout design, and (3) global rendering incorporating scene geometry, new furniture objects, and real-time outdoor photograph. We demonstrate the effectiveness of our workflow in rendering indoor scenes under different outdoor illumination conditions. Additionally, we contribute a new calibrated HDR (Cali-HDR) dataset that consists of 137 calibrated indoor panoramas and their associated …

Analyzing physical impacts using transient surface wave imaging

Authors

Tianyuan Zhang,Mark Sheinin,Dorian Chan,Mark Rau,Matthew O’Toole,Srinivasa G Narasimhan

Published Date

2023

The subtle vibrations on an object's surface contain information about the object's physical properties and its interaction with the environment. Prior works imaged surface vibration to recover the object's material properties via modal analysis, which discards the transient vibrations propagating immediately after the object is disturbed. Conversely, prior works that captured transient vibrations focused on recovering localized signals (eg, recording nearby sound sources), neglecting the spatiotemporal relationship between vibrations at different object points. In this paper, we extract information from the transient surface vibrations simultaneously measured at a sparse set of object points using the dual-shutter camera described by Sheinin [31]. We model the geometry of an elastic wave generated shortly after an object's surface is disturbed (eg, a knock or a footstep), and use the model to localize the disturbance source for various materials (eg, wood, plastic, tile). We also show that transient object vibrations contain additional cues about the impact force and the impacting object's material properties. We demonstrate our approach in applications like localizing the strikes of a ping-pong ball on a table mid-play and recovering the footsteps' locations by imaging the floor vibrations they create.

Optimized virtual optical waveguides enhance light throughput in scattering media

Authors

Adithya Pediredla,Matteo Giuseppe Scopelliti,Srinivasa Narasimhan,Maysamreza Chamanzar,Ioannis Gkioulekas

Journal

Nature Communications

Published Date

2023/9/14

Ultrasonically-sculpted gradient-index optical waveguides enable non-invasive light confinement inside scattering media. The confinement level strongly depends on ultrasound parameters (e.g., amplitude, frequency), and medium optical properties (e.g., extinction coefficient). We develop a physically-accurate simulator, and use it to quantify these dependencies for a radially-symmetric virtual optical waveguide. Our analysis provides insights for optimizing virtual optical waveguides for given applications. We leverage these insights to configure virtual optical waveguides that improve light confinement fourfold compared to previous configurations at five mean free paths. We show that virtual optical waveguides enhance light throughput by 50% compared to an ideal external lens, in a medium with bladder-like optical properties at one transport mean free path. We corroborate these simulation findings with real …

Megahertz Light Steering Without Moving Parts

Authors

Adithya Pediredla,Srinivasa G Narasimhan,Maysamreza Chamanzar,Ioannis Gkioulekas

Published Date

2023

We introduce a light steering technology that operates at megahertz frequencies, has no moving parts, and costs less than a hundred dollars. Our technology can benefit many projector and imaging systems that critically rely on high-speed, reliable, low-cost, and wavelength-independent light steering, including laser scanning projectors, LiDAR sensors, and fluorescence microscopes. Our technology uses ultrasound waves to generate a spatiotemporally-varying refractive index field inside a compressible medium, such as water, turning the medium into a dynamic traveling lens. By controlling the electrical input of the ultrasound transducers that generate the waves, we can change the lens, and thus steer light, at the speed of sound (1.5 km/s in water). We build a physical prototype of this technology, use it to realize different scanning techniques at megahertz rates (three orders of magnitude faster than commercial alternatives such as galvo mirror scanners), and demonstrate proof-of-concept projector and LiDAR applications. To encourage further innovation towards this new technology, we derive the theory for its fundamental limits and develop a physically-accurate simulator for virtual design. Our technology offers a promising solution for achieving high-speed and low-cost light steering in a variety of applications.

Energy optimized imaging system with synchronized dynamic control of directable beam light source and reconfigurably masked photo-sensor

Published Date

2023/9/5

An energy optimized imaging system that includes a light source that has the ability to illuminate specific pixels in a scene, and a sensor that has the ability to capture light with specific pixels of its sensor matrix, temporally synchronized such that the sensor captures light only when the light source is illuminating pixels in the scene.

Agile depth sensing using triangulation light curtains

Published Date

2022/6/30

A method to dynamically and adaptively sample the depths of a scene using the principle of triangulation light curtains is described. The approach directly detects the presence or absence of obstacles (or scene points) at specified 3D lines in a scene by sampling the scene. The scene can be sampled sparsely, non-uniformly, or densely at specified regions. The depth sampling can be varied in real-time, enabling quick object discovery or detailed exploration of areas of interest. Once an object is discovered in the scene, adaptive light curtains comprising dense sampling of a region of the scene containing the object, can be used to better define the position, shape and size of the discovered object.

Reply to: The overwhelming role of ballistic photons in ultrasonically guided light through tissue

Authors

Maysamreza Chamanzar,Matteo Giuseppe Scopelliti,Adithya Pediredla,Hengji Huang,Srinivasa G Narasimhan,Ioannis Gkioulekas,Mohammad-Reza Alam,Michel M Maharbiz

Journal

Nature communications

Published Date

2022/4/6

In a previous publication 1, we introduced the concept of virtual optical waveguides capable of guiding and confining light without physically inserting optical components into the medium. We showed that ultrasound can locally change the refractive index in transparent and scattering media to sculpt in situ gradient-index (GRIN) optical waveguides. These waveguides can therefore be formed where external waveguides cannot be placed non-invasively. Follow-up work has also demonstrated the utility of this technique for confining light in tissue 2, 3. Edrei and Scarcelli 4 have tried to understand the underlying mechanisms of ultrasonic light guiding. Their letter acknowledges that virtual waveguides can guide scattered photons, and subsequently focuses on clarifying the relative effect of guiding ballistic versus scattered photons. Unfortunately, their letter draws incorrect and overly broad conclusions about the …

Programmable light curtains

Published Date

2022/11/8

Embodiments described herein are generally directed to a device that monitors for the presence of objects passing through or impinging on a virtual shell near the device, referred to herein as a “light curtain”, which is created by rapidly rotating a line sensor and a line laser in synchrony. The boundaries of the light curtain are defined by a sweeping line defined by the intersection of the sensing and illumination planes.

WALT: Watch And Learn 2D amodal representation from Time-lapse imagery

Authors

N Dinesh Reddy,Robert Tamburo,Srinivasa G Narasimhan

Published Date

2022

Current methods for object detection, segmentation, and tracking fail in the presence of severe occlusions in busy urban environments. Labeled real data of occlusions is scarce (even in large datasets) and synthetic data leaves a domain gap, making it hard to explicitly model and learn occlusions. In this work, we present the best of both the real and synthetic worlds for automatic occlusion supervision using a large readily available source of data: time-lapse imagery from stationary webcams observing street intersections over weeks, months, or even years. We introduce a new dataset, Watch and Learn Time-lapse (WALT), consisting of 12 (4K and 1080p) cameras capturing urban environments over a year. We exploit this real data in a novel way to automatically mine a large set of unoccluded objects and then composite them in the same views to generate occlusions. This longitudinal self-supervision is strong enough for an amodal network to learn object-occluder-occluded layer representations. We show how to speed up the discovery of unoccluded objects and relate the confidence in this discovery to the rate and accuracy of training occluded objects. After watching and automatically learning for several days, this approach shows significant performance improvement in detecting and segmenting occluded people and vehicles, over human-supervised amodal approaches.

See List of Professors in Srinivasa Narasimhan University(Carnegie Mellon University)

Srinivasa Narasimhan FAQs

What is Srinivasa Narasimhan's h-index at Carnegie Mellon University?

The h-index of Srinivasa Narasimhan has been 36 since 2020 and 52 in total.

What are Srinivasa Narasimhan's top articles?

The articles with the titles of

Toward Planet-Wide Traffic Camera Calibration

A single dose for me, A wealth of protection for us: The public health cost of individualism in the rollout of COVID-19 vaccine

TPSeNCE: Towards Artifact-Free Realistic Rain Generation for Deraining and Object Detection in Rain

Robot Safety Monitoring using Programmable Light Curtains

WALT3D: Generating Realistic Training Data from Time-Lapse Imagery for Reconstructing Dynamic Objects under Occlusion

Addressing Source Scale Bias via Image Warping for Domain Adaptation

One-Step Image Translation with Text-to-Image Models

Learned two-plane perspective prior based image resampling for efficient object detection

...

are the top articles of Srinivasa Narasimhan at Carnegie Mellon University.

What are Srinivasa Narasimhan's research interests?

The research interests of Srinivasa Narasimhan are: Computer vision, computer graphics, computational photography, sensors, computational imaging

What is Srinivasa Narasimhan's total number of citations?

Srinivasa Narasimhan has 16,906 citations in total.

What are the co-authors of Srinivasa Narasimhan?

The co-authors of Srinivasa Narasimhan are Takeo Kanade, Shree Nayar, Alexei A. Efros, Sanjiv Singh, Yoav Y. Schechner, Jean-François Lalonde.

    Co-Authors

    H-index: 169
    Takeo Kanade

    Takeo Kanade

    Carnegie Mellon University

    H-index: 134
    Shree Nayar

    Shree Nayar

    Columbia University in the City of New York

    H-index: 106
    Alexei A. Efros

    Alexei A. Efros

    University of California, Berkeley

    H-index: 71
    Sanjiv Singh

    Sanjiv Singh

    Carnegie Mellon University

    H-index: 49
    Yoav Y. Schechner

    Yoav Y. Schechner

    Technion - Israel Institute of Technology

    H-index: 41
    Jean-François Lalonde

    Jean-François Lalonde

    Université Laval

    academic-engine

    Useful Links