Arrow Research search

Author name cluster

Navyata Sanghvi

Possible papers associated with this exact author name in Arrow. This page groups case-insensitive exact name matches and is not a full identity disambiguation profile.

2 papers
1 author row

Possible papers

2

AAAI Conference 2021 Conference Paper

Inverse Reinforcement Learning with Explicit Policy Estimates

  • Navyata Sanghvi
  • Shinnosuke Usami
  • Mohit Sharma
  • Joachim Groeger
  • Kris Kitani

Various methods for solving the inverse reinforcement learning (IRL) problem have been developed independently in machine learning and economics. In particular, the method of Maximum Causal Entropy IRL is based on the perspective of entropy maximization, while related advances in the field of economics instead assume the existence of unobserved action shocks to explain expert behavior (Nested Fixed Point Algorithm, Conditional Choice Probability method, Nested Pseudo-Likelihood Algorithm). In this work, we make previously unknown connections between these related methods from both fields. We achieve this by showing that they all belong to a class of optimization problems, characterized by a common form of the objective, the associated policy and the objective gradient. We demonstrate key computational and algorithmic differences which arise between the methods due to an approximation of the optimal soft value function, and describe how this leads to more efficient algorithms. Using insights which emerge from our study of this class of optimization problems, we identify various problem scenarios and investigate each method’s suitability for these problems.

AAMAS Conference 2017 Conference Paper

Exploiting Robotic Swarm Characteristics for Adversarial Subversion in Coverage Tasks

  • Navyata Sanghvi
  • Sasanka Nagavalli
  • Katia Sycara

Multi-robot systems, such as swarms, with large number of members that are homogeneous and anonymous are robust to deletion and addition of members. However, these same properties that make the system robust, create vulnerabilities under certain circumstances. In this paper, we study such a case, namely the insertion by adversarial agents, called moles, that subvert the performance of the system. The adversary monitors the swarm’s movements during surveillance operations for the presence of holes, i. e. areas that were left uncovered by the swarm. The adversary then adds moles that get positioned in the swarm, in such a way as to deceive the swarms regarding the existence of holes and thus preventing the swarm from discovering and repairing the holes. This problem has significant military applications. Our contributions are as follows: First, to the best of our knowledge, this is the first paper that studies this problem. Second, we provide a formalization of the problem. Third, we provide several algorithms, and characterize them formally and also experimentally.