no code implementations • 27 Mar 2024 • Hazhar Rahmani, Abhishek N. Kulkarni, Jie Fu
In the second step, we prove that finding a most preferred policy is equivalent to computing a Pareto-optimal policy in a multi-objective MDP that is constructed from the original MDP, the preference automaton, and the chosen stochastic ordering relation.
no code implementations • 3 Apr 2023 • Chongyang Shi, Abhishek N. Kulkarni, Hazhar Rahmani, Jie Fu
Furthermore, if such a strategy does not exist, winning for P1 must entail the price of revealing his secret to the observer.
no code implementations • 4 Oct 2022 • Abhishek N. Kulkarni, Jie Fu
We construct a model called an improvement MDP, in which the synthesis of SPI and SASI strategies that guarantee at least one improvement reduces to computing positive and almost-sure winning strategies in an MDP.
no code implementations • 25 Sep 2022 • Hazhar Rahmani, Abhishek N. Kulkarni, Jie Fu
We prove that a weak-stochastic nondominated policy given the preference specification is Pareto-optimal in the constructed multi-objective MDP, and vice versa.
no code implementations • 7 Aug 2020 • Abhishek N. Kulkarni, Jie Fu
Given qualitative security specifications in formal logic, we show that the solution concepts from hypergames and reactive synthesis in formal methods can be extended to synthesize effective dynamic defense strategy using cyber deception.