Learning DNN Abstractions using Gradient DescentRecorded Talk
Deep Neural Networks (DNNs) are being trained and trusted for performing fairly complex tasks, even in business- and safety-critical applications. This necessitates that they be formally analyzed before deployment. Scalability of such analyses is a major bottleneck in their widespread use. There has been a lot of work on abstraction, and counterexample-guided abstraction refinement (CEGAR) of DNNs to address the scalability issue. However, these abstraction-refinement techniques explore only a subset of possible abstractions, and may miss an \emph{optimal} abstraction. In particular, the refinement updates the abstract DNN based only on local information derived from the spurious counterexample in each iteration. The lack of a global view may result in a series of bad refinement choices, limiting the search to a region of sub-optimal abstractions. We propose a novel technique that parameterizes the construction of the abstract network in terms of continuous real-valued parameters. This allows us to use gradient descent to search through the space of possible abstractions, and ensures that the search never gets restricted to sub-optimal abstractions. Moreover, our parameterization can express more general abstractions than the existing techniques, enabling us to discover better abstractions than previously possible.
Wed 30 OctDisplayed time zone: Pacific Time (US & Canada) change
15:30 - 16:30 | |||
15:30 15mTalk | UFront: Toward A Unified MLIR Frontend for Deep Learning Research Papers Guoqing Bao Shanghai Enflame Technology Co., Ltd., Heng Shi Shanghai Jiao Tong University, Shanghai Enflame Technology Co., Ltd., Chengyi Cui Shanghai Enflame Technology Co., Ltd., Yalin Zhang Shanghai Enflame Technology Co., Ltd., Jianguo Yao Shanghai Jiao Tong University; Shanghai Enflame Technology | ||
15:45 15mTalk | FIPSER: Improving Fairness Testing of DNN by Seed Prioritization Research Papers Junwei Chen East China Normal University, Yueling Zhang East China Normal University, Lingfeng Zhang East China Normal University, Min Zhang East China Normal University, Chengcheng Wan East China Normal University, Ting Su East China Normal University, Geguang Pu East China Normal University, China | ||
16:00 15mTalk | Prioritizing Test Inputs for DNNs Using Training Dynamics Research Papers Jian Shen Nanjing University, Zhong Li , Minxue Pan Nanjing University, Xuandong Li Nanjing University | ||
16:15 15mTalk | Learning DNN Abstractions using Gradient DescentRecorded Talk NIER Track Diganta Mukhopadhyay TCS Research, Pune, India, Sanaa Siddiqui Indian Institute of Technology Delhi, New Delhi, India, Hrishikesh Karmarkar TCS Research, Kumar Madhukar Indian Institute of Technologiy Delhi, New Delhi, India, Guy Katz The Hebrew University of Jerusalem |