“Kuldeep

Kuldeep S. Meel

Sung Kah Kay Assistant Professor
Affiliate, Institute of Data Science
Computer Science Department, School of Computing
National University of Singapore
Office: COM2-03-41
Phone (O): +65-651 61422
Email: meel@comp.nus.edu.sg
Please consult my calendar before suggesting a meeting time.
Research Group Website

Short Bio

Kuldeep Meel is Sung Kah Kay Assistant Professor in the Computer Science Department of School of Computing at National University of Singapore. He is an affiliate at Institute of Data Science and also holds a long-term visiting appointment at Computer Science Department of IIT Bombay.


His research interests are at the intersection of artificial intelligence and formal methods, which also lies at the intersection of theory and practice. His group strongly believes (and acts upon the beliefs) in developing mathematical frameworks whose faithful implementations can handle practical applications. He is also a receipient of 2019 NRF Fellowship for AI.

Open Positions

  • Multiple post-doc positions available in the project "Provably Verified and Explainable Probabilistic Reasoning". Read advertisement for more details.
  • One post-doc position (jointly advised with Prof. Roland Yap) available in the broad area of applying machine learning to SAT solvers, approximate counting techniques, and CP. Read advertisement for more details.
  • Research Assistant position available for inter-disciplinary project on computational Sustainability (jointed advised with Massimo Lupascu from Department of Geography). Read advertisement for more details.
  • We are always looking for highly motivated Ph.D. students, research assistants with time commitment of at least 6 months and summer internship for exceptional undergraduate interns in our group.
  • Read this before sending me an email.

[top]

Research Interests

Some of the recent papers are linked below. Full list of publications is available here

[top]

Students & Postdocs

Research Group Website
I am proud of my highly motivated students and postdocs:
  • Vignesh Sivaraman, Research Fellow (since Sep 2020)

  • Teodora Baluta, PhD Student, co-advised with Prateek Saxena
  • Bishwamittra Ghosh, PhD Student
  • Priyanka Golia, Joint NUS-IITK PhD Student, co-advised with Subhajit Roy (IITK)
  • Yash Pote, PhD Student
  • Yang Suwei, PhD Student

  • Zhanzhong Pang, Research Intern (since Feb 2020)
  • Jiong Yang, Resesarch Intern (since May 2020)
  • Biswadeep Sen, Research Intern (since Sep 2020)

Alumni

  • Dr. Mate Soos, Senior Research Fellow (Jul -- Sep 2019; Mar -- June 2018)
  • Dr. Yong Lai, Research Fellow, Research Fellow (Aug 2019-Aug 2020), now: Associate Professor at Jilin University

  • Alexis de Colnet, MComp@NUS (Graduated: Dec 2018); now a PhD student at Centre de Recherche en Informatique de Lens
  • Rahul Gupta, BTech+MTech@IITK (Graduated: June 2019) , co-advised with Subhajit Roy (IITK)
  • Shubham Sharma, BTech+MTech@IITK (Graduated: May 2019), co-advised with Subhajit Roy (IITK)
  • Delannoy Remi Christian, MComp@NUS, (Graduated: Dec 2019)
  • Lorenzo Ciampiconi, MS@Politecnico di Milano (Graduated: Dec 2019)

  • Bhavishya, Summer Intern (May 2018 - July 2018)
  • Mohimenul Kabir, Research Intern (September 2018 - July 2018)
  • Do Andre Khoi Nguyen, Undergraduate Researcher (May 2018 -- Apr 2019)
  • Himanshu Arora, Research Intern (Feb 2019 -- July 2019)
  • Durgesh Agrawal, Summer Intern (May 2019 -- July 2019)
  • Kurt Warren, Research Intern (July 2019 - Sep 2019)

[top]

Teaching

  • CS3243: Introduction to Artificial Intelligence (Fall 2020)
  • CS 4244: Knowledge Representation and Reasoning (Spring 2020; Spring 2019)
  • CS 4269/5469: Fundamentals of Logic in Computer Science (Fall 2019)
  • CS 6283: Advanced Topics in Computer Science: Logic in AI (Fall 2018)
  • CS4244: Knowledge-Based Systems (Spring 2018), (co-taught with Dr. Henry Chia)

[top]

Awards and Honors

[top]

Selected Recent Publications

Google Scholar

1
Sparse Hashing for Scalable Approximate Model Counting: Theory and Practice
Kuldeep S. Meel ⓡ S. Akshay
Proceedings of Logic in Computer science (LICS), 2020.
Given a CNF formula F on n variables, the problem of model counting, also referred to as #SAT, is to compute the number of models or satisfying assignments of F. Model counting is a fundamental but hard problem in computer science with varied applications. Recent years have witnessed a surge of effort towards developing efficient algorithmic techniques that combine the classical 2-universal hashing (from [Stockmeyer 1983]) with the remarkable progress in SAT solving over the past decade. These techniques augment the CNF formula F with random XOR constraints and invoke an NP oracle repeatedly on the resultant CNF-XOR formulas. In practice, the NP oracle calls are replaced by calls to a SAT solver and it is observed that runtime performance of modern SAT solvers (based on conflict-driven clause learning) on CNF-XOR formulas is adversely affected by the size of XOR constraints. However, the standard construction of 2-universal hash functions chooses every variable with probability p =1/2 leading to XOR constraints of size n/2 in expectation. Consequently, the main challenge is to design sparse hash functions, where variables can be chosen with smaller probability and lead to smaller sized XOR constraints, which can then replace 2-universal hash functions. In this paper, our goal is to address this challenge both from a theoretical and a practical perspective. First, we formalize a relaxation of universal hashing, called concentrated hashing, a notion implicit in prior works to design sparse hash functions. We then establish a novel and beautiful connection between concentration measures of these hash functions and isoperimetric inequalities on boolean hypercubes. This allows us to obtain tight bounds on variance as well as the dispersion index and show that p = O(log m /m ) suffices for the design of sparse hash functions from 2^n to 2^m belonging to concentrated hash family. Finally, we use sparse hash functions belonging to this concentrated hash family to develop new approximate counting algorithms. A comprehensive experimental evaluation of our algorithm on 1896 benchmarks with computational effort of over 20,000 computational hours demonstrates significant speedup compared to existing approaches. To the best of our knowledge, this work is the first study to demonstrate runtime improvement of approximate model counting algorithms through the usage of sparse hash functions, while still retaining strong theoretical guarantees (a la 2-universal hash functions).
2
BIRD: Engineering an Efficient CNF-XOR SAT Solver and its Applications to Approximate Model Counting
Mate Soos and Kuldeep S. Meel
Proceedings of AAAI Conference on Artificial Intelligence (AAAI), 2019.
Given a Boolean formula F, the problem of model counting, also referred to as #SAT is to compute the number of solutions of F. Model counting is a fundamental problem in artificial intelligence with a wide range of applications including probabilistic reasoning, decision making under uncertainty, quantified information flow, and the like. Motivated by the success of SAT solvers, there has been surge of interest in the design of hashing-based techniques for approximate model counting for the past decade. We profiled the state of the art approximate model counter ApproxMC3 and observed that over 99.99% of time is consumed by the underlying SAT solver, CryptoMinisat. This observation motivated us to ask: Can we design an efficient underlying CNF-XOR SAT solver that can take advantage of the structure of hashing-based algorithms and would this lead to an efficient approximate model counter? The primary contribution of this paper is an affirmative answer to the above question. We present a novel architecture, called BIRD, to handle CNF-XOR formulas arising from hashing-based techniques. The resulting hashing-based approximate model counter, called ApproxMC3, employs the BIRD framework in its underlying SAT solver, CryptoMinisat. To the best of our knowledge, we conducted the most comprehensive study of evaluation performance of counting algorithms involving 1896 benchmarks with computational effort totaling 86400 computational hours. Our experimental evaluation demonstrates significant runtime performance improvement for ApproxMC3 over ApproMC2. In particular, we solve 648 benchmarks more than ApproMC2, the state of the art approximate model counter and for all the formulas where both ApproMC2 and ApproxMC3 did not timeout and took more than 1 seconds, the mean speedup is 284.40 -- more than two orders of magnitude.
3
Quantitative Verification of Neural Networks And its Security Applications
Teodora Baluta, Shiqi Shen, Shweta Shine, Kuldeep S. Meel, and Prateek Saxena
ACM Conference on Computer and Communications Security (CCS), 2019.
Neural networks are increasingly employed in safety-critical domains. This has prompted interest in verifying or certifying logically encoded properties of neural networks. Prior work has largely focused on checking existential properties, wherein the goal is to check whether there exists any input that violates a given property of interest. However, neural network training is a stochastic process, and many questions arising in their analysis require probabilistic and quantitative reasoning, i.e., estimating how many inputs satisfy a given property. To this end, our paper proposes a novel and principled framework to quantitative verification of logical properties specified over neural networks. Our framework is the first to provide PAC-style soundness guarantees, in that its quantitative estimates are within a controllable and bounded error from the true count. We instantiate our algorithmic framework by building a prototype tool called NPAQ that enables checking rich properties over binarized neural networks. We show how emerging security analyses can utilize our framework in 3 concrete point applications: quantifying robustness to adversarial inputs, efficacy of trojan attacks, and fairness/bias of given neural networks.
4
On testing of Uniform Samplers
Sourav Chakraborty and Kuldeep S. Meel
Proceedings of AAAI Conference on Artificial Intelligence (AAAI), 2019.
Recent years have seen an unprecedented adoption of artificial intelligence in wide variety of applications ranging from medical diagnosis, automobile, security to aircraft collision avoidance. Probabilistic reasoning is a key component of such modern artificial intelligence systems. Sampling techniques form the core of the state of the art probabilistic reasoning systems. In contrast to testing for deterministic programs, where one trace is sufficient to prove the existence of a bug; such is not the case for samplers as one sample is typically not sufficient to prove non-conformity of the sampler to the desired distribution. This makes one wonder: whether it is possible to design testing methodology to test whether a sampler under test generates samples close to a given distribution. The primary contribution of this paper is a positive answer to the above question: We design, to the best of our knowledge, the first algorithmic framework, Barbarik, to test whether the distribution generated by $\varepsilon-$close or $\eta-$far from the uniform distribution. In contrast to the sampling techniques that require an exponential or sub-exponential number of samples for sampler whose support can be represented by $n$ bits, Barbarik requires only $\mathcal{O}(1/(\eta-\varepsilon)^2)$ samples. We present a prototype implementation of Barbarik and use it to test three state of the art uniform samplers over the support defined by combinatorial constraints. Barbarik is able to provide a certificate of uniformity to one sampler and demonstrate non-uniformity for the other two samplers.
5
On Computing Minimal Independent Support and Its Applications to Sampling and Counting
Alexander Ivrii, Sharad Malik, Kuldeep S. Meel, and Moshe Y. Vardi
Proceedings of International Conference on Constraint Programming (CP), 2015.
Best Student Paper Award
Constrained sampling and counting are two fundamental problems arising in domains ranging from artificial intelligence and security, to hardware and software testing. Recent approaches to approximate solutions for these problems rely on employing SAT solvers and universal hash functions that are typically encoded as XOR constraints of length n/2 for an input formula with n variables. As the runtime performance of SAT solvers heavily depends on the length of XOR constraints, recent research effort has been focused on reduction of length of XOR constraints. Consequently, a notion of Independent Support was proposed, and it was shown that constructing XORs over independent support (if known) can lead to a significant reduction in the length of XOR constraints without losing the theoretical guarantees of sampling and counting algorithms. In this paper, we present the first algorithmic procedure (and a corresponding tool, called MIS) to determine minimal independent support for a given CNF formula by employing a reduction to group minimal unsatisfiable subsets (GMUS). By utilizing minimal independent supports computed by MIS, we provide new tighter bounds on the length of XOR constraints for constrained counting and sampling. Furthermore, the universal hash functions constructed from independent supports computed by MIS provide two to three orders of magnitude performance improvement in state-of-the-art constrained sampling and counting tools, while still retaining theoretical guarantees.

Full list of publications is available here

Please visit Software and Github Repository for the accompanied softwares. Our group strongly believes in developing algorithms whose faithful implementations (i.e., no hiding of constant factors, magic parameters) can solve real-world problems.

[top]

Service

[top]

News

  • 25 September 2020

    Three Papers accepted to NeurIPS 2020.

    1. The first paper titled, On Testing of Samplers, describes how to test samplers without exponentially many samles Authors: Kuldeep S Meel, Yash Pralhad Pote and Sourav Chakraborty
    2. The second paper focuses on Taming Discrete Integration via the Boon of Dimensionality. Authors: Jeffrey M. Dudek, Dror Fried and Kuldeep S. Meel
    3. The third paper focuses on Efficient Distance Approximation for Structured High-Dimensional Distributions via Learning. Authors: Arnab Bhattacharyya, Sutanu Gayen, Kuldeep S. Meel and N. V. Vinodchandran
  • 18 September 2020

    Our Paper on Model Counting meets F0 Estimation (public version forthcoming!) is accepted to PODS 2021. Authors: Arnab Bhattacharyya, Kuldeep Meel, A. Pavan and N.V. Vinodchandran

  • 10 September 2020

    We presented our work on Phase Transition Behaviour in Knowledge Compilation at CP 2020.

  • 20 July 2020

    We will be presenting three papers at CAV 2020.

    1. The first paper builds on our CNF-XOR solving paradigm (BIRD) and as a result, the new versions of ApproxMC and UniGen are faster than ever.
    2. The second paper proposes the first algorithm for approximate MUS counting.
    3. The third one proposes a data-driven approach for Boolean functional synthesis, which works at the intersection of constrained sampling, machine learning and automated reasoning.

All news…

Recent: MeelGroup Wins Model Counting Competition
Profile: Kuldeep Meel: The New CS Professor
Old Profile: Kuldeep Meel: High-Risk, High-Reward Research

[top]

Recruitment

We are always looking for highly motivated Ph.D. students, research assistants and summer internship for exceptional undergraduate interns in our group. We work at the intersection of algorithmic design and systems; therefore, an ideal candidate should have deeper expertise in one area and willingness to learn the other. A strong background in statistics, algorithms/formal methods and prior experience in coding is crucial to make a significant contribution to our research. Read here for more details.

[top]