Monotone Submodular Function
Monotone submodular functions, characterized by diminishing returns and increasing values with set size, are central to many optimization problems in machine learning and related fields. Current research focuses on developing efficient algorithms for maximizing these functions under various constraints (e.g., cardinality, matroid, fairness), including dynamic settings where the input set changes over time, and extending these methods to non-monotone and stochastic scenarios. These advancements leverage techniques such as greedy algorithms, gradient ascent, and neural network architectures like FLEXSUBNET, aiming for improved approximation guarantees and scalability. The resulting algorithms have significant implications for applications like data summarization, resource allocation, and recommendation systems.
Papers
A Dynamic Algorithm for Weighted Submodular Cover Problem
Kiarash Banihashem, Samira Goudarzi, MohammadTaghi Hajiaghayi, Peyman Jabbarzade, Morteza Monemizadeh
Sliding Window Bi-Objective Evolutionary Algorithms for Optimizing Chance-Constrained Monotone Submodular Functions
Xiankun Yan, Aneta Neumann, Frank Neumann