ResearchBib Share Your Research, Maximize Your Social Impacts
Sign for Notice Everyday Sign up >> Login

DISCML 2012 - 4th Workshop on Discrete Optimization in Machine Learning (DISCML): Structure and Scalability

Date2012-12-07

Deadline2012-09-16

VenueLake Tahoe, USA - United States USA - United States

Keywords

Websitehttp://www.discml.cc

Topics/Call fo Papers

4th Workshop on
Discrete Optimization in Machine Learning (DISCML):
Structure and Scalability
at the Annual Conference on Neural Information Processing Systems (NIPS 2012)
http://www.discml.cc
Submission Deadline: Sunday 16th September
===================================================================
- We apologize for multiple postings -
Optimization problems with ultimately discretely solutions are
becoming increasingly important in machine learning: At the core of
statistical machine learning is to infer conclusions from data, and
when the variables underlying the data are discrete, both the tasks of
inferring the model from data, as well as performing predictions using
the estimated model are discrete optimization problems. Two factors
complicate matters: first, many discrete problems are in the general
case very hard, and second, machine learning applications often demand
solving such problems at large scale. The focus of this year's
workshop lies on structures that enable scalability. Which properties
of the problem make it possible to still efficiently obtain exact or
decent approximate solutions? What are the challenges posed by
parallel and distributed processing? Which discrete problems in
machine learning are in need of more scalable algorithms? How can we
make discrete algorithms scalable? Some heuristics perform well but are
as yet devoid of a theoretical foundation. What explains this
behavior?
We would like to encourage high quality submissions of short papers
relevant to the workshop topics. Accepted papers will be presented as
spotlight talks and posters. Of particular interest are new algorithms
with theoretical guarantees, as well as applications of discrete
optimization to machine learning problems.
Areas of interest include
Optimization
? Combinatorial algorithms
? Submodular / supermodular optimization
? Discrete Convex Analysis
? Pseudo-boolean optimization
? Parallel & distributed discrete optimization
Continuous relaxations
? Sparse approximation & compressive sensing
? Regularization techniques
? Structured sparsity models
Learning in discrete domains
? Online learning / bandit optimization
? Generalization in discrete learning problems
? Adaptive / stochastic optimization
Applications
? Graphical model inference & structure learning
? Clustering
? Feature selection, active learning & experimental design
? Structured prediction
? Novel discrete optimization problems in ML, Computer Vision, NLP, ...
Submission deadline: September 16, 2012
Length & Format: max. 6 pages NIPS 2012 format
Time & Location: December 7 or 8 2012, Lake Tahoe, Nevada, USA
Submission instructions: Email submit-AT-discml.cc
Invited talks by
? Satoru Fujishige
? Amir Globerson
? Alex Smola
Organizers:
Andreas Krause (ETH Zurich, Switzerland),
Jeff A. Bilmes (University of Washington),
Pradeep Ravikumar (University of Texas, Austin),
Stefanie Jegelka (UC Berkeley)

Last modified: 2012-09-10 12:19:46