P. Alquier, V. Cottet, and G. Lecué, Estimation bounds and sharp oracle inequalities of regularized procedures with lipschitz loss functions, 2017.

P. Alquier, J. Ridgway, and N. Chopin, On the properties of variational approximations of Gibbs posteriors, Journal of Machine Learning Research, vol.17, issue.239, pp.1-41, 2016.

J. Audibert and A. B. Tsybakov, Fast learning rates for plug-in classifiers, The Annals of Statistics, vol.35, issue.2, pp.608-633, 2007.
DOI : 10.1214/009053606000001217

URL : https://hal.archives-ouvertes.fr/hal-00160849

R. Francis and . Bach, Consistency of trace norm minimization, Journal of Machine Learning Research, vol.9, pp.1019-1048, 2008.

F. Barthe, O. Guédon, S. Mendelson, and A. Naor, A probabilistic approach to the geometry of the l n p -ball. The Annals of Probability, pp.480-513, 2005.

S. Barthelmé and N. Chopin, Expectation Propagation for Likelihood-Free Inference, Journal of the American Statistical Association, vol.21, issue.505, pp.315-333, 2014.
DOI : 10.3758/BF03194544

L. Peter, S. Bartlett, and . Mendelson, Empirical minimization. Probability Theory and Related Fields, pp.311-334, 2006.

L. Peter, . Bartlett, I. Michael, J. D. Jordan, and . Mcauliffe, Large margin classifiers: Convex loss, low noise, and convergence rates, NIPS, pp.1173-1180, 2003.

L. Peter, O. Bartlett, S. Bousquet, and . Mendelson, Local Rademacher complexities, Ann. Statist, vol.33, issue.4, pp.1497-1537, 2005.

L. Peter, . Bartlett, I. Michael, J. D. Jordan, and . Mcauliffe, Convexity, classification, and risk bounds, Journal of the American Statistical Association, vol.101, issue.473, pp.138-156, 2006.

A. Mark, J. Beaumont, J. Cornuet, C. P. Marin, and . Robert, Adaptive approximate Bayesian computation, Biometrika, vol.96, issue.4, pp.983-990, 2009.

A. Belloni and V. Chernozhukov, -1-penalized quantile regression in high-dimensional sparse models. The Annals of Statistics, pp.82-130, 2011.

C. M. Bishop, Pattern Recognition and Machine Learning (Information Science and Statistics), 2006.

M. Bogdan, E. Van-den-berg, C. Sabatti, W. Su, and E. J. Candès, Slope?adaptive variable selection via convex optimization. The annals of applied statistics, p.1103, 2015.
DOI : 10.1214/15-aoas842

URL : https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4689150/pdf

S. Boucheron, G. Lugosi, and P. Massart, Concentration inequalities: A nonasymptotic theory of independence
DOI : 10.1093/acprof:oso/9780199535255.001.0001

URL : https://hal.archives-ouvertes.fr/hal-00794821

S. Boucheron, O. Bousquet, and G. Lugosi, Theory of Classification: a Survey of Some Recent Advances, ESAIM: Probability and Statistics, vol.49, pp.323-375, 2005.
DOI : 10.1109/TIT.2003.813564

URL : https://hal.archives-ouvertes.fr/hal-00017923

S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers, Machine Learning, pp.1-122, 2011.
DOI : 10.1561/2200000016

S. Burer, D. Renato, and . Monteiro, A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization, Mathematical Programming, pp.329-357, 2003.
DOI : 10.1007/s10107-002-0352-8

J. Cai, E. J. Candès, and Z. Shen, A Singular Value Thresholding Algorithm for Matrix Completion, SIAM Journal on Optimization, vol.20, issue.4, pp.1956-1982, 2010.
DOI : 10.1137/080738970

T. Cai and W. Zhou, A max-norm constrained minimization approach to 1-bit matrix completion, Journal of Machine Learning Research, vol.14, pp.3619-3647, 2013.

J. Emmanuel, Y. Candès, and . Plan, Matrix completion with noise, Proceedings of the IEEE, pp.925-936, 2010.

J. Emmanuel, B. Candès, and . Recht, Exact matrix completion via convex optimization, Communications of the ACM, vol.55, issue.6, pp.111-119, 2012.

J. Emmanuel, T. Candès, and . Tao, The power of convex relaxation: Near-optimal matrix completion, IEEE Transactions on Information Theory, vol.56, issue.5, pp.2053-2080, 2010.

O. Catoni, Statistical Learning Theory and Stochastic Optimization. Saint-Flour Summer School on Probability Theory, Lecture Notes in Mathematics, 2001.
DOI : 10.1007/b99352

URL : https://hal.archives-ouvertes.fr/hal-00104952

O. Catoni, PAC-Bayesian supervised classification: the thermodynamics of statistical learning, Institute of Mathematical Statistics Lecture Notes?Monograph Series, vol.56, 2007.
URL : https://hal.archives-ouvertes.fr/hal-00206119

D. Chafaï, O. Guédon, G. Lecué, and A. Pajor, Interactions between compressed sensing random matrices and high dimensional geometry, of Panoramas et Synthèses [Panoramas and Syntheses]. Société Mathématique de France

V. Chandrasekaran, B. Recht, P. A. Parrilo, and A. S. Willsky, The Convex Geometry of Linear Inverse Problems, Foundations of Computational Mathematics, vol.1, issue.10, pp.805-849, 2012.
DOI : 10.1007/978-1-4613-8431-1

S. Chatterjee, Matrix estimation by universal singular value thresholding. The Annals of Statistics, pp.177-214
DOI : 10.1214/14-aos1272

URL : http://arxiv.org/pdf/1212.1247.pdf

V. Cottet and P. Alquier, 1-bit Matrix Completion: PAC-Bayesian Analysis of a Variational Approximation ArXiv e-prints, 2016.

B. Cseke and T. Heskes, Approximate marginals in latent Gaussian models, J. Mach. Learn. Res, vol.12, pp.417-454, 2011.

F. Cucker and S. Smale, On the mathematical foundations of learning, Bulletin of the American Mathematical Society, vol.39, issue.01, pp.1-49, 2002.
DOI : 10.1090/S0273-0979-01-00923-5

A. Dalalyan and A. B. Tsybakov, Aggregation by exponential weighting, sharp PAC-Bayesian bounds and sparsity, Machine Learning, pp.39-61, 2008.
DOI : 10.1007/978-3-540-45167-9_23

URL : https://hal.archives-ouvertes.fr/hal-00291504

A. Mark, Y. Davenport, and . Plan, Ewout van den Berg, and Mary Wootters . 1-bit matrix completion, Information and Inference, vol.3, issue.3, pp.189-223, 2014.

T. A. Dean, S. S. Singh, A. Jasra, and G. W. Peters, Parameter estimation for hidden Markov models with intractable likelihoods. Scand, J. Stat, vol.41, issue.4, pp.970-987, 2014.
DOI : 10.1111/sjos.12077

G. Dehaene and S. Barthelmé, Expectation propagation in the large-data limit. arXiv preprint, 2015.
URL : https://hal.archives-ouvertes.fr/hal-01235066

P. Guillaume, S. Dehaene, and . Barthelmé, Bounding errors of expectation-propagation, Advances in Neural Information Processing Systems, pp.244-252, 2015.

P. Del-moral, A. Doucet, and A. Jasra, An adaptive sequential Monte Carlo method for approximate Bayesian computation, Statistics and Computing, vol.6, issue.5, pp.1009-1020, 2012.
DOI : 10.1098/rsif.2008.0172

R. M. Dudley, Real analysis and probability, volume 74 of Cambridge Studies in Advanced Mathematics, 2002.

L. Dümbgen, Bounding standard gaussian tail probabilities, 2010.

A. Elsener and S. Van-de-geer, Robust low-rank matrix estimation . arXiv preprint, 2016.

J. Robert, R. L. Erhardt, and . Smith, Approximate bayesian computing for spatial extremes, Computational Statistics & Data Analysis, vol.56, issue.6, pp.1468-1481, 2012.

M. Fazel, H. Hindi, P. Stephen, and . Boyd, A rank minimization heuristic with application to minimum order system approximation, Proceedings of the 2001 American Control Conference. (Cat. No.01CH37148), pp.4734-4739, 2001.
DOI : 10.1109/ACC.2001.945730

M. Garcia-magariños, A. Antoniadis, R. Cao, and W. González-manteiga, Lasso Logistic Regression, GSoft and the Cyclic Coordinate Descent Algorithm: Application to Gene Expression Data, Statistical Applications in Genetics and Molecular Biology, vol.9, issue.1, 2010.
DOI : 10.2202/1544-6115.1536

A. Gelman, J. B. Carlin, H. S. Stern, D. B. Dunson, A. Vehtari et al., Bayesian data analysis. Texts in Statistical Science Series, 2014.

U. Michael, J. Gutmann, and . Corander, Bayesian optimization for likelihood-free inference of simulator-based statistical models, 2015.

N. Halko, . Per-gunnar-martinsson, and . Tropp, Finding Structure with Randomness: Probabilistic Algorithms for Constructing Approximate Matrix Decompositions, SIAM Review, vol.53, issue.2, pp.217-288, 2011.
DOI : 10.1137/090771806

N. Heess, D. Tarlow, and J. Winn, Learning to pass expectation propagation messages, Advances in Neural Information Processing Systems 26, pp.3219-3227, 2013.

R. Herbrich and T. Graepel, A PAC-Bayesian margin bound for linear classifiers, IEEE Transactions on Information Theory, vol.48, issue.12, pp.3140-3150, 2002.
DOI : 10.1109/TIT.2002.805090

URL : http://www.icos.ethz.ch/research/ieee_final_submission.pdf

M. Herbster, S. Pasteris, and M. Pontil, Mistake bounds for binary matrix completion, Advances In Neural Information Processing Systems, pp.3954-3962, 2016.

C. Hsieh, A. Peder, and . Olsen, Nuclear norm minimization via active subspace selection, ICML, pp.575-583, 2014.

C. Hsieh, N. Natarajan, S. Inderjit, and . Dhillon, PU learning for matrix completion, Proceedings of The 32nd International Conference on Machine Learning, pp.2445-2453, 2015.

J. Peter and . Huber, Robust estimation of a location parameter, Ann. Math. Statist, vol.35, pp.73-101, 1964.

T. S. Jaakkola and M. I. Jordan, Bayesian parameter estimation via variational methods, Statistics and Computing, vol.10, issue.1, pp.25-37, 2000.
DOI : 10.1023/A:1008932416310

A. Jasra, Approximate Bayesian Computation for a Class of Time Series Models, International Statistical Review, vol.12, issue.37, 2015.
DOI : 10.1080/10618600.2014.938811

H. Raghunandan, A. Keshavan, S. Montanari, and . Oh, Matrix completion from noisy entries, Journal of Machine Learning Research, vol.11, pp.2057-2078, 2010.

O. Klopp, Noisy low-rank matrix completion with general sampling distribution, Bernoulli, vol.20, issue.1, pp.282-303, 2014.
DOI : 10.3150/12-BEJ486

URL : https://hal.archives-ouvertes.fr/hal-00675413

O. Klopp, Matrix completion by singular value thresholding: Sharp bounds, Electronic Journal of Statistics, vol.9, issue.2, pp.2348-2369, 2015.
DOI : 10.1214/15-EJS1076

URL : https://hal.archives-ouvertes.fr/hal-01111757

O. Klopp, J. Lafond, É. Moulines, and J. Salmon, Adaptive multinomial matrix completion, Electronic Journal of Statistics, vol.9, issue.2, pp.2950-2975, 2015.
DOI : 10.1214/15-EJS1093

URL : https://hal.archives-ouvertes.fr/hal-01058447

V. Koltchinskii and D. Panchenko, Empirical margin distributions and bounding the generalization error of combined classifiers, Ann. Statist, vol.30, issue.1, pp.1-50, 2002.

V. Koltchinskii, Local Rademacher complexities and oracle inequalities in risk minimization, The Annals of Statistics, vol.34, issue.6, pp.2593-2656, 2006.
DOI : 10.1214/009053606000001019

URL : http://doi.org/10.1214/009053606000001019

V. Koltchinskii, Oracle inequalities in empirical risk minimization and sparse recovery problems ISBN 978-3-642-22146-0. Lectures from the 38th Probability Summer School held in Saint-Flour, Lecture Notes in Mathematics, vol.2033, 2008.
DOI : 10.1007/978-3-642-22147-7

V. Koltchinskii, K. Lounici, and . Tsybakov, Nuclear-norm penalization and optimal rates for noisy low-rank matrix completion. The Annals of Statistics, pp.2302-2329, 2011.
DOI : 10.1214/11-aos894

URL : https://hal.archives-ouvertes.fr/hal-00676868

M. Kyung, J. Gill, M. Ghosh, and G. Casella, Penalized regression, standard errors, and Bayesian lassos, Bayesian Analysis, vol.5, issue.2, pp.369-412, 2010.
DOI : 10.1214/10-BA607

URL : http://doi.org/10.1214/10-ba607

J. Lafond, O. Klopp, E. Moulines, and J. Salmon, Probabilistic low-rank matrix completion on finite alphabets, Advances in Neural Information Processing Systems, pp.1727-1735, 2014.
URL : https://hal.archives-ouvertes.fr/hal-01081805

P. Latouche, S. Robin, and S. Ouadah, Goodness of fit of logistic regression models for random graphs, Journal of Computational and Graphical Statistics, vol.41, issue.12, 2015.
DOI : 10.1016/j.patrec.2010.01.026

URL : https://hal.archives-ouvertes.fr/hal-01583801

G. Lecué, Optimal rates of aggregation in classification under low noise assumption, Bernoulli, vol.13, issue.4, pp.1000-1022, 2007.
DOI : 10.3150/07-BEJ6044

G. Lecué, Interplay between concentration, complexity and geometry in learning theory with applications to high dimensional data analysis. Habilitation à Diriger des Recherches Université, 2011.

G. Lecué and S. Mendelson, General nonexact oracle inequalities for classes with a subexponential envelope. The Annals of Statistics, pp.832-860, 2012.

G. Lecué and S. Mendelson, Learning subgaussian classes: Upper and minimax bounds, 2013.

G. Lecué and S. Mendelson, Regularization and the smallball method II: complexity dependent error rates, 2015.

G. Lecué and S. Mendelson, Regularization and the smallball method I: sparse recovery, 2015.

M. Ledoux and M. Talagrand, Probability in Banach spaces, Ergebnisse der Mathematik und ihrer Grenzgebiete (3) [Results in Mathematics and Related Areas, 1991.
DOI : 10.1007/978-3-642-20212-4

J. Yew, Y. W. Lim, and . Teh, Variational Bayesian approach to movie rating prediction, Proceedings of KDD cup and workshop, pp.15-21, 2007.

T. Tien, M. , and P. Alquier, A Bayesian approach for noisy matrix completion: Optimal rate under general sampling distribution, Electronic Journal of Statistics, vol.9, pp.823-841, 2015.

C. Mak, Polychotomous logistic regression via the Lasso, ProQuest LLC, 1999.

E. Mammen and A. Tsybakov, Smooth discrimination analysis. The Annals of Statistics, pp.1808-1829, 1999.

P. Marjoram, J. Molitor, V. Plagnol, and S. Tavaré, Markov chain Monte Carlo without likelihoods, Proceedings of the National Academy of Sciences, pp.15324-15328, 2003.
DOI : 10.1214/aoms/1177700372

URL : http://www.pnas.org/content/100/26/15324.full.pdf

R. Mazumder, T. Hastie, and R. Tibshirani, Spectral regularization algorithms for learning large incomplete matrices, Journal of machine learning research, vol.11, pp.2287-2322, 2010.

D. A. Mcallester, Some PAC-Bayesian theorems, Proceedings of the eleventh annual conference on Computational learning theory , COLT' 98, pp.230-234, 1998.
DOI : 10.1145/279943.279989

L. Meier, S. Van-de-geer, and P. Bühlmann, The group lasso for logistic regression, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.68, issue.1, pp.53-71, 2008.
DOI : 10.1093/oxfordjournals.pan.a004868

URL : http://onlinelibrary.wiley.com/doi/10.1111/j.1467-9868.2007.00627.x/pdf

S. Mendelson, Improving the sample complexity using global data, IEEE Transactions on Information Theory, vol.48, issue.7, pp.1977-1991, 2002.
DOI : 10.1109/TIT.2002.1013137

URL : http://axiom.anu.edu.au/~shahar/Mendelson.ps

S. Mendelson, On the performance of kernel classes, J. Mach. Learn. Res, vol.4, issue.5, pp.759-771, 2004.

S. Mendelson, Obtaining fast error rates in nonconvex situations, Journal of Complexity, vol.24, issue.3, pp.380-397, 2008.
DOI : 10.1016/j.jco.2007.09.001

URL : https://doi.org/10.1016/j.jco.2007.09.001

S. Mendelson and J. Neeman, Regularization in kernel learning, The Annals of Statistics, vol.38, issue.1, pp.526-565, 2010.
DOI : 10.1214/09-AOS728

URL : http://doi.org/10.1214/09-aos728

T. P. Minka, Expectation propagation for approximate Bayesian inference, Proceedings of Uncertainty in Artificial Intelligence, vol.17, pp.362-369, 2001.

S. Negahban, J. Martin, and . Wainwright, Restricted strong convexity and weighted matrix completion: Optimal bounds with noise, Journal of Machine Learning Research, vol.13, pp.1665-1697, 2012.

S. A. Padoan, M. Ribatet, and S. A. Sisson, Likelihood-Based Inference for Max-Stable Processes, Journal of the American Statistical Association, vol.105, issue.489, pp.263-277, 2010.
DOI : 10.1198/jasa.2009.tm08577

URL : https://hal.archives-ouvertes.fr/hal-00361245

T. Park and G. Casella, The Bayesian Lasso, Journal of the American Statistical Association, vol.103, issue.482, pp.681-686, 2008.
DOI : 10.1198/016214508000000337

J. Dante, . Paz, G. Ariel, and . Sánchez, Improving the precision matrix for precision cosmology, Monthly Notices of the Royal Astronomical Society, vol.454, issue.4, pp.4326-4334, 2015.

M. M. Rao and Z. D. Ren, Theory of Orlicz spaces, volume 146 of Monographs and Textbooks in Pure and Applied Mathematics, 1991.

M. M. Rao and Z. D. Ren, Applications of Orlicz spaces, volume 250 of Monographs and Textbooks in Pure and Applied Mathematics, 2002.

B. Recht and C. Ré, Parallel stochastic gradient algorithms for large-scale matrix completion, Mathematical Programming Computation, vol.8, issue.2, pp.201-226, 2013.
DOI : 10.1137/S1052623495294797

URL : http://pages.cs.wisc.edu/~brecht/papers/11.Rec.Re.IPGM.pdf

B. Recht, M. Fazel, and P. A. Parrilo, Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization, SIAM Review, vol.52, issue.3, pp.471-501, 2010.
DOI : 10.1137/070697835

P. Christian, G. Robert, and . Casella, Monte Carlo Statistical Methods (Springer Texts in Statistics), 2005.

A. Rohde, B. Alexandre, and . Tsybakov, Estimation of highdimensional low-rank matrices. The Annals of Statistics, pp.887-930, 2011.
DOI : 10.1214/10-aos860

URL : https://hal.archives-ouvertes.fr/hal-00606063

E. Ruli, N. Sartori, and L. Ventura, Approximate Bayesian computation with composite score functions, Statistics and Computing, vol.21, issue.3, pp.1-14, 2015.
DOI : 10.1016/j.jspi.2008.08.003

URL : http://arxiv.org/pdf/1311.7286

N. Sabbe, O. Thas, and J. Ottoy, EMLasso: logistic lasso with missing data, Statistics in Medicine, vol.26, issue.6, pp.3143-3157, 2013.
DOI : 10.1093/bioinformatics/btp041

R. Salakhutdinov and A. Mnih, Bayesian probabilistic matrix factorization using Markov chain Monte Carlo, Proceedings of the 25th international conference on Machine learning, ICML '08, pp.880-887, 2008.
DOI : 10.1145/1390156.1390267

URL : http://icml2008.cs.helsinki.fi/papers/600.pdf

M. Schlather, Models for stationary max-stable random fields, Extremes, vol.5, issue.1, pp.33-44, 2002.
DOI : 10.1023/A:1020977924878

M. Seeger, Expectation propagation for exponential families, 2005.

Y. Seldin, F. Laviolette, N. Cesa-bianchi, J. Shawe-taylor, and P. Auer, PAC-Bayesian Inequalities for Martingales, IEEE Transactions on Information Theory, vol.58, issue.12, pp.7086-7093, 2012.
DOI : 10.1109/TIT.2012.2211334

URL : http://mercurio.srv.di.unimi.it/~cesabian/Pubblicazioni/pacBayesMartin.pdf

Y. Seldin and N. Tishby, PAC?Bayesian analysis of coclustering and beyond, Journal of Machine Learning Research, vol.11, pp.3595-3646, 2010.

J. Shawe-taylor and J. Langford, PAC-Bayes & margins Advances in neural information processing systems, p.439, 2003.

S. A. Sisson, Y. Fan, and M. M. Tanaka, Sequential Monte Carlo without likelihoods, Proc. Natl. Acad. Sci. USA, pp.1760-1765, 2007.
DOI : 10.1111/j.1467-9868.2005.00497.x

URL : http://www.pnas.org/content/104/6/1760.full.pdf

N. Srebro and A. Shraibman, Rank, Trace-Norm and Max-Norm, International Conference on Computational Learning Theory, pp.545-560, 2005.
DOI : 10.1007/11503415_37

URL : http://people.csail.mit.edu/~nati/Publications/SrebroShraibmanCOLT05.pdf

N. Srebro, J. Rennie, and T. S. Jaakkola, Maximum-margin matrix factorization, Advances in neural information processing systems, pp.1329-1336, 2005.

I. Steinwart and A. Christmann, Support vector machines Information Science and Statistics, 2008.

W. Su and E. Candès, SLOPE is adaptive to unknown sparsity and asymptotically minimax, The Annals of Statistics, vol.44, issue.3, pp.1038-1068, 2016.
DOI : 10.1214/15-AOS1397SUPP

URL : http://arxiv.org/pdf/1503.08393

M. Talagrand, The generic chaining, Upper and lower bounds of stochastic processes, 2005.

M. Guo-liang-tian, . Tang, . Hong-bin, M. Fang, and . Tan, Efficient methods for estimating constrained parameters with applications to regularized (lasso) logistic regression, Computational Statistics & Data Analysis, vol.52, issue.7, pp.3528-3542, 2008.
DOI : 10.1016/j.csda.2007.11.007

A. B. Tsybakov, Optimal aggregation of classifiers in statistical learning, The Annals of Statistics, vol.32, issue.1, pp.135-166, 2004.
DOI : 10.1214/aos/1079120131

URL : https://hal.archives-ouvertes.fr/hal-00102142

B. Alexandre and . Tsybakov, Introduction to nonparametric estimation, 2009.

S. Van-de-geer, Estimation and testing under sparsity ISBN 978- 3-319-32773-0; 978-3-319-32774-7. Lecture notes from the 45th Probability Summer School held in Saint-Four, Lecture Notes in Mathematics, 2015.

A. Sara and . Van-de-geer, Applications of empirical process theory, volume 6 of Cambridge Series in Statistical and Probabilistic Mathematics, 2000.

A. Sara and . Van-de-geer, High-dimensional generalized linear models and the lasso. The Annals of Statistics, pp.614-645, 2008.

N. Vladimir and . Vapnik, Statistical learning theory Adaptive and Learning Systems for Signal Processing, Communications, and Control, 1998.

C. Varin, N. Reid, and D. Firth, An overview of composite likelihood methods, Statist. Sinica, vol.21, issue.1, pp.5-42, 2011.

G. Watson, Characterization of the subdifferential of some matrix norms. Linear algebra and its applications, pp.33-45, 1992.

S. R. White, T. Kypraios, and S. P. Preston, Piecewise Approximate Bayesian Computation: fast inference for discretely observed Markov models using a factorised posterior distribution, Statistics and Computing, vol.9, issue.1, pp.289-301, 2015.
DOI : 10.1007/978-1-4899-4493-1

URL : https://link.springer.com/content/pdf/10.1007%2Fs11222-013-9432-2.pdf

R. D. Wilkinson, Accelerating ABC methods using Gaussian processes . ArXiv preprint 1401, 1436.

S. Y?ld?r?m, S. Sumeetpal, T. Singh, A. Dean, and . Jasra, Parameter Estimation in Hidden Markov Models With Intractable Likelihoods Using Sequential Monte Carlo, Journal of Computational and Graphical Statistics, vol.24, issue.3, pp.846-865, 2014.
DOI : 10.1214/13-EJS819

M. Yuan and Y. Lin, Model selection and estimation in regression with grouped variables, Journal of the Royal Statistical Society: Series B (Statistical Methodology), vol.58, issue.1, pp.49-67, 2006.
DOI : 10.1198/016214502753479356

URL : http://www2.isye.gatech.edu/~myuan/papers/glasso.final.pdf

T. Zhang, Statistical behavior and consistency of classification methods based on convex risk minimization. The Annals of Statistics, pp.56-85, 2004.