E. Inaki, A. Aguerri, and . Zaidi, Distributed variational representation learning, IEEE Transactions on Pattern Analysis and Machine Intelligence, 2019.

T. Berger, Z. Zhang, and H. Viswanathan, The CEO problem, IEEE Transactions on Information Theory, vol.42, issue.3, pp.887-902, 1996.

Y. Oohama, Rate-distortion theory for Gaussian multiterminal source coding systems with several side informations at the decoder, IEEE Transactions on Information Theory, vol.51, issue.7, pp.2577-2593, 2005.

V. Prabhakaran, D. Tse, and K. Ramachandran, Rate region of the quadratic Gaussian CEO problem, Proceedings of IEEE International Symposium on Information Theory, p.117, 2004.

J. Chen and J. Wang, On the vector Gaussian CEO problem, Proceedings of IEEE International Symposium on Information Theory, pp.2050-2054, 2011.

J. Wang and J. Chen, On the vector Gaussian L-terminal CEO problem, Proceedings of IEEE International Symposium on Information Theory, pp.571-575, 2012.

T. Liu and P. Viswanath, An extremal inequality motivated by multiterminal information-theoretic problems, IEEE Transactions on Information Theory, vol.53, issue.5, pp.1839-1851, 2007.

Y. Xu and Q. Wang, Rate region of the vector Gaussian CEO problem with the trace distortion constraint, IEEE Transactions on Information Theory, vol.62, issue.4, pp.1823-1835, 2016.

A. Thomas, R. D. Courtade, and . Wesel, Multiterminal source coding with an entropy-based distortion measure, Proceedings of IEEE International Symposium on Information Theory, pp.2040-2044, 2011.

A. Thomas, T. Courtade, and . Weissman, Multiterminal source coding under logarithmic loss, IEEE Transactions on Information Theory, vol.60, issue.1, pp.740-761, 2014.

E. Ekrem and S. Ulukus, An outer bound for the vector Gaussian CEO problem, IEEE Transactions on Information Theory, vol.60, issue.11, pp.6870-6887, 2014.

S. Tavildar and P. Viswanath, On the sum-rate of the vector Gaussian CEO problem, Proceedings of 39-th Asilomar Conference on Signals, Systems, and Computers, pp.3-7, 2005.

H. Weingarten, Y. Steinberg, and S. Shamai, The capacity region of the gaussian multiple-input multiple-output broadcast channel, IEEE Transactions on Information Theory, vol.52, issue.9, pp.3936-3964, 2006.

D. Perez-palomar, J. M. Cioffi, and M. A. Lagunas, Joint Tx-Rx beamforming design for multicarrier MIMO channels: A unified framework for convex optimization, IEEE Transactions on Signal Processing, vol.51, issue.9, pp.2381-2401, 2003.

A. Scaglione, P. Stoica, S. Barbarossa, G. B. Giannakis, and H. Sampath, Optimal designs for space-time linear precoders and decoders, IEEE Transactions on Signal Processing, vol.50, issue.5, pp.1051-1064, 2002.

S. Md, A. B. Rahman, and . Wagner, On the optimality of binning for distributed hypothesis testing, IEEE Transactions on Information Theory, vol.58, issue.10, pp.6282-6303, 2012.

N. Tishby, F. C. Pereira, and W. Bialek, The information bottleneck method, Proceedings of the 37-th Annual Allerton Conference on Communication, Control and Computing, pp.368-377, 1999.

P. Harremoes and N. Tishby, The information bottleneck revisited or how to choose a good distortion measure, Proceedings of IEEE International Symposium on Information Theory, pp.566-570, 2007.

R. E. Blahut, Computation of channel capacity and rate-distortion functions, IEEE Transactions on Information Theory, vol.18, issue.4, pp.460-473, 1972.

S. Arimoto, An algorithm for computing the capacity of arbitrary discrete memoryless channels, IEEE Transactions on Information Theory, vol.18, issue.1, pp.14-20, 1972.

G. Chechik, A. Globerson, N. Tishby, and Y. Weiss, Information bottleneck for Gaussian variables, Journal of Machine Learning Research, vol.6, pp.165-188, 2005.

A. Winkelbauer and G. Matz, Rate-information-optimal Gaussian channel output compression, Proceedings of the 48-th Annual Conference on Information Sciences and Systems, 2014.

S. Cheng, V. Stankovic, and Z. Xiong, Computing the channel capacity and rate-distortion function with two-sided state information, IEEE Transactions on Information Theory, vol.51, issue.12, pp.4418-4425, 2005.

M. Chiang and S. Boyd, Geometric programming duals of channel capacity and rate distortion, IEEE Transactions on Information Theory, vol.50, issue.2, pp.245-258, 2004.

F. Dupuis, W. Yu, and F. M. Willems, Blahut-Arimoto algorithms for computing channel capacity and rate-distortion with side information, Proceedings of IEEE International Symposium on Information Theory, p.181, 2004.

M. Rezaeian and A. Grant, A generalization of Arimoto-Blahut algorithm, Proceedings of IEEE International Symposium on Information Theory, p.180, 2004.

I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot et al., ?-vae: Learning basic visual concepts with a constrained variational framework, Proceedings of the 5-th International Conference on Learning Representations, 2017.

A. A. Alemi, B. Poole, I. Fischer, I. Fischer, J. V. Dillon et al., Fixing a broken ELBO, Proceedings of the 35-th International Conference on Machine Learning, 2018.

P. Diederik, M. Kingma, and . Welling, Auto-encoding variational bayes, Proceedings of the 2-nd International Conference on Learning Representations, 2014.

A. A. Alemi, I. Fischer, J. V. Dillon, and K. Murphy, Deep variational information bottleneck, Proceedings of the 5-th International Conference on Learning Representations, 2017.

Z. Jiang, Y. Zheng, H. Tan, B. Tang, and H. Zhou, Variational deep embedding: An unsupervised and generative approach to clustering, Proceedings of the 26-th International Joint Conference on Artificial Intelligence, pp.1965-1972, 2017.

N. Slonim, The Information Bottleneck: Theory and Applications, 2002.

J. Xie, R. Girshick, and A. Farhadi, Unsupervised deep embedding for clustering analysis, Proceedings of the 33-rd International Conference on Machine Learning, pp.478-487, 2016.

H. S. Witsenhausen, Indirect rate distortion problems, IEEE Transactions on Information Theory, issue.5, pp.518-521, 1980.

Y. Steinberg, Coding and common reconstruction, IEEE Transactions on Information Theory, vol.55, issue.11, pp.4995-5010, 2009.

I. Sutskover, S. Shamai, ;. , and J. Ziv, Extremes of information combining, IEEE Transactions on Information Theory, vol.51, issue.4, pp.1313-1325, 2005.

I. Land and J. Huber, Information combining, Foundations and Trends in Communication and Information Theory, vol.3, issue.3, pp.227-330, 2006.

I. Land, S. Huettinger, P. A. Hoeher, and J. B. Huber, Bounds on information combining, IEEE Transactions on Information Theory, vol.51, issue.2, pp.612-619, 2005.

A. D. Wyner, On source coding with side information at the decoder, IEEE Transactions on Information Theory, vol.21, issue.3, pp.294-300, 1975.

R. Ahlswede and J. Korner, Source coding with side information and a converse for degraded broadcast channels, IEEE Transactions on Information Theory, vol.21, issue.6, pp.629-637, 1975.

E. Erkip and T. Cover, The efficiency of investment information, IEEE Transactions on Information Theory, vol.44, issue.3, pp.1026-1040, 1998.

A. Makhdoumi, S. Salamatian, N. Fawaz, and M. Medard, From the information bottleneck to the privacy funnel, Proceedings of IEEE Information Theory Workshop, pp.501-505, 2014.

Y. Bengio, A. Courville, and P. Vincent, Representation learning: A review and new perspectives, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.35, issue.8, pp.1798-1828, 2013.

C. Xu, D. Tao, and C. Xu, A survey on multi-view learning, 2013.

W. Wang, R. Arora, K. Livescu, and J. Bilmes, On deep multi-view representation learning, Proceedings of the 32-nd International Conference on Machine Learning, 2015.

Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, Gradient-based learning applied to document recognition, Proceedings of the IEEE, vol.86, pp.2278-2324, 1998.

D. D. Lewis, Y. Yang, T. G. Rose, and F. Li, A new benchmark collection for text categorization research, The Journal of Machine Learning Research, vol.5, pp.361-397, 2004.

A. Coates, A. Ng, and H. Lee, An analysis of single-layer networks in unsupervised feature learning, Proceedings of the 14-th International Conference on Artificial Intelligence and Statistics, pp.215-223, 2011.

G. Pichler, P. Piantanida, and G. Matz, Distributed information-theoretic biclustering, Proceedings of IEEE International Symposium on Information Theory, pp.1083-1087, 2016.

G. Pichler, P. Piantanida, and G. Matz, A multiple description CEO problem with log-loss distortion, Proceedings of IEEE International Symposium on Information Theory, pp.111-115, 2017.
URL : https://hal.archives-ouvertes.fr/hal-01742330

J. Jiao, T. A. Courtade, K. Venkat, and T. Weissman, Justification of logarithmic loss via the benefit of side information, IEEE Transactions on Information Theory, vol.61, issue.10, pp.5357-5365, 2015.

A. No and T. Weissman, Universality of logarithmic loss in lossy compression, Proceedings of IEEE International Symposium on Information Theory, pp.2166-2170, 2015.

Y. Shkel, M. Raginsky, and S. Verdu, Universal lossy compression under logarithmic loss, Proceedings of IEEE International Symposium on Information Theory, pp.1157-1161, 2017.

N. Cesa, -. Bianchi, and G. Lugosi, Prediction, Learning and Games, 2006.

T. Andre, M. Antonini, M. Barlaud, and R. M. Gray, Entropy-based distortion measure for image coding, Proceedings of IEEE International Conference on Image Processing, pp.1157-1160, 2006.

K. Kittichokechai, Y. Chia, T. J. Oechtering, M. Skoglund, and T. Weissman, Secure source coding with a public helper, IEEE Transactions on Information Theory, vol.62, issue.7, pp.3930-3949, 2016.

A. Painsky and G. Wornell, On the universality of the logistic loss function, Proceedings of IEEE International Symposium on Information Theory, pp.936-940, 2018.

X. Cheuk-ting-li, A. Wu, A. E. Ozgur, and . Gamal, Minimax learning for remote prediction, Proceedings of IEEE International Symposium on Information Theory, pp.541-545, 2018.

C. Tian and J. Chen, Remote vector Gaussian source coding with decoder side information under mutual information and distortion constraints, IEEE Transactions on Information Theory, vol.55, issue.10, pp.4676-4680, 2009.

A. Sanderovich, S. Shamai, ). Shitz, Y. Steinberg, and G. Kramer, Communication via decentralized processing, IEEE Transactions on Information Theory, vol.54, issue.7, pp.3008-3023, 2008.

O. Simeone, E. Erkip, and S. Shamai, On codebook information for interference relay channels with out-of-band relaying, IEEE Transactions on Information Theory, vol.57, issue.5, pp.2880-2888, 2011.

A. Inaki-estella-aguerri, G. Zaidi, S. Caire, and . Shamai, On the capacity of cloud radio access networks with oblivious relaying, Proceedings of IEEE International Symposium on Information Theory, pp.2068-2072, 2017.

A. Inaki-estella-aguerri, G. Zaidi, S. Caire, and . Shamai, On the capacity of cloud radio access networks with oblivious relaying, IEEE Transactions on Information Theory, vol.65, issue.Shitz, pp.4575-4596, 2019.

P. Flavio, A. Calmon, M. Makhdoumi, M. Medard, M. Varia et al., Principal inertia components and applications, IEEE Transactions on Information Theory, vol.63, issue.8, pp.5011-5038, 2017.

R. Ahlswede and I. Csiszar, Hypothesis testing with communication constraints, IEEE Transactions on Information Theory, issue.4, pp.533-542, 1986.

. Te-sun-han, Hypothesis testing with multiterminal data compression, IEEE Transactions on Information Theory, issue.6, pp.759-772, 1987.

C. Tian and J. Chen, Successive refinement for hypothesis testing and lossless one-helper problem, IEEE Transactions on Information Theory, vol.54, issue.10, pp.4666-4681, 2008.

S. Salehkalaibar, M. Wigger, and R. Timo, On hypothesis testing against conditional independence with multiple decision centers, IEEE Transactions on Communications, vol.66, issue.6, pp.2409-2420, 2018.
URL : https://hal.archives-ouvertes.fr/hal-02440688

R. Gilad-bachrach, A. Navot, and N. Tishby, An information theoretic tradeoff between complexity and accuracy, Proceedings of Conference on Learning Theory, pp.595-609, 2003.

A. Winkelbauer, S. Farthofer, and G. Matz, The rate-information tradeoff for Gaussian vector channels, Proceedings of IEEE International Symposium on Information Theory, pp.2849-2853, 2014.

M. Meidlinger, A. Winkelbauer, and G. Matz, On the relation between the Gaussian information bottleneck and MSE-optimal rate-distortion quantization, Proceedings of IEEE Workshop on Statistical Signal Processing, pp.89-92, 2014.

A. Zaidi, I. E. Aguerri, and S. Shamai, On the information bottleneck problems: Models, connections, applications and information theoretic views, Entropy, vol.22, issue.2, p.151, 2020.

A. D. Wyner and J. Ziv, The rate-distortion function for source coding with side information at the decoder, IEEE Transactions on Information Theory, issue.1, pp.1-10, 1976.

M. Benammar and A. Zaidi, Rate-distortion of a Heegard-Berger problem with common reconstruction constraint, Proceedings of International Zurich Seminar on Communications, pp.150-154, 2016.

M. Benammar and A. Zaidi, Rate-distortion function for a Heegard-Berger problem with two sources and degraded reconstruction sets, IEEE Transactions on Information Theory, vol.62, issue.9, pp.5080-5092, 2016.
URL : https://hal.archives-ouvertes.fr/hal-01807725

P. Flavio-du, N. Calmon, and . Fawaz, Privacy against statistical inference, Proceedings of the 50-th Annual Allerton Conference on Communication, Control and Computing, pp.1401-1408, 2012.

S. Asoodeh, M. Diaz, F. Alajaji, and T. Linder, Information extraction under privacy constraints, Information, vol.7, issue.15, 2016.

A. Achille and S. Soatto, Information dropout: Learning optimal representations through noisy computation, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.40, issue.12, pp.2897-2905, 2018.

S. Watanabe, Information theoretical analysis of multivariate correlation, IBM Journal of Research and Development, vol.4, issue.1, pp.66-82, 1960.

T. Q. Ricky, X. Chen, R. Li, D. Grosse, and . Duvenaud, Isolating sources of disentanglement in VAEs, Proceedings of the 32-nd Conference on Neural Information Processing Systems, 2018.

O. Shamir, S. Sabato, and N. Tishby, Learning and generalization with the information bottleneck, Proceedings of the 19-th International Conference on Algorithmic Learning Theory, pp.92-107, 2008.

N. Tishby and N. Zaslavsky, Deep learning and the information bottleneck principle, Proceedings of IEEE Information Theory Workshop, 2015.

P. Diederik, J. Kingma, and . Ba, Adam: A method for stochastic optimization, Proceedings of the 3-rd International Conference on Learning Representations, 2015.

R. Schwartz-ziv and N. Tishby, Opening the black box of deep neural networks via information, 2017.

A. Michael-saxe, Y. Bansal, J. Dapello, M. Advani, A. Kolchinsky et al., On the information bottleneck theory of deep learning, Proceedings of the 6-th International Conference on Learning Representations, 2018.

G. Lewandowsky and . Bauch, Information-optimum LDPC decoders based on the information bottleneck method, IEEE Access, vol.6, pp.4054-4071, 2018.

M. Meidlinger, A. Balatsoukas-stimming, A. Burg, and G. Matz, Quantized message passing for LDPC codes, Proceedings of 49-th Asilomar Conference on Signals, Systems, and Computers, pp.1606-1610, 2015.

J. Korner and K. Marton, How to encode the modulo-two sum of binary sources, IEEE Transactions on Information Theory, vol.25, issue.02, pp.219-221, 1979.

M. Gastpar, The Wyner-Ziv problem with multiple sources, IEEE Transactions on Information Theory, vol.50, issue.11, pp.2762-2768, 2004.

D. Russo and J. Zou, How much does your data exploration overfit? Controlling bias via information usage, 2015.

A. Xu and M. Raginsky, Information-theoretic analysis of generalization capability of learning algorithms, Proceedings of the 31-st Conference on Neural Information Processing Systems, pp.2524-2533, 2017.

R. Amir, E. Asadi, S. Abbe, and . Verdu, Chaining mutual information and tightening generalization bounds, Proceedings of the 32-nd Conference on Neural Information Processing Systems, 2018.

T. Berger, Decentralized estimation and decision theory, Proceedings of IEEE Spring Workshop on Information Theory, 1979.

M. H. Hossam, A. Shalaby, and . Papamarcou, Multiterminal detection with zerorate data compression, IEEE Transactions on Information Theory, vol.38, issue.2, pp.254-267, 1992.

W. Zhao and L. Lai, Distributed testing with zero-rate compression, Proceedings of IEEE International Symposium on Information Theory, pp.2792-2796, 2015.

P. Escamilla, M. Wigger, and A. Zaidi, Distributed hypothesis testing with concurrent detections, Proceedings of IEEE International Symposium on Information Theory, pp.166-170, 2018.
URL : https://hal.archives-ouvertes.fr/hal-01806545

P. Escamilla, M. Wigger, and A. Zaidi, Distributed hypothesis testing with collaborative detection, Proceedings of the 56-th Annual Allerton Conference on Communication, Control, and Computing, pp.512-518, 2018.
URL : https://hal.archives-ouvertes.fr/hal-02288547

J. Liao, L. Sankar, P. Flavio, . Calmon, Y. F. Vincent et al., Hypothesis testing under maximal leakage privacy constraints, Proceedings of IEEE International Symposium on Information Theory, pp.779-783, 2017.

S. Sreekumar, A. Cohen, and D. Gunduz, Distributed hypothesis testing with a privacy constraint, Proceedings of IEEE Information Theory Workshop, 2018.

A. Zaidi and I. E. Aguerri, Optimal rate-exponent region for a class of hypothesis testing against conditional independence problems, Proceedings of IEEE Information Theory Workshop, 2019.

T. Berger, Multiterminal source coding. The Information Theory Approach to Communications, CSIM Courses and Lectures, vol.229, 1978.

S. Y. Tung, Multiterminal source coding, 1978.

Y. Zhou, Y. Xu, W. Yu, and J. Chen, On the optimal fronthaul compression and decoding strategies for uplink cloud radio access networks, IEEE Transactions on Information Theory, vol.62, issue.12, pp.7402-7418, 2016.

T. A. Courtade, Gaussian multiterminal source coding through the lens of logarithmic loss, Information Theory and Applications Workshop, 2015.

T. A. Courtade, A strong entropy power inequality, IEEE Transactions on Information Theory, vol.64, issue.4, pp.2173-2192, 2018.

A. B. Wagner, S. Tavildar, and P. Viswanath, Rate region of the quadratic Gaussian two-encoder source-coding problem, IEEE Transactions on Information Theory, vol.54, issue.5, pp.1938-1961, 2008.

A. Thomas, J. Courtade, and . Jiao, An extremal inequality for long Markov chains, Proceedings of the 52-nd Annual Allerton Conference on Communication, Control and Computing, pp.763-770, 2014.

Y. Oohama, The rate-distortion function for the quadratic gaussian ceo problem, IEEE Transactions on Information Theory, vol.44, issue.3, pp.1057-1070, 1998.

S. Tavildar, P. Viswanath, and A. B. Wagner, The gaussian many-help-one distributed source coding problem, IEEE Transactions on Information Theory, vol.56, issue.1, pp.564-581, 2010.

S. Md, A. B. Rahman, and . Wagner, Rate region of the vector gaussian one-helper source-coding problem, IEEE Transactions on Information Theory, vol.61, issue.5, pp.2708-2728, 2015.

E. Inaki, A. Aguerri, and . Zaidi, Distributed information bottleneck method for discrete and Gaussian sources, Proceedings of International Zurich Seminar on Information and Communication, 2018.

N. Slonim and N. Tishby, The power of word clusters for text classification, Proceedings of 23-rd European Colloquium on Information Retrieval Research, pp.191-200, 2001.

Y. Baram, R. El-yaniv, and K. Luz, Online choice of active learning algorithms, Journal of Machine Learning Research, vol.5, pp.255-291, 2004.

J. Chen and T. Berger, Successive Wyner-Ziv coding scheme and its application to the quadratic Gaussian CEO problem, IEEE Transactions on Information Theory, vol.54, issue.4, pp.1586-1603, 2008.

M. Meisam-razaviyayn, Z. Hong, and . Luo, A unified convergence analysis of block successive minimization methods for nonsmooth optimization, SIAM Journal on Optimization, vol.23, issue.2, pp.1126-1153, 2013.

M. Grant and S. Boyd, CVX: Matlab software for disciplined convex programming, 2014.

M. Chalk, O. Marre, and G. Tkacik, Relevant sparse codes with variational information bottleneck, Proceedings of the 30-th Conference on Neural Information Processing Systems, 2016.

A. Xue-bin-peng, S. Kanazawa, P. Toyer, S. Abbeel, and . Levine, Variational discriminator bottleneck: Improving imitation learning, inverse rl, and gans by constraining information flow, Proceedings of the 7-th International Conference on Learning Representations, 2019.

B. Dai, C. Zhu, and D. P. Wipf, Compressing neural networks using the variational information bottleneck, Proceedings of the 35-th International Conference on Machine Learning, 2018.

P. Diederik, T. Kingma, R. Salimans, X. Jozefowicz, I. Chen et al., Improved variational inference with inverse autoregressive flow, Proceedings of 30-st Conference on Neural Information Processing Systems, 2016.

G. Papamakarios, T. Pavlakou, and I. Murray, Masked autoregressive flow for density estimation, Proceedings of 31-st Conference on Neural Information Processing Systems, 2017.

D. Sculley, Web-scale K-means clustering, Proceedings of the 19-th International Conference on World Wide Web, pp.1177-1178, 2010.

Z. Huang, Extensions to the K-means algorithm for clustering large datasets with categorical values, Data Mining and Knowledge Discovery, vol.2, issue.3, pp.283-304, 1998.

J. A. Hartigan and M. A. Wong, Algorithm AS 136: A k-means clustering algorithm, Journal of the Royal Statistical Society, vol.28, pp.100-108, 1979.

A. P. Dempster, N. M. Laird, and D. B. Rubin, Maximum likelihood from incomplete data via the EM algorithm, Journal of the Royal Statistical Society, vol.39, pp.1-38, 1977.

C. Ding and X. He, K-means clustering via principal component analysis, Proceedings of the 21-st International Conference on Machine Learning, 2004.

K. Pearson, On lines and planes of closest fit to systems of points in space, Philosophical Magazine, vol.2, issue.11, pp.559-572, 1901.

S. Wold, K. Esbensen, and P. Geladi, Principal component analysis, Chemometrics and Intelligent Laboratory Systems, vol.2, pp.37-52, 1987.

S. Roweis, EM algorithms for PCA and SPCA, Advances in Neural Information Processing Systems, vol.10, pp.626-632, 1997.

T. Hofmann, B. Schölkopf, and A. J. Smola, Kernel methods in machine learning, The Annals of Statistics, vol.36, pp.1171-1220, 2008.

N. Slonim and N. Tishby, Document clustering using word clusters via the information bottleneck method, Proceedings of the 23-rd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, pp.208-215, 2000.

D. J. Rezende, S. Mohamed, and D. Wierstra, Stochastic backpropagation and approximate inference in deep generative models, Proceedings of the 31-st International Conference on Machine Learning, pp.1278-1286, 2014.

X. Guo, L. Gao, X. Liu, and J. Yin, Improved deep embedded clustering with local structure preservation, Proceedings of the 26-th International Joint Conference on Artificial Intelligence, pp.1753-1759, 2017.

N. Dilokthanakul, P. A. Mediano, M. Garnelo, C. H. Matthew, H. Lee et al., Deep unsupervised clustering with Gaussian mixture variational autoencoders, 2017.

E. Min, X. Guo, Q. Liu, G. Zhang, J. Cui et al., A survey of clustering with deep learning: From the perspective of network architecture, IEEE Access, vol.6, pp.39-501, 2018.

R. John, P. A. Hershey, and . Olsen, Approximating the Kullback Leibler divergence between Gaussian mixture models, Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing, pp.317-320, 2007.

K. He, X. Zhang, S. Ren, and J. Sun, Deep residual learning for image recognition, Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp.770-778, 2016.

P. Vincent, H. Larochelle, I. Lajoie, Y. Bengio, and P. Manzagol, Stacked denoising autoencoders: Learning useful representations in a deep network with a local denoising criterion, The Journal of Machine Learning Research, vol.11, pp.3371-3408, 2010.

L. Van-der-maaten and G. Hinton, Visualizing data using t-SNE, Journal of Machine Learning Research, vol.9, pp.2579-2605, 2008.

T. Cheuk, A. E. Li, and . Gamal, Strong functional representation lemma and applications to coding theorems, IEEE Transactions on Information Theory, vol.64, issue.11, pp.6967-6978, 2018.

X. Cheuk-ting-li, A. Wu, A. E. Ozgur, and . Gama, Minimax learning for remote prediction, 2018.

A. Homri, M. Peleg, and S. Shamai, Oblivious fronthaulconstrained relay for a Gaussian channel, IEEE Transactions on Communications, vol.66, issue.11, pp.5112-5123, 2018.

R. Karasik, O. Simeone, and S. Shamai, Robust uplink communications over fading channels with variable backhaul connectivity, IEEE Transactions on Wireless Communications, vol.12, issue.Shitz, pp.5788-5799, 2013.

Y. Chen, A. J. Goldsmith, and Y. C. Eldar, Channel capacity under sub-nyquist nonuniform sampling, IEEE Transactions on Information Theory, vol.60, issue.8, pp.4739-4756, 2014.

A. Kipnis, Y. C. Eldar, and A. J. Goldsmith, Analog-to-digital compression: A new paradigm for converting signals to bits, IEEE Signal Processing Magazine, vol.35, issue.3, pp.16-39, 2018.

M. Gastpar, On Wyner-Ziv networks, Proceedings of 37-th Asilomar Conference on Signals, Systems, and Computers, pp.855-859, 2003.

A. Dembo, T. M. Cover, and J. A. Thomas, Information theoretic inequalities, IEEE Transactions on Information Theory, vol.37, issue.6, pp.1501-1518, 1991.

P. Daniel, S. Palomar, and . Verdu, Gradient of mutual information in linear vector gaussian channels, IEEE Transactions on Information Theory, vol.52, issue.1, pp.141-154, 2006.

Y. Ugur, E. Iñaki, A. Aguerri, and . Zaidi, Vector Gaussian CEO problem under logarithmic loss and applications, IEEE Transactions on Information Theory, 2020.

Y. Ugur, E. Iñaki, A. Aguerri, and . Zaidi, A generalization of Blahut-Arimoto algorithm to compute rate-distortion regions of multiterminal source coding under logarithmic loss, Proceedings of IEEE Information Theory Workshop, pp.349-353, 2017.
URL : https://hal.archives-ouvertes.fr/hal-01806549

Y. Ugur, E. Iñaki, A. Aguerri, and . Zaidi, Vector Gaussian CEO problem under logarithmic loss, Proceedings of IEEE Information Theory Workshop, pp.515-519, 2018.

Y. Ugur, G. Arvanitakis, and A. Zaidi, Variational information bottleneck for unsupervised clustering: Deep Gaussian mixture embedding, Entropy, vol.22, issue.2, p.213, 2020.