Information and Coding Theory Seminar  


Abstracts of talks

Fall 2005

2/2 Grigory Kabatiansky (IPPI, Moscow), Collusion secure digital fingerprinting: the case of two pirates revisited

Abstract: We consider the problem of digital fingerprinting codes in the simplest case of two "pirates". We calculate the rate achievable via the family of random linear codes and discuss upper bounds in order to estimate the capacity of the corresponding channel. A related problem of so-called tracing traitors codes will be also discussed.

2/9 Adrian Papamarcou (UMD), Writing on Dirty Paper: A Geometrical Approach

Abstract: It was shown by Costa (1981) that the capacity of the power- limited AWGN channel is not reduced by the presence of additive Gaussian interference known to the transmitter but not the receiver. The proof of that result was based on a previously known single-letter characterization of the capacity (due to Gel'fand and Pinsker) which involved an auxiliary variable; the iid realization of that variable was also instrumental in the random coding argument of the proof. In this talk, we briefly review the role played by n- dimensional hyperspheres in the representation and visualization of codes for the AWGN, and proceed to develop and interpret the Costa result using a geometrical approach that does not rely on auxiliary variables or distributions.

2/16 Jon Feldman (Columbia U.), Linear programming decoding

Abstract: In this talk we introduce the method of LP decoding, and show how the error rate of an LP decoder can be characterized as the mass of a certain "normal cone" under a probability distribution defined by the channel. We then apply LP decoding to low-density parity-check (LDPC) codes, and study the relationship between "polytope vertices" and "graph covers," (Koetter and Vontobel) unified under the notion of a "pseudocodeword." We then use a "dual witness" to prove a lower bound on the number of errors that can be corrected by an LP decoder, as long as the graph expands sufficiently.

(Joint work with David Karger, Tal Malkin, Rocco Servedio, Cliff Stein, and Martin Wainwright)

3/2 Olgica Milenkovic (U. Colorado, Boulder), Constrained and Error-Control Coding for DNA Computers

Abstract: The last century was marked by the birth of two outstanding scientific and engineering disciplines: silicon based computing and the theory and technology of genetic data analysis. The research field very likely to dominate the area of scientific computing in the foreseeable future is the merger of these two disciplines, leading to unprecedented possibilities for applications in varied areas of biology and engineering. The first steps toward this goal were made in 1994, when Leonard Adleman solved a quite unremarkable computational problem with an outstanding method. The computational problem considered by Adleman was a simple instant of the directed travelling salesmen problem. The technique used for solving the problem was a new technological paradigm, termed DNA computing. Adleman's experiment represents a landmark demonstration of data processing and communication on the level of biological molecules.
    Following Adleman's initial work, many other DNA computing schemes were developed, including solvers for the 3-SAT problem, DNA systems for analyzing chess games, DNA-based logical components, as well as an intelligent, interactive DNA "game machine", called MAYA. Other applications of DNA techniques include DNA-based storage, DNA biosensing, molecular signal processing and intra-cell disease diagnostics and treatment.
    One of the major obstacles to efficient DNA computing, storage and signal processing is the very low reliability of DNA hybridization operations. DNA computing experiments require the creation of a controlled environment that allows for a set of DNA oligonucleotide strands (codewords) to hybridize with their complements in an appropriate fashion. If the DNA codewords are not carefully chosen, unwanted, or non-selective, hybridization may occur. Even more detrimental is the fact that an oligonucleotide sequence may fold back onto itself, forming a secondary structure which prevents the sequence from participating in the computational process.
    In this talk, we describe some new DNA code design methods which meet the hybridization-constraint requirements. Several classes of such codes are cyclic, which is a particularly useful property in light of the fact that the presence of a cyclic structure reduces the complexity of testing DNA codes for secondary structure formation. Based on some simple properties of a dynamic-programming algorithm, known as Nussinov's method, which is an effective predictor of secondary structure given the sequence of bases in a DNA codeword, we identify some code design criteria that reduce the possibility of secondary structure formation in a codeword. These design criteria can be formulated in terms of the requirement that the Watson-Crick distance between a DNA codeword and a number of its shifts be larger than a given threshold or that the DNA codewords do not contain long subsequences and their reverse complements. In this talk we address both the issue of enumerating DNA sequences with such properties and the problem of practical DNA code construction under such constraints.
Joint work with Navin Kashyap from Queen’s University, Kingston, Canada.

3/9 Damianos Karakos (Johns Hopkins), An Introduction to Decision Trees

Abstract:
Classification (decision) trees are usually built in a supervised manner, where a set of observation-label pairs {(X,Y)} is provided. For each input observation X, a series of decisions is taken, as to which branch of the tree X will follow at each internal node. Once X reaches a leaf, a leaf-specific "label" hat{Y} is chosen as the output. The tree is designed to minimize (at least approximately) the expected loss between Y and hat{Y}. In this talk, we will present the fundamentals of decision trees, and we will describe various locally optimal practical procedures for growing them. We will also see some problems that have been successfully tackled using decision trees.


4/1 Tom Richardson (Flarion Technologies), Design and Modeling Issues for Mobile Wireless Data: An introduction to Flash OFDM
(CSHCN Colloquium)


Abstract: Recent experience in a deployment of Flash OFDM in the Raleigh area by Nextel has shown that Flash OFDM offers the best technological solution for mobile wireless data. The transition from wireless voice to wireless data necessitates a reconsideration of both system networking design and the wireless channel. This talk will introduce some of the key ideas that underpin and differentiate the Flash OFDM system.

4/6 Alexander Barg (UMD), Quantum Codes: An Introduction

Abstract: A tutorial on the basics of quantum algorithms and error correction.

4/13 Alexander Barg (UMD), Quantum Codes: An Introduction (continued)

4/20 Punarbasu Purkayastha (UMD), Weight distributions of LDPC codes

Abstract: This talk will mainly focus on how the asymptotic distance spectrum of regular LDPC codes can be estimated using saddle point approximation, and how these estimates can be used to provide bounds on the error rate under ML Decoding.

References: [1] David Burshtein, Gadi Miller, "Asymptotic Enumeration Methods for Analyzing LDPC Codes", IEEE Transactions on Information Theory, June 2004.
[2] Ohad Barak, David Burshtein, "Lower Bounds on the Spectrum and Error Rate of LDPC Code Ensembles"

4/27 Prakash Narayan (UMD), The Minimum Description Length (MDL) Principle

Abstract: Rissanen's Minimum Description Length (MDL) principle, which lies at the interface of information theory and statistics, states that the statistical information contained in data is best extracted in terms of a possibly short and invertible description (code) of the data. The statistical model (e.g., probability distribution) inferred from the data (as best fitting the data) is the one that leads to the shortest description, taking into account that the inferred model (e.g., probability distribution) itself must also be described. The entrails of the MDL principle will be joyously examined.

References:
[1] I. Csiszar and P.C. Shields, Notes on Information Theory and Statistics, available in Foundations and Trends in Communications and Information Theory, S. Verdu, Ed., 2004. (The relevant material appears in Section 7 - pdf)
[2] A. Barron, J. Rissanen and B. Yu, "The Minimum Description Length Principle in Coding and Modeling",IEEE Trans IT, vol. 44, no. 6, pp. 2743-2760, Oct. 1998.

5/4 Martin Wainwright (UC Berkeley), Tree-reweighted message-passing algorithms in graphical models: Some theory and applications (CSPL seminar)

Abstract: Graphical models (e.g., factor graphs, Markov random fields) provide a flexible framework for capturing statistical dependencies among large collections of random variables, and are widely used in various fields, including statistical signal processing, coding and communication theory, and machine learning. Central to the wide-spread use of graphical models are distributed ``message-passing'' algorithms (e.g., sum-product, max-product), in which nodes perform local computations and communicate with their neighbors. The purpose of these algorithms is to solve various statistical inference problems (e.g., image denoising, error-control decoding) that arise in applying a graphical model. These methods are well-known to be exact for trees, but they are also widely applied to graphs with cycles, where they yield approximate answers.
    To motivate our work, we describe some potential pitfalls associated with applying the standard forms of message-passing to graphs with cycles (sum-product: multiple fixed points and instability; max-product: misleading outputs). We then describe a novel family of tree-reweighted algorithms, and present some theoretical guarantees on their behavior for graphs with cycles. The tree-reweighted sum-product algorithm has a unique fixed point for any problem, and it can be understood as computing an optimal upper bound on the log partition function. On the other hand, the tree-reweighted max-product algorithm has the desirable property of never deceiving the user; this behavior can be understood in terms of a particular linear programming relaxation of the mode-finding problem. We conclude with an overview of some applications of these modified algorithms. Talk based on joint work with: Jon Feldman, Tommi Jaakkola, Alan Willsky.

5/12 Pierre Moulin (UIUC) (CSPL seminar, note the unusual day), Communication games with side information

In this talk we consider some communication problems with channel uncertainty and attempt to derive the worst noise distribution within a prescribed class of distributions. For point-to-point communications this is a classical problem, for which elegant solutions have been derived over the last forty years.
    Motivated by problems such as data hiding and watermarking, we have recently been interested in extending this work to communication problems with side information. Optimal strategies for the communicator and the adversary (jammer) may then be derived by selecting an appropriate cost function and solving minmax optimization problems with respect to the variables of the game. The inner maximization problem (over the noise distribution) is infinite-dimensional but concave. For detection problems the worst noise distributions are strongly non-Gaussian; in some cases exotic solutions are obtained. We also quantify the benefits of randomized strategies for the communicator. This methodology is illustrated by applications to scalar and hexagonal quantization codes for data hiding.