MATHEMATICS AND COMPUTATION IN IMAGING SCIENCE AND
 INFORMATION PROCESSING
(July - December 2003)

~ Abstracts ~

Huang's Empirical Mode Decomposition and splines
Sherman D. Riemenschneider, Dept. of Mathematics, West Virginia University, USA

Cubic splines play a major role in the Empirical Mode Decomposition introduced by Norden Huang for non-stationary and nonlinear signals as means to define an instantaneous frequency via the Hilbert Spectrum. In this talk we give some elementary results about the Hilbert transforms of splines, and other roles that splines may play in the obtaining an EMD.

 « Back

Generating a good triangular mesh
Gilbert Strang, Dept. of Mathematics, Massachusetts Institute of Technology, USA

An essential first step in many problems of numerical analysis and computer graphics is to cover a region with a reasonably regular mesh.

We describe every line of a short MATLAB code that begins with a "distance function" to describe the region: d(x,y) is the distance to the boundary (with d < 0 inside the region). The code must decide the location of nodes and the choice of edges. At each iteration the mesh becomes a truss (with forces in the edges that move the nodes). Then the Delaunay algorithm resets the topology of the triangles. The input includes a size function h(x,y) to give desired meshlengths over the region.

The code is adaptable and public. It extends directly to n dimensions. The 2D meshes are remarkably regular. The theory is not complete.

 « Back

Orthogonal frames of translates
Eric S. Weber, Dept. of Mathematics, University of Wyoming, USA

A frame sequence (or Bessel sequence) defines an operator, called an analysis operator, which maps the Hilbert space into a space of coefficient sequences. The composition of the adjoint of the operator with an analysis operator from another frame sequence is a bounded operator on the Hilbert space which is a sum of rank one tensors. This operator, sometimes called a ``Mixed Dual Grammian'', is an important operator in frame theory. The duality condition of two frames corresponds to this operator being the identity.

Two Bessel sequences are said to be orthogonal if this operator is the 0 operator. The motivation for orthogonal Bessel sequences comes from the following problems:

1) Given a frame, can we find all of the dual frames?

2) How can frames be used in orthogonal division multiple access communications?

3) How can perfect reconstruction of vectors in a subspace be accomplished when the vectors are not in the subspace?

We shall describe these problems, and also describe some other applications which arise from the problems, such as cryptography and watermarking. We shall then give characterizations of when two Bessel sequences are orthogonal when the Bessel sequences have the form of translates of a finite number of functions in L2(Rd). The characterizations are applied to Bessel sequences which have an affine structure, a quasi-affine structure, a Gabor structure, and also an irregular structure. Moreover, we characterize perfect reconstruction (duality) of subspace frames for translation invariant subspaces of L2(Rd).

 « Back

Uncertainty principles and wavelets
Tim N. T. Goodman, Dept of Mathematics, The University of Dundee

Heisenberg’s uncertainty principle gives a lower bound on the ‘uncertainty product’ of a function, a measure of its spread in both time and frequency. It is seen that for certain sequences of functions satisfying self-similarity equations, the uncertainty products converge to the minimum. We then introduce the related wavelets: functions which allow decomposition simultaneously in time and frequency with orthogonality between different frequencies. For such bandpass filters the uncertainty principle is modified and we examine the convergence to the minimum of the appropriate uncertainty products of the wavelets. We also discuss uncertainty principles for periodic functions. In these cases the minimum uncertainty is not attained and we consider sequences of trigonometric polynomials whose uncertainty products converge to the minimum. These results are extended to functions on general spheres, and we also suggest an uncertainty principle on the circle for which minimum uncertainty is attained.

 « Back

CT and MR imaging
Chye Hwang Yan, DSO National Laboratories and Borys Shuter, Department of Radiology, National University Hospital

The purpose of this tutorial is to present a technical overview on computed tomography and magnetic resonance imaging. It will cover the underlying physics and also the mathematical techniques used for image reconstruction. A portion of the tutorial will address the limitations, such as artifacts and noises, of these imaging modalities. The tutorial will end with a survey on the current state of the art CT and MR machines.

 « Back

Image processing and soft computing techniques for medical applications
Kap Luk Chan, School of Electrical and Electronic Engineering Nanyang Technological University

Early detection and treatment of colorectal cancer helps reducing the number of deaths from this disease. It is thus meaningful to develop image processing and analysis techniques and apply them to endoscopic (colonscopic) images acquired through colonoscopy for this purpose. In this presentation, we report several techniques developed for detection of abnormalities in the colon from colonoscopic images which may assist subsequent analysis for detection of colorectal cancer by doctors and health care providers. We have developed edge-based techniques for abnormality detection through edge curvature analysis, region-based techniqiues through region segmentation using neural network and fuzzy-logics and etc. Both color and textural image properties are explored. In a more recent attempt, a new concept-learning approach is taken to overcome the ambiguity in the problem definition. Through learning from image data labeled by the medical experts, diagnosis models can be extracted automatically. We evaluate our methods on a set of colonoscope images. The experimental results show the feasibility of our methods and suggest promising applicability for clinical implementation.

 « Back

Current clinical applications and limitations of CT/MR imaging
Poh Sun Goh, Department of Radiology National University Hospital

This talk covers the advantages and limitations of CT/MR Imaging.

 « Back

Advanced signal processing algorithms for fMRI
Sadasivan Puthusserypady, Department of Electrical and Computer Engineering National University of Singapore

fMRI is a new technique for localising brain activity; whereas independent component analysis (ICA) is a relatively new technique for blind source separation (BSS). Principle of brain modularity states that different region of the brain perform different function independently, thus spatial ICA can be applied on fMRI. Brain signal has proven to be nonlinear, hence applying nonlinear ICA on fMRI signal should arrive at a better results. However, nonlinear ICA has the problem of non-unique solutions therefore post-nonlinear (PNL) ICA model is used. Furthermore, PNL-ICA model closely resemble to the balloon model. The balloon model is a biomechanical model of the haemodynamic of the blood system, which includes the transient states.

 « Back

Automated retinal image analysis
Sanjaya P.M.D. Perera, School of Computing, National University of Singapore

Diabetic retinopathy is a major cause of blindness in the world. Regular screening and timely intervention can halt or reverse the progression of this disease. Digital retinal imaging technologies have become an integral part of eye screening programs worldwide due to their greater accuracy and repeatability in staging diabetic retinopathy. Existing automated techniques to detect diabetic retinopathy lesions and retinal structures are neither adaptable nor sufficiently sensitive and specific for real-life screening application. This talk presents accurate and robust algorithms for automated detection of hard exudates, retinal vessels, optic disc and macular region. Hard exudates detection algorithm which incorporates domain knowledge has achieved 100% sensitivity and 72% specificity with 543 consecutive digital retinal images. Large retinal vessel detection algorithm and optic disc detection algorithm has achieved above 90% accuracy and finally macular region detection has the ability to classify hard exudates further into macular exudates with 92% sensitivity and 92% specificity.

 « Back

Non-rigid registration in MRI mammography
Ek Tsoon Tan, Department of Electrical and Computer Engineering National University of Singapore

Magnetic resonance imaging (MRI) has been suggested as an alternative to x-ray mammography because of its 3D tomography, better contrast between tissue types, and because it is radiation-free. The use of a contrast agent for highlighting regions with malignant lesions results in an intensity increase between the pre- and post-contrast scans. Image registration is required to capture involuntary patient movement and the non- uniform intensity increases, so as to correctly identify malignant lesions. Recent work in this volume registration task employed rigid and non-rigid models with entropy-based cost functions. The long computational time is the main drawback for entropy-based algorithms.

The talk presents a new adaptive Normalized Mutual Information (NMI) optimization scheme that will significantly reduce the computational requirements. The new scheme makes use of the current histogram to (1) perform a clever prediction of the change introduced by the contrast agent, and (2) adapts the standard deviation of the Parzen window for each intensity level.

 « Back

Cramer-Rao Lower Bound for parameter estimation of multidimensional NMR data
ZhiPing Lin, School of Electrical and Electronic Engineering Nanyang Technological University

Nuclear magnetic resonance (NMR) is one of the two major experimental techniques for determining proteins' 3D structures. One of the key steps in multidimensional NMR data processing is parameter estimation. This talk presents a new method for the calculation of the Fisher information matrix, the inverse of which gives the Cramer-Rao lower bound. Our method is based on a novel concept of derivative system and solutions to certain Lyapunov equations. The techniques are illustrated by a simulation example from NMR spectroscopy.

 « Back

On fast algorithms to compute canonical windows for gabor frames
A.J.E.M. Janssen, Philips Research Laboratories, The Netherlands

We present iterative algorithms, some of which were already known to Feichtinger and Strohmer, for the computation of the canonical tight (gt) and dual (gd) window associated with a Gabor frame (g, a, b). To this end we develop a mechanism for proposing iterative algorithms [involving the current frame operator and (for gt) its inverse] with a desired conver- gence rate. The algorithms so obtained with cubic convergence and using frame operators only (no inverses) are conditionally convergent, requiring a minimal value of the initial frame bound ratio A/B. These minimal values can be estimated by using a sharp form of Kantorovich's inequality.

 « Back

Tight frames and their symmetries
Shayne Waldron, Dept. of Mathematics, University of Auckland

Frame representations are useful because they are technically similar to orthogonal expansions (they simply have more terms) and can be constructed to have  desirable properties that may be impossible for an orthogonal basis, e.g., in the case of wavelets certain smoothness and small support properties.

Here we show that frames are of interest where the desirable properties are symmetries of the underlying space (which an orthogonal basis cannot express). One important example is orthogonal polynomials of several variables for a weight which has some symmetries. It turns out that there are a finite number of tight frames of n vectors in d dimensions generated by an abelian subgroup of its symmetries. We discuss how this list of so called harmonic frames was calculated using the symbolic algebra package MAGMA, and how to use it.

 « Back

Recent progress in iteratively regularized image reconstruction
Martin Hanke-Bourgeois, Johannes-Gutenberg-Universität, Germany

We survey various options for the regularized reconstruction of blurred and noisy images, given complete information about the blurring operator. The topics we will focus on include preconditioning, in particular multilevel preconditioning, and the possibility of regularization by iteration or by hybrid combinations of the Lanczos method with other regularizing schemes. We shall also briefly comment on the delicate issue of how to incorporate nonnegativity constraints.

 « Back

A Multigrid technique for restoring images with Dirichlet boundary conditions
Stefano Serra Capizzano, Dipartimento di Chimica, Fisica, e Matematica Universita dell'Insubria - Sede di Como

We propose a method of multigrid type which has an optimal computational cost and whose convergence speed is robust both with respect to the size of the problem and to the regularization parameter.

 « Back

Boundary conditions and fast deblurring models
Stefano Serra Capizzano, Dipartimento di Chimica, Fisica, e Matematica Universita dell'Insubria - Sede di Como

Anti-reflecting boundary conditions (AR-BC) for blurring models were recently introduced (see Serra-Capizzano, SISC, to appear): the idea seems promising both from the computational and approximation viewpoint. The key point is that, under certain symmetry conditions, the AR-BC matrices can be essentially simultaneously diagonalized by the (fast) sine transform DST I and, moreover, a C1 continuity at the border is guaranteed in the 1D case.

Here we give more details for the 2D case and we perform extensive numerical simulations which illustrate that the AR-BC can be superior to Dirichlet, periodic and reflective BCs in certain applications.

 « Back

Applications of 3D digital technologies in cultural heritage
Hongbin Zha, Center for Information Science, National Laboratory on Machine Perception, Peking University, China

In the presentation, I will report new progress in our research on applications of 3D digital technologies in heritage digitization and digital museum. The main topics include: 3D modeling of cultural artifacts, relics and sites; digital geometry processing for large heritage models; photorealistic visualization of cultural scenes.

 « Back

The fast and parallel processing for bit-plane coding in JPEG2000
Chao Xu, Center for Information Science, National Laboratory on Machine Perception, Peking University, China

Abstact: In JPEG2000 image compression standard the fractional bit-plane coding is adopted for the context forming for arithmetic coding. The fractional bit-plane coding generates more quality and resolution scalability, while makes more computation difficulty for fast implementation. In this paper, we present a parallel processing for bit-plane coding that reduce the computation time greatly. A preprocessor is proposed to make each bit-plane coding independent for parallel processing. The encoder is divided into multiple modules to analyse, and a method of assignment modules in terms of the statistic of module application is proposed. The parallel processing is suitable for architecture implementation. With increasing about 4 times of circuit resource it can achieve 8 times of speed-up.

 « Back

Omni-directional vision based motion detection for multiple human bodies
Hong Liu, Center for Information Science, National Laboratory on Machine Perception, Peking University, China

Computer vision systems for moving targets tracking are widely used in many applications such as smart surveillance, virtual reality, advanced user interfaces, motion analysis and model-based image coding. One of the most important issues in visual surveillance is to expand the view fields of visual sensors. Since traditional cameras suffer from the problems of having limited visual fields, lose of global visual information, hard to control for active one, we have proposed a new real-time visual system based on omni-directional vision, for detecting and tracking multiple human bodies in indoor environment. First, an omni-directional camera is used to obtain 360º view images of the scene to enlarge the monitored area. Then, omni-directional images are changed into the cylindrical panoramic images. Secondly, an adaptive subtraction method is utilized to improve the robustness for segmenting moving targets in dynamic environments. Moreover, shape and color information are used for tracking and identifying multiple peoples whether they are in groups or separated from each other. Experiments show that the system can track multiple moving human bodies with large view fields in a dynamic environment in real-time. Finally, potential applications and research plan on omni-directional vision will be discussed.

 « Back

M-band wavelet construction and compound image compression
Tong Lin, Center for Information Science, National Laboratory on Machine Perception, Peking University, China

Two parts of work are reported in this talk. One is on the algebraic construction of M-band orthogonal wavelets. A system of constraint equations is obtained for M-band orthonormal filters, and then a solution based on SVD is developed to produce innumerable wavelet bases of given length. Also the property of 2 vanishing moments is integrated into our wavelet construction process, which provides another way to compute 2-regular M-band filter banks. Another part is on compound image compression for real-time computer-generated screen transmission. A block-based lossless coding algorithm is designed for textual and graphical contents, which contains several techniques such as shape-based coding, palette-based coding, run-length coding, and palette reuse. The coded bitstream is fed into a LZW coder for further compression. Then we integrate the lossless algorithm with lossy JPEG into a hybrid image compression scheme. The key is to extract the text and graphics as foreground from photographic pictures within a low complexity constraint. The shape-based coding in our lossless algorithm elegantly provides the intelligence to extract computer-generated text/graphics without efforts.

 « Back

Medical image series segmentation based on modified active contour model
Jie Tian, Institute of Automation, Chinese Academy of Science, National Laboratory of Pattern Recognition, China

Abstract: In this paper, we propose an algorithm for the semiautomatic segmentation of medical image series based on the combination of the live wire algorithm and the active contour model. We modified the traditional live wire algorithm by combining watershed transform and obtain accurate segmentation of one or more slices in a medical image series by the live wire algorithm. Based on the segmentation of previous slices, the computer will segment the nearby slice using the modified active contour model automatically. To make full use of the correlative information between contiguous slices, we introduce a gray-scale model to the boundary points of the active contour model to record the local region characters of the desired object in the segmented slice and replace the external energy of the traditional active contour model with the energy decided by the likelihood of the gray-scale model. Moreover we introduce the active region concept of the snake to improve the segmentation accuracy and reduce the complexity of dynamic program. Experiment shows that our algorithm can obtain the boundary of the desired object from a series of medical images quickly and reliably with only little user intervention.

 « Back

Wavelet-type expansions
Lasse Borup, Dept. of Mathematical Sciences, Aalborg University, Denmark

We consider an orthonormal basis for L2(R) consisting of functions that are well localized in the spatial domain and have compact support in the frequency domain. The construction is based on smooth local cosine bases and is inspired by F. Meyer and R. Coifman's brushlets, which are local exponentials in the frequency domain.

We will see that brushlet systems associated with an exponential type partition of the frequency axis, bear some resemblance to wavelet systems. E.g. such a system constitutes an unconditional basis for various function spaces - Lebesgue, Besov, Triebel-Lizorkin, Hardy, BMO - and the norm in these spaces can be expressed in terms of the expansion coefficients.

If instead, the brushlet system is associated with a partition of the frequency axis based on a polynomial rule with a parameter α (0<α≤1) we will see that the system forms an unconditional basis for the α-modulation spaces introduced by P. Gröbner.

 « Back

Gabor analysis, Rieffel induction and Feichtinger's algebra as a link
Franz Luef, NuHAG, Dept. of Mathematics, University of of Vienna

Download/View PDF

 

 « Back

Frames and flexibility
Ole Christensen, Dept. of Mathematics, Technical University of Denmark

The increased flexibility (compared to orthonormal bases) is often an argument for using frames. However, in most cases we also want our frames to have some structure, and there are cases where this additional constraint limits (or removes) the freedom. For this reason we seek to extend classical frame theory by allowing duals belonging to a different space than the frame.

Given a frame for a subspace W of a Hilbert space H, we characterize the set of oblique dual frame sequences (i.e., dual frame sequences that are not constrained to lie in W). We then consider frame sequences in shift invariant spaces, and characterize the translation invariant oblique dual frame sequences. For a given translation invariant frame sequence an easily verifiable condition on another shift-invariant frame sequence implies that its closed linear span contains a generator for a translation invariant dual of the frame sequence we start with; in particular, this result shows that classical frame theory does not provide any freedom if we want the dual to be translation invariant. In the case of frame sequences generated by B-splines we can use our approach to obtain dual generators of arbitrary regularity

 « Back

Camera calibration and pose determination from N points
Yihong Wu, National Laboratory of Pattern Recognition, Institute of Automation, Chinese Academy of Sciences

Perspective-N-Point camera calibration and pose determination, simply called PNP problem, has attracted lot of attention in the literature. PNP problem has different definitions, and different definition usually implies different set of solutions, different computational complexity, and different robustness.

When N<3, the PNP problem is under constrained, and When N>=6 the PNP problem can be linearly determined for the intrinsic parameters and pose parameters. When N=3,4,5, the constraints from N points are not enough to determine all the intrinsic parameters and pose parameters. So, usually, the literatures assume that the intrinsic parameters are known, then determine the camera pose; or determine the intrinsic and pose parameters from prior knowledge on the intrinsic parameters.

In this work, with known intrinsic parameters, we investigate the PNP problem for N=3,4,5 from both geometric and algebraic standpoints, and has the following contributions: Firstly, we formally proved that the PNP problem under distance-based definition is equivalent to the PnP problem under orthogonal-transformation-based definition when N>3, and is equivalent to the PnP problem under rotation-transformation-based definition when N=3. In other words, if the rotation matrix in perspective matrix is relaxed to an orthogonal matrix, then the distance-based definition and the transformation based definition become equivalent when N>3. Secondly, we introduce a depth-ratio based technique to solve PnP problem. This technique is shown to be advantageous over the traditional techniques of either directly determining the distances or transformation matrix. Thirdly, from the geometrical point of view, we obtained the upper bounds of the PnP problem under different definitions, and provided some insights into the fact that the different bounds are generally associated with different definitions. Fourthly, a construction of multiple solutions is presented for P3P problem. Lastly, the degenerated cases, i.e., the coplanar control points, or collinear control points, are discussed. Surprisingly enough, it is shown that if all the control points are collinear, the PnP problem under distance-based definition has a unique solution, but the PnP problem under transformation-based definition is determined only up to one free parameter.

 « Back

Content-based visual information retrieval
Yu Jin Zhang, Dept. of Electronic Engineering, Tsinghua University, China

While advances in compression algorithms have relaxed the storage and transmission requirements to some extent, the large volume of images and videos makes it difficult for a user to browse through the entire database. On the other side, when such a collection is available or accessible, the question asked by a user is often how a particular image or a clip of video is located. In the last ten years, researches on content-based image and video retrieval have made a number of progresses and provided some suitable ways to answer efficiently the above question. This talk will be based on a recent book “Content-based Visual Information Retrieval”. Using the information collected for writing this book as well as the research experiments and results that supply suitable materials for this book, a general overview in this research direction will be introduced, different techniques for querying and retrieving image as well as for organizing and searching video will be described. In addition, some existing problems and further research trends will be discussed.

 « Back

A Quasi-Wavelet algorithm for second kind boundary integral equations
Han-lin Chen, Institute of Mathematics, Academia Sinica and Silong Peng, National Asia Design Engineering Center, Chinese Academy of Science

In solving integral equations with a logarithmic kernel, we combine the Galerkin Approximation with periodic quasi-wavelet (PQW). We develop an algorithm for solving the integral equations with only O(N log N) arithmetic operations where N is the number of knots. We also proved that the Galerkin approximation has a polyno- mial rate of convergence.

 « Back

Spectral radii of refinement and subdivision operators
Victor Didenko, Faculty of Science, Universiti Brunei Darussalam

The spectral radii of refinement and subdivision operators can be estimated by using norms of their symbols. In several cases the exact value of the spectral radius is found.

 « Back

The construction of wavelets and its some applications to pattern recognition
Daren Huang, Sun Yat-sen University

In this paper, the construction of wavelets, especially multiband wavelets, and their corresponding properties are discussed. As wavelet applications, multiband wavelets are used to classify textures successfully; an algorithm for image recognition, such as face images, is developed based on nonlinear wavelets. Also, researches on the application of wavelets to similar numeral discrimination are presented.

 « Back

Radial basis, radial basis interpolation and meshless method
Zongmin Wu, Dept. of Mathematics, Fudan University

This talk will introduce the radial basis function, which is composed of linear combination of shifts of radial function Φ(||x||). Radial basis interpolation method and related topics is reviewed, especially the method, how to interpolate the linear functional data is introduced. The radial basis interpolations scheme for linear functional data will be applied to solve linear partial differential equation as a numerical meshless method. Existence, uniqueness as well as the convergence and related topics for the method are discussed.

 « Back

Convergence rates of subdivsion schemes associated with nonhomogeneous refinement equation
Song Li, Zhejiang University

Download/View PDF

 

 « Back

On the spectrums of frame multiresolution analyses
Hong Oh Kim, Division of Applied Mathematics, KAIST

We characterize the spectrum of the 'central' space of an FMRA and then give a new condition for an FMRA to admit a single wavelet solely in terms of the spectrum of an FMRA. This improves the results prevously obtained by Benedetto and Treiber and also by us. Our methods and results are applied to the problem of the 'containments' of FMRAs in MRAs. We first prove that any FMRA is always contained in an MRA, and then we characterize those MRAs that contain 'genuine' FMRAs in terms the unique masks of the MRAs and the spectrums of the central spaces of the FMRAs to be contained.

 « Back

Adaptive short time Fourier transform for Whale call classification
Zhenyong Lin, Temasek Laboratories

Short-time Fourier transform is a classic time-frequency analysis method for non-stationary signal processing and representation. However, since its time and frequency resolution is fixed, it is not adaptive to signals. In this report, we developed an adaptive short-time Fourier transform algorithm, and use it as an analysis tool to underwater whale call classification. Experiments show that our method can achieve excellent results.

 « Back

Dual Gramian analysis
Zuowei Shen, Dept. of Mathematics, National University of Singapore

The dual Gramian analysis has been used in the study of the frame properties of shift invariant systems. It allows us to decompose the frame operator into a collection of simpler operators (fibers) in the Fourier domain. Each one of the `fiber’ operators acts from a sequence space to the same sequence space and its matrix representation can be explicitly described in terms of the Fourier transforms of the generators of the shift invariant system. Application to the Gabor systems which are shift invariant shows the power of the analysis. However, it cannot be directly applied to many other systems such as the wavelet systems, which are not shift invariant. This leads us to consider the dual Gramian analysis in a more general setting, i.e. general shift invariant systems which include wavelet system as a special case. In this talk, I will give the development and the recent results on the dual Gramian analysis.

 « Back

A Panoramic View on Gabor Analysis (from linear algebra to pseudo-differential operators)
Hans G. Feichtinger, Dept. of Mathematics, University of Vienna, Austria

It is the purpose of this series of lectures (the duration and choice of specific topics will be flexible, depending on the audience) to point the relevance of Banach frames and Riesz projection bases in the context of Gabor analysis, and to provide background on the importance of certain Banach spaces of functions (called Wiener amalgam spaces) in the treatment of families of function spaces (so-called modulation spaces) described by the time-frequency behaviour of their elements. These spaces in turn are relevant for the description of the behaviour of various kinds of operators, in particular slowly time varying systems.

Rough outline:

A) From linear algebra to Gabor analysis over finite Abelian groups (generating systems, frames, pseudo-inverse and frames, Gram matrix and biorthogonal systems; specific properties of coherent systems; redundancy; algebraic properties of Gabor systems). Introductory part (avoiding questions of convergence and continuity)

B) Continuous transforms (in particular STFT), sampling, the role of Wiener amalgam spaces and their basic properties, modulation spaces defined by means of the STFT (1-2 hours)

C) Fundamental properties of Gabor-Weyl-Heisenberg systems (Janssen representation, fundamental identity, adjoint TF-lattice, Ron-Shen principle, etc.) and the role of modulation spaces in the description of these facts; in particular the usefulness of the Gelfand triple (S_0,L_2,S_0') (Feichtinger's algebra and its dual). (1 hour)

D) Description of operators using time-frequency concepts (e.g. the Kohn-Nirenberg symbol resp. the spreading function), indications on pseudo-differential operators (Heil, Groechenig) in the context of time-frequency analysis (1+ hour)

As time permits we may come also to a discussion of the following topics: E) Possible continuation (rather for discussions): irregular Gabor families, questions of excess and redundancy, changing the lattice constants, automatic continuity properties of the frame operator and much more

HINT: final talk on Gabor Multipliers will make use of many of the results described in this tutorial session.

A theory of Gabor multipliers
Hans G. Feichtinger, Dept. of Mathematics, University of Vienna, Austria

Gabor multipliers are linear operators which arise by pointwise multiplication of Gabor coefficients of a function/distribution by a certain sequence of real numbers (over the TF-lattice in use). We will emphasize questions of the following type:

  1. what are the mapping properties of such operators (e.g. in term of modulation spaces)
  2. what about the uniqueness of the representation of a Gabor multiplier
  3. approximation properties of general mappings by Gabor multipiers
  4. similarity of Gabor multipliers using similar lattices

 « Back

Gabor Deconvolution: the recovery of seismic reflectivity in an attenuating medium
Gary Margrave, Dept. of Geology and Geophysics, The University of Calgary

Seismic wave propagation is associated with a progressive loss of frequency content as the wave propagates. This means that the reflected waveforms recorded at the earth's surface form a seismogram with nonstationary spectral character. While Wiener deconvolution was designed to sharpen these reflections by estimating and inverting their waveform, its inherent stationarity means that it can only succeed in a limited and local sense. We present an extension of the Wiener process that is accomplished in the Gabor transform domain. Using a pseudodifferential operator model for the attenuation process, we show that Gabor deconvolution can estimate and invert a nonstationary waveform with surprising success. In essence, we use the Gabor transform to accomplish an approximate factorization of the pseudodifferential operator. The resulting reflectivity estimates have greater resolution than can be achieved with stationary processes. Tests on real seismic data have produced excellent images. In a controlled experiment, in which a seismic wavefield was measured simultaneously in a borehole and on the earth's surface, we find that the wavforms estimated in the borehole closely match those estimated by Gabor deconvolution on the surface data.

 « Back

The representation of linear operators by Gabor multipliers
Michael P. Lamoureux, Dept. Mathematics and Statistics, University of Calgary

For a given pair of analysis and synthesis window, we examine what class of linear operators can be represented as a Gabor multiplier, and their relationship to the symbol of a Gabor multiplier. The notion of a compatible window pair is introduced, which includes Gaussians, as well as other examples of smooth windows. A complete answer to the representation question is obtained for compatible windows. These results are used to justified numerical techniques in seismic imaging that use Gabor multipliers to represent non-stationary filters and wavefield extrapolators.

 « Back

Filter approximation by Wavelet filter banks
Silong Peng, National Asia Design Engineering Center, Chinese Academy of Science

In many applications, we need to construct signal adaptive filters, especially adaptive wavelet filters. It mainly leads to an optimal problem which is difficult to be solved. For real time image processing, it is better to have a fast algorithm. In this paper, we give a direct method to approximate a given filter with wavelet filter banks. Experiments show the excellent performance.

 « Back

Content based image data retrieval and classification using nonnegative encoding
Robert Plemmons, Dept. of Maths & Computer Science, Wake Forest University, USA

This study introduces novel clustering methodology for image and sensor data retrieval, analysis and classification. The methodology is based upon encoding the data using low rank nonnegative matrix factorization algorithms to preserve natural data nonnegativity and thus avoid subtractive basis vector and encoding interactions present in methodologies such as principal component analysis. Some existing nonnegative matrix factorization techniques are reviewed and some new ones are proposed. Numerical experiments are reported using iris image data in biometric identification systems, phase screen correlation data in atmospheric imaging applications, and spectral data for identification and classification of orbital space objects.

 « Back

General uncertainty principles for Hilbert spaces
Say Song Goh, Dept. of Mathematics, National University of Singapore

The classical Heisenberg uncertainty principle plays an important role in time-frequency analysis of signals. This inequality could be generalized and formulated under an abstract mathematical setting. In this talk, we shall present a general uncertainty inequality bounding the commutator of two linear operators acting on a Hilbert space, where no additional assumptions are made on the two operators. The Heisenberg uncertainty principle and the periodic uncertainty principle proposed by Breitenberger are special cases of this general inequality. We shall also discuss about a characterization of equality in the general inequality, and asymptotic equality in the absence of equality for specific examples. Another issue of interest that will be highlighted is the description of geometric regions formed by quantities appearing in the inequality when the two operators involved are self-adjoint. Finally, we shall extend the general uncertainty principle to handle the multivariate case and also provide results on equality and asymptotic equality. Frequent references to the settings of the Heisenberg uncertainty principle and the periodic uncertainty principle will be made throughout the talk as illustrations. This talk is based on collaborations with Charles A. Micchelli and Tim N. T. Goodman.

 « Back

Loop groups, conjugate quadrature filters and wavelet approximation
Wayne M. Lawton, Dept. of Mathematics, National University of Singapore

Orthonormal wavelet bases associated to a multiresolution analysis are constructed from sequences c whose Fourier transforms C correspond to matrices [C(z) zC(-z); -zC(-z) C(z)] that map the unit circle into the special unitary group. The set of such functions, under pointwise multiplication, forms an infinite dimensional loop group. Smooth wavelets arise from C that admit divisors having the form (1 + z) while compactly supported wavelets arise from C that are Laurent polynomials and this subgroup is dense by a classical approximation result that is based on Trotter’s formula. We extend this result to show that every wavelet can be approximated, in appropriate Sobelev Spaces, by a sequence of compactly supported wavelets. We obtain this existence by combining methods and results from the theory of loop groups with a classical result from alegrabic topology and an extension of a recent approximation theoretic result concerning Bezout identities.

 « Back

An integrated optical/digital approach for improved image restoration
Robert J. Plemmons, Wake Forest University

A novel and successful optical/digital approach for removing certain aberrations in optical imaging systems involves placing an optical mask between an image-recording device and an object to encode the wave front phase before the image is recorded, followed by digital image restoration to decode the phase. We have shown that when appropriately engineered, such an optical mask can also act as a form of preconditioner for certain types of image restoration algorithms. Moreover it can boost information in the signal before it is recorded well above the noise level, leveraging digital restorations of very high quality. In this talk, we 1) examine the influence that a phase mask has on the incoming signal and how it subsequently affects the performance of restoration algorithms, and 2) explore the design of optical masks, a difficult nonlinear optimization problem with multiple design parameters, for removing certain aberrations and for maximizing information content in recorded images. The talk involves joint work with Paul Pauca, Sudhakar Prasad, Todd Torgersen, and Joe van der Gracht, on Pupil Phase Engineering.

 « Back

Preconditioning regularized least squares problems arising from high-resolution image reconstruction from low-resolution frames
Furong Lin, Dept. of Mathematics, Shantou University, China

In this paper, we study the problem of reconstructing a high- resolution image from multiple undersampled, shifted, degraded frames with subpixel displacement errors from multisensors. Preconditioned conjugate gradient methods with cosine transform based precondition- ers and incomplete factorization based preconditioners are applied to solve this image reconstruction problem. Numerical examples are given to demonstrate the effciency of these preconditioners. We find that cosine transform based preconditioners are e®ective when the num- ber of shifted low-resolution frames are large, but are less effective when the number is small. However, incomplete factorization based preconditioners work quite well independent of the number of shifted low-resolution frames.

 « Back

Wavelet algorithms for high resolution image reconstruction
Zuowei Shen, Dept. Mathematics, National University of Singapore

High-resolution image reconstruction refers to the reconstruction of high-resolution images from multiple low-resolution, shifted, degraded samples of a true image. In this talk, we give a method based on the wavelet analysis. By expressing the true image as a function, we derive iterative algorithms which recover the function completely from the given low-resolution functions. These algorithms decompose the function obtained from the previous iteration into different frequency components in the wavelet transform domain via a redundant or irredundant system and add them into the new iterate to improve the approximation. We apply wavelet (packet) thresholding methods to denoise the function obtained in the previous step before adding it into the new iterate. Our numerical results show that the reconstructed images from our wavelet algorithms are better than that from the Tikhonov least-squares approach. Some other applications of this method are also discussed.

 « Back

Scalable multimedia communication over heterogeneous network
Fuchun Shang, Temasek Laboratories, Singapore

In heterogeneous network environment, the hardware capabilities of network devices to consume multimedia information are very different. For efficient utilization of system resource, it is important to transmit the just enough information that the receiving end can handle. Issues of scalable media compression, media stream organization and media quality negotiation and control in a multimedia communication system that can provide different media quality streams to different users at both codec and system level will be presented.

 « Back

A numerical method for maximum entropy deconvolution
Jian-Guo Liu, Dept. of Mathematics, University of Maryland

In this talk, I will discuss some numerical methods for image deconvolution based on maximum entropy. Numerical issues such as efficiency, reliability in designing of this kind of the numerical methods will discussed. Comparison with other kind deconvolution technologies will be presented by using examples in image deconvolution in the surface physics.

 « Back

Research of the centre for wavelets, approximation and information processing
Say Song Goh, Dept. of Mathematics, National University of Singapore

The Centre for Wavelets, Approximation and Information Processing of the National University of Singapore is a research centre of the Faculty of Science in the Department of Mathematics. It is a multidisciplinary research centre that emphasizes the synergy of mathematics, engineering and computer science. Its research activities are on mathematical signal, image and information processing, as well as their analysis and understanding. In this talk, we shall provide an overview of the centre and highlight its various areas of research. Some of the centre’s research capabilities include image, document image and video compression, video summarization, denoising and resolution enhancement, underwater signal processing and recognition, and signal identification and classification.

 « Back

Image restoration for historical documents using directional wavelet transform
Tao Xia, CWAIP, National University of Singapore

In this talk we will introduce a novel algorithm to clean up a large collection of historical handwritten documents. Due to the seepage of ink over long period of storage, the front page of each document has been severely marred by the reverse side writing. Earlier attempts have been made to match both sides of a page to identify the offending strokes to eliminate them with the aid of a wavelet transform. Perfect matching, however, is difficult due to document skews, differing resolutions, inadvertently missing out reverse side and warped pages during image capture. A new approach is proposed to avoid double side mapping by using a directional wavelet transform to distinguish the foreground and reverse side strokes much better than the conventional wavelet transform. Experiments have shown that the algorithm enhances the readability of each document significantly without the mapping with its reverse side.

 « Back

FOI/ROI video compression
Li Hui, CWAIP, National University of Singapore

In video compression techniques, frames-of-interest (FOI) and region-of-interest (ROI) could be addressed for more attention. Thus, high priority is given to FOI or ROI by allocating more bits than others. It can obtain higher subjective video quality at the same rate, by sacrificing the quality of the other parts. We developed new FOI/ROI algorithms for video compression. Simulation results show excellent performance.

 « Back

Multiwavelets, diamond search, and multi-scalable video codec platform
Jo-Yew Tham, Centre for Industrial Mathematics, National University of Singapore

The recent explosion in digital video storage and delivery has presented strong motivations for high performance video compression solutions. Coupling this with the growing heterogeneity of user and network requirements, there arises a pressing need for a flexible video compression framework that can support a “generate once, scale many” concept. This talk presents an overview of a highly scalable video compression platform, covering the design and application of good multiwavelet filters, a novel diamond search algorithm for fast block-based motion compensation, and a multi-scalable video codec architecture that supports simultaneous video scalability in terms of bit rate, distortion, frame rate, color, spatial resolution, and computational complexity within a single compressed video bit stream.

 « Back

Wavelet-Galerkin method and natural boundary integral equations
Wei Lin, Dept. of Mathematics, Zhongshan University

In this paper we apply the wavelet- Galerkin method to numerical solutions for Harmonic Equations, Biharmonic Equations, Stokes Equations, Elasticity Equations and Helmholtz Equations respectively. All of them are reduced to the natural boundary integral equations with the strong-singular kernel. Hermite wavelets, Shannon wavelets and spline wavelets are utilized.

 « Back

Fingerprint recognition technology
Jufu Feng, Center for Information Science, National Laboratory on Machine Perception, Peking University, China

In this talk, I will introduce the fingerprint history and automated fingerprint identification system(AFIS). Ancien people were aware of the individuality of fingerprint. From the late sixteenth centurty to the late ninteenth century, a lot of findings established the foundation of morden fingerprint recognition. In the early 1960s, automated fingerprint identification system(AFIS) was developed. Since then, AFIS has benn used worldwide.

 « Back

Wavelet decompositions and non-linear approximations
Dung Dinh, Information Technology Institute, Vietnam National University

In this talk, we will discuss the following questions:

  • Nonlinear approximations of functions by sets of finite cardinality and the classical entropy number measuring their optimality,
  • Nonlinear approximations of functions by sets of finite pseudo-dimension and the related nonlinear widths measuring their optimality, and
  • Constructions and use of wavelet decompositions for constructing asymptotically optimal methods for these nonlinear approximation characterizations

 « Back

Tiles, self-affine sets and scaling functions
Ka Sing Lau, Dept. of Mathematics, The Chinese University of Hong Kong

Scaling functions are used to construct wavelets. Such a function is gererated by the subdivsion scheme. The most studied case is for scaling 2 and with integral translations; the multivariate case is more complicated. In this talk we will consider the support of such functions which can be considered, more generally, as the invariant sets of iterated function systems with expanding integral dilation matrices and integral translations; these are self-affine sets. A special case is the self-affine tiles. We will discuss the geometry of such sets and the related properties.

 « Back

A wavelet set construction and tight frames
Songkiat Sumetkijakan, Chulalongkorn University, Thailand

Since wavelet sets were introduced in the first existence proof of multidimensional single dyadic orthonormal wavelets by Dai, Larson, and Speegle, many examples and a few general constructions have been given. An intuitive wavelet set construction is due to Benedetto and Leon. Starting from a set K_0 and a map T with some specified properties, it is an iterative construction of sets K_n that approach a wavelet set. Though the scope of this construction has some limitations, we extend it to a more general setting. In particular, the construction can produce a multidimensional Journ'e wavelet set, the wedding cake set, and a wavelet set with only two connected components. In addition, each finite iteration of the construction is proved to give rise to a tight frame wavelet set.

 « Back

Wavelet frames: theory and constructions
Zuowei Shen, Dept. Mathematics, National University of Singapore

This talk is about wavelet frames. The redundant representation offered by wavelet frames has already been put to good use for signal denoising, and is currently explored for image compression. We restrict our attention to wavelet frames constructed via multiresolution analysis, because this guarantees the existence of fast implementation algorithms. We shall explore the `power of redundancy' to establish general principles and specific algorithms for constructing framelets and tight framelets. In particular, we shall give several systematic constructions of spline and pseudo-spline tight frames and symmetric bi-frames with short supports and high approximation orders. Several explicit examples are discussed. This talk is based on joint work with Ingrid Daubechies, Bin Han and Amos Ron.

 « Back

Pointwise properties of bounded refinable function vectors
Di-Rong Chen, Dept. of Applied Mathematics, Beijing University of Aeronautics and Astronautics

Download/View PDF

 

 « Back

Eigenvalues and Biorthogonal Eigensystems of scaling operators
Seng Luan Lee, Dept. of Mathematics, National University of Singapore

Download/View PDF

 

 « Back

Approximation and interpolation by radial basis function on Sobolev space
Jungho Yoon, Dept. of mathematics, Ewha W. University, S. Korea

Approximation by translates of suitable radial basis functions is an important approach towards solving the (multivariate) scattered data problem. Its strengths are as follows:

  1. it facilitates the evaluation of the approximant;
  2. the accuracy of approximation is usually very satisfactory provided the approximand f is reasonably smooth;
  3. there is enough flexibility in the choice of basis functions.

In this talk, we discuss the approximation power of radial basis function interpolation and approximation on Sobolev space.

 « Back

An orthogonal splines approach to periodic wavelet frames
Say Song Goh, Dept. of Mathematics, National University of Singapore

Wavelet frames are of much current interest in wavelet research. They contain redundancy and could provide more flexibility in some applications. In this talk, we shall present an approach based on orthogonal splines for the study of frame multiresolution analyses and wavelet frames of periodic functions. Results on frames for shift-invariant subspaces of periodic functions will be obtained in terms of orthogonal splines. We shall also discuss about equivalent conditions for the existence of a wavelet frame starting from a frame multiresolution analysis, and an algorithmic construction when such a wavelet frame exists. Trigonometric polynomial frames will be used to illustrate the ideas presented.

 « Back

Multiwavelets associated with countable groups of unitary operators in Hilbert spaces
Wai Shing Tang, Dept. of Mathematics, National University of Singapore

We give a characterization of biorthogonality among Riesz multiwavelets in terms of certain invariant properties of their associated core spaces. A large family of non-biorthogonal Riesz multiwavelets is exhibited. We also discuss some results on linear perturbation of orthonormal multiwavelets. This is joint work with David Larson and Eric Weber.

 « Back

Singularity and Instantaneous Frequency Detection from Modulus Maxima of Continuous Wavelet Transforms
Wen-Liang Hwang, Institute of Information Sciences Academia Sinica

Many problems in signal processing, image processing, and pattern recognitions use wavelet transform to detection local features. Among them, the most important features may be singularity and instantaneous frequency. Singularity measures the local irregularity of a signal and instantaneous frequency measures the local dominant variations. Singularity can be detected from the modulus maxima of a real-valued wavelet, while the instantaneous frequency can be obtained from that of complex-valued wavelets. We proof that isolated singularity can be detected from modulus maxima of complex-valued wavelets. Thus, using complex-valued wavelet transform, one is about to simultaneously detection isolated singularity and instantaneous frequency from its modulus. We will also show applications in which shape are derived from texture variations measured by instantaneous frequencies in textured images.

 « Back

What's Math got to do with it ? Mathematics at the frontiers of sciences and technology
Tony Chan, Department of Mathematics, University of California, USA

Mathematics is probably the least understood (and appreciated) among the hard sciences by the general public. Its public image is probably a mixture of inscrutability, fear and irrelevance to everyday life. And this despite that most adults have been exposed to many years of mathematics during their education. In fact, mathematics is at the foundation of our highly technological society. It is at the core of much of the current information technology revolution and many predict that it will play an equally important role in the burgeoning bio-medical revolution ushered in by the advent of modern genetics.

So why is Mathematics having such a hard time getting the message across to the public? The cynics will say it is because of the abysmal and often traumatic experience many people experienced in mathematics classes in school. It is true that the rigor, the abstraction, and the scope of mathematics is quite challenging for many students. And much of frontier research mathematics is close to impossible for outsiders to understand. However, the APPLICATION of mathematics can be found in almost all walks of life, and often in the most unexpected places. My conviction is that if the public is more aware of the "usefulness" and "ubiquity" of mathematics, not only will the public image of mathematics improve, but it'll also provide the public with the necessary understanding to make informed political decisions regarding how society should invest in its technological future.

In this talk, I'll provide some examples of interesting applications of frontier research mathematics in areas that are quite close to everyday life. Examples include the movies, the stock market, the internet, medicine, communication, etc.

 « Back

Mathematics in the real world and the fake world
Stanley Osher, University of California, Los Angeles, USA

With the advent of the computer we can now develop algorithms that perform incredible tasks-special effects in Hollywood, catching bad guys on video, predicting all kinds of natural and unnatural phenomena. A common theme in these algorithms is rather elementary geometry. In this talk I'll discuss geometric algorithms and their applications in every day life.

 « Back

Impulse noise removal by median-type noise detector and edge-preserving regularization
Chung-Wa Ho, Department of Mathematics, The Chinese University of Hong Kong

This paper proposes a two-phase scheme for impulsive noise removal. In the first phase, an adaptive median filter is used to identify the noise candidates. In the second phase, the image is restored by an edge-preserving regularization method that applies only to the noise candidates. In terms of edge preserving and noise suppression, our restored images are significantly better than those restored by using just nonlinear filters or edge-preserving regularization only.

 « Back

The numerical algorithm for an inverse problem in optical mammography
Jerry Yin-Tzer Shih, Department of Medical Imaging Technology, Chung Shan Medical University

We discuss the implementation of a finite element algorithm with C1 Elements for solving a coupled system of generalized biharmonic type equations related to a certain inverse problem. We consider both direct and preconditioned conjugate gradient methods for rapid solution of the resulting rather large linear system and our development of an efficient preconditioner.

 « Back

Super-resolution image restoration from blurred low-resolution images
Andy Chin Ko Yau, Department of Mathematics, The Chinese University of Hong Kong

Download/View PDF

 

 « Back

Latest results in image superresolution
Nirmal Bose, Pennsylvania State University, USA

The wavelet superresolution algorithm proposed by Nguyen et al in 2000 ignored the effects of using different choices of mother wavelets and scaling functions. In their work, only the Daubechies wavelet and scaling function were applied and tested. Here, a detailed comparison of the desirable properties in superresolution of several finitely supported scaling function and wavelet families is given. Second, adaptation to nonsmooth domains, irregularly sampled data, and weighted measures is considered. Third, a comparison in performance, features, and constraints of several superresolution algorithms that have been popularized during the last decade is provided. Finally, problems to be overcome are summarized.

 « Back

Non smooth cost-functions in image and signal reconstruction
Mila Nikolova, Centre de Mathematiques et de Leurs Applications CMLA (CNRS UMR 8536) ENS de Cachan, France

We consdier the recovery of an image or a signal x*, based on data y, by minimizing a cost function of the form F(x,y)=D(x,y)+c R(x), where D is a data-fidelity term, R is a regularization term and c>0 is a parameter. Examples of applications are image restoration, segmentation, motion estimation, color reproduction, optical imaging, tomography, seismic and nuclear imaging. Classical ways to construct cost-functions F are based on calculus of variations or on probabilistic considerations. In the past, different heuristics have been adapted to specify cost-functions for different applications. Instead, we analyze the properties of minimizers x* as an implicit function of both the data y and the shape of the function F. In this talk, we focus on the smoothness of the cost function F. Points of interest are the recovery of homogeneous regions and edges, and the processing of outliers and impulse noise.

(a) Non-smooth regularization to recover strongly homogeneous regions

Regularization R accounts for priors on x such as smoothness and other properties. Usually R is the sum of terms of the form f(||g_i x||) where g_i are linear operators which yield the differences between neighboring samples in x and f is a potential function. We show that typically, the local minimizers x* of F(.,y) satisfy g_i x*=0 for numerous indexes i. So, if g_i yield first-order differences, x* involves constant regions. This explains the stair-casing effect observed in total variation methods. If g_i yield second-order differences, x* involves planar regions, etc. We call such regions strongly homogeneous. This is a striking property, both visually and mathematically. In comparison, such a behavior does not occur if R is smooth.

(b) Non-smooth data-fidelity to satisfy exactly a certin number of the data entries

Data-fidelity D is generally a smooth function. Non-smooth data-fidelity terms are avoided in image processing. We focus on functions D combining terms of the form d(a_i x-y_i) where a_i are linear operators modelling the data acquisition, and d is a function which is non-smooth at zero. We show that typically, the local minimizers x* of F(.,y) fit exactly a certain number of the data entries: there are many indexes i for which a_i x*=y_i. This is a strong mathematical property which produces a surprising visual effect. Based on it, we construct cost-functions allowing outliers and impulse noise to be detected and selectively smoothed.

All theoretical results are illustrated using various numerical experiments in the context of practical image and signal reconstruction problems.

 « Back

Analysis of half-quadratic minimization methods for signal and image recovery
Mila Nikolova, Centre de Mathematiques et de Leurs Applications CMLA (CNRS UMR 8536) ENS de Cachan, France

We address the minimization of regularized convex cost functions which are customarily used for edge-preserving restoration and reconstruction of signals and images. In order to accelerate computation, multiplicative and additive half-quadratic reformulation of the original cost-function have been pioneered in Geman & Reynolds (1992) and Geman & Yang (1995). The alternate minimization of the resultant (augmented) cost-functions has a simple explicit form. Our goal is to provide a systematic analysis of the convergence rate achieved by each one of these methods. We show that the number of iterations required for convergence is quite smaller for the multiplicative form than for the additive form. However, the computational cost of each iteration is much higher for the multiplicative form than for the additive form. The global assessment is that minimization using the additive form of half-quadratic regularization is faster than using the multiplicative form. When the additive form is applicable, it is hence recommended. Extensive experiments demonstrate that both methods are substantially faster than the standard Matlab routines. (joint work with Michael Ng, University of Hong Kong)

 « Back

Structured matrix computation and image recovery applications
Michael Ng, Dept. of Mathematics, University of Hong Kong

In this talk, we address some structured matrix computation issues related to image recovery applications:

  1. the multiplicative and the additive forms in the minimization of regularized convex cost functions which are customarily used for edge-preserving restoration and reconstruction of signals and images;
  2. the factorized banded inverse preconditioners for matrices with Toeplitz structure arising in nonlinear image restoration applications.
  3. the preconditioning regularized least squares problems arising from high-resolution image reconstruction from low-resolution frames.

 « Back

Data hiding - theory and applications
Pierre Moulin, Dept. of Electrical and Computer Engineering University of Illinois at Urbana-Champaign
Nasir Memon, Computer Science Dept., Polytechnic University, USA

In the first part of the tutorial we will look at applications of data hiding. Specifically, we look at what mechanisms could be used to securely realize these applications. We will examine the following four applications (50 minute sessions each):

  1. Ownership assertion: Basic framework. Invertibility attack and copy attack. Countering invertibility attack. Protocol for secure ownership assertion. Zero knowledge protocols.
  2. Authentication: Basic framework. Attacks. Key management problem. Distortion bounded authentication.
  3. Steganography: Review of techniques. Steganalysis techniques. Universal steganalysis. Notions of steganography capacity and security. Discussion on the future of steganography.
  4. Copy protection: Basic framework. Attacks. Review of data hiding based copy protection mechanisms in consumer appliances and discussion of their security.

In the second part of this tutorial we will present some recent approaches to modeling watermarking and data hiding as communication problems. We first overview the watermarking problem (low payload) and show how it can be formulated as a problem of shaping signal statistics. We further describe how several popular watermarking techniques fit into this model and describe their performance. We also review the data hiding problem (high payload) and reduce it to a channel coding problem with side information. The unifying theme is that watermarking algorithms can (should) be based on fundamentals of detection theory, estimation theory, game theory, and information theory in general. The second part will be organized into five 50 minute sessions as follows:

  1. Early Work: spread-spectrum techniques and LSB techniques
  2. Binning schemes; modern quantization-based codes: ideas and basic constructions
  3. Performance analysis for watermarking systems: error probabilities
  4. Data Hiding: relation to communication with side information, capacity analysis, random binning, application to images.
  5. Advanced topics: desynchronization attacks, sensitivity attacks, steganography

 « Back

Wavelet algorithms for high-resolution image reconstruction
Raymond Chan, Dept. of Mathematics, Chinese University of Hong Kong

High-resolution image reconstruction refers to the reconstruction of high-resolution images from multiple low-resolution, shifted, degraded samples of a true image. In this talk, we analyze this problem from the wavelet point of view. By expressing the true image as a function in L2( R2), we derive iterative algorithms which recover the function completely in the L2 sense from the given low-resolution functions. These algorithms decompose the function obtained from the previous iteration into different frequency components in the wavelet transform domain and add them into the new iterate to improve the approximation. We apply wavelet (packet) thresholding methods to denoise the function obtained in the previous step before adding it into the new iterate. Our numerical results show that the reconstructed images from our wavelet algorithms are better than that from the Tikhonov least-squares approach.

 « Back

Algorithms for association rules
Markus Hegland Mathematical Sciences Institute, Australian National University

Association rules are "if-then rules" with two measures which quantify the support and confidence of the rule for a given data set. Having their origin in market basked analysis, association rules are now one of the most popular tools in data mining. This popularity is to a large part due to the availability of efficient algorithms following from the development of the Apriori algorithm.

In these lectures we will introduce basic concepts of association rule discovery including support, confidence, interestingness, the apriori property, and hierarchical and quantitative association rules. The core of the lectures consists of a review of the most important algorithms for association rule discovery. Some familiarity with concepts like predicates, probability, expectation and random variables is assumed.

 « Back

Domain decomposition methods for image processing
Shiu-Hong Liu, Dept. of Mathematics, University of Manitoba, Canada

We investigate the application of domain decomposition methods for nonlinear parabolic PDEs arising from a combination of schemes to denoise digital images. We initially apply the mean curvature flow equation to annihilate salt and pepper noise. Then we apply a Perona--Malik type scheme to remove additive Gaussian noise. At each time step, we employ the Schwarz alternating method to solve a linear system of equations. Some numerical examples illustrate the efficiency of the method.

 « Back

Wavelet adaptation for hybrid wavelet-large margin classifiers
Julia Neumann, Dept. of Mathematics & Computer Science, University of Mannheim, Germany

Hybrid wavelet-large margin classifiers have recently proven to solve difficult signal classification problems in cases where merely using a large margin classifier like, e.g., the Support Vector Machine may fail. In our case, we investigate the classification of electrophysiological signals. The features for our hybrid classifier are selected from the outputs of all orthonormal filter banks of fixed length with respect to criteria measuring class separability and generalisation error.

We evaluate a range of such adaptation criteria to perform feature selection for hybrid wavelet – large margin classifiers. The two main points we focus on are (i) approximation of the radius-margin error bound as the ultimate criterion for the target classifier, and (ii) computational costs of the approximating criterion for feature selection relative to those for the classifier design.

We show that by virtue of the adaptivity of the filter bank, criteria which are more efficient than computing the radius-margin are sufficient for wavelet adaptation and, hence, feature selection.

Another important issue for the extraction of signal features based on wavelet filtering is translation invariance. In order to provide a library of decimated shift insensitive perfect reconstruction wavelet transforms, we propose the use of complex wavelets defined in the frequency domain. These enable us to select shift insensitive wavelet features for the classifi- cation problem at hand.

 « Back

Edge and ramp preserving noise removal based on nonlinear diffusion with Gauss curvature conductance
Suk-Ho Lee, Dept. of Mathematics, Yonsei University, Korea

In this paper we propose the gauss curvature as the conductance of the diffusion equation for noise removal. The motivation for using the gauss curvature is that it goes to zero if any one of the principle curvatures goes to zero. Thus it has the property of preserving curved edges, ramps, features with small gradient magnitude and other small-scaled features, which other noise removal scheme such as the total variation based noise removal, mean curvature evolution, curvature based evolution fail to preserve. The properties and performance of the proposed scheme will be given theoretically and experimentally.

 « Back

Introduction to digital watermarking
Ee-Chien Chang, Dept. of Computer Science, National University of Singapore
Mohan Kankanhalli, Dept. of Computer Science, National University of Singapore

This tutorial provides the background information necessary to understand the applications and problems of watermarking and data-hiding.

 « Back

Content-dependent filter (CDF) design and analysis: CDF design and analysis based on the anisotropic diffusion model
Charles Chui, Stanford University and University of Missouri - St. Louis, USA

Several limitations of numerical solution of the anisotropic diffusion (AD) PDE as well as the bilateral filter will be discussed. To address these limitations, the notion of "content-dependent filtering" (CDF) is introduced in the first presentation. Both design criteria and theoretical analysis of these innovative filters will be addressed. The second presentation will focus on technical details. It will be clear that with appropriate modifications, CDFs also apply to image enhancement and image segmentation.

 « Back

Content-dependent filter (CDF) design and analysis: Design of CDFs for image noise reduction
Jianzhong Wang, Mathematics, Computer Science and Statistics, Sam Houston State University

Several limitations of numerical solution of the anisotropic diffusion (AD) PDE as well as the bilateral filter will be discussed. To address these limitations, the notion of "content-dependent filtering" (CDF) is introduced in the first presentation. Both design criteria and theoretical analysis of these innovative filters will be addressed. The second presentation will focus on technical details. It will be clear that with appropriate modifications, CDFs also apply to image enhancement and image segmentation.

 « Back

Fast fourth-order time-stepping for stiff PDE
Nick Trefethen, Oxford University

Many partial differential equations combine higher-order linear terms with lower-order nonlinear terms. Examples include the KdV, Allen-Cahn, Burgers, and Kuramoto-Sivashinsky equations. High accuracy is typically needed because the solutions may be highly sensitive to small perturbations. For simulations in simple geometries, spectral discretization in space is excellent, but what about the time discretization? Too often, second-order methods are used because higher order seems impractical. In fact, fourth-order methods are entirely practical for such problems, and we present a comparison of the competing methods of linearly implicit schemes, split step schemes, integrating factors, "sliders", and ETD or exponential time differencing. In joint work with A-K Kassam we have found that a fourth-order Runge-Kutta scheme known as ETDRK4, developed by Cox and Matthews, performs impressively if its coefficients are stably computed by means of contour integrals in the complex plane. Online examples show that accurate solutions of challenging nonlinear PDE can be computed by a 30-line Matlab code in less than a second of laptop time. For reaction-diffusion equation in 2D or 3D, we can often beat standard finite-difference methods by a factor of 10 or more in speed.

 « Back

Sparse grids and the analysis of high dimensional large scale data
Markus Hegland, Mathematical Sciences Institute, Australian National University

The search for interesting variable stars, the discovery of relations between geomorphological properties, satellite observations and mineral concentrations, and the analysis of biological networks all require the solution of a large number of complex learning problems with large amounts of data. A major computational challenge faced in these investigations is posed by the curse of dimensionality. A well known aspect of this curse is the exponential dependence of the size of regular grids on the dimension of the domain. This makes traditional finite element approaches infeasible for high-dimensional domains. It is less known that this curse also affects computations of radial basis function approximations -- in a slightly more subtle way.

Sparse grid functions can deal with the major problems of the curse of dimensionality. As they are the superposition of traditional finite element spaces, many well-known algorithms can be generalised to the sparse grid context. Sparse grids have been successfully used to solve partial differential equations in the past and, more recently, have been shown to be competitive for learning problems as well. The talk will provide a general introduction to the major properties of sparse grids and will discuss connections with kernel based methods and parallel learning algorithms. It will conclude with a short review over some recent work on algorithms based on the combination technique.

References:

  • Additive Sparse Grid Fitting, M. Hegland, Curve and Surface Fitting: Saint-Malo 2002, pp. 209-219,Nashboro Press, Brentwood, 2003.
  • Adaptive sparse grids, M. Hegland, Proceedings of 10th Computational Techniques and Applications Conference CTAC-2001, ANZIAM Journal, vol. 44, C335--C353, Available online.

(for other references see http://datamining.anu.edu.au)

 « Back

Relations between nonlinear denoising methods in signal and image processing
Gabriele Steidl, Fakultaet fuer Mathematik und Informatik, Universitaet Mannheim

Wavelet shrinkage, nonlinear diffusion filtering, and nonquadratic regularization methods are three useful techniques for edge-preserving denoising of signals and images.

In the first part of the talk we analyze relations between the important 1-D prototypes soft Haar wavelet shrinkage, space-discrete total variation (TV) diffusion and discrete total variation (TV) regularization: we prove that space-discrete TV diffusion and TV regularization are identical. We show that applying translationally invariant soft Haar wavelet shrinkage on a single scale can be regarded as an absolutely stable explicit discretisation of TV diffusion. Afterwards we show that wavelet shrinkage on multiple scales can be regarded as a single step diffusion filtering of the Laplacian pyramid of the signal.

In the second part of the talk we study general relations between arbitrary shrinkage functions and diffusivities: we show that one step of a (stabilized) explicit discretisation of nonlinear diffusion can be expressed in terms of wavelet shrinkage on a single spatial level. This equivalence allows a fruitful exchange of ideas between the two fields. For example. we derive new wavelet shrinkage functions from existing diffusivitv functions, and identify some previously used shrinkage functions as corresponding to well known diffusivities. Moreover, by transferring stability notions from diffusion filtering to wavelet shrinkage, we derive conditions on the shrinkage function that ensure that shift invariant single-level Haar wavelet shrinkage is maximum-minimum stable. monotonicity preserving, and variation diminishing.

Finally, we present some novel diffusion-inspired wavelet shrinkage methods with improved rotation invariance in 2D.

 « Back

Underwater acoustic signal classification
Zhenyong Lin, Temasek Laboratories

we applied a fast adaptive short-time Fourier transform algorithm for underwater whale song signal representation, and extract first and second order statistics from its time-frequency representation as features for underwater Humped Back whale call classification. After the optimisation of parameters for classification, we have achieve excellent results.

 « Back

Bayesian resolution enhancement of compressed video
Aggelos Katsaggelos, Electrical and Computer Engineering, Northwestern University, USA

Super-resolution algorithms recover high-frequency information from a sequence of low-resolution observations. In this presentation we consider some recent results on the impact of video compression on the super-resolution task. Hybrid motion-compensation and transform coding schemes are the focus, as these methods provide observations of the underlying displacement values as well as a variable noise process. We utilize the Bayesian framework to incorporate this information and fuse the super-resolution and post-processing problems. A tractable solution is defined, and the relationship between algorithm parameters and information in the compressed bit-stream is established. The association between resolution recovery and compression ratio is also explored. Simulations illustrate the performance of the procedure with both synthetic and non-synthetic sequences.

 « Back

Challenges and new solutions predicting Tsunami waves
M. M. Lavrentiev, Jr., Sobolev Institute for Mathematics SD RAS, Novosibirsk, Russia

In this paper the problem of tsunamigenic earthquake is concerned. Perhaps, only on the basis of multimethod approach the reasonable forecast of tsunami danger will be available. Among the possible methods to be used we list determination of initial displacement at source solving the corresponding inverse problem (A. Avdeev et. al., 2001, F. Gonzalez et. al, 2003), neural network analysis of the time series.

Pacific Marine Environmental Laboratory (PMEL) has developed the Deep-ocean Assessment and Reporting Tsunami (DART) system and has deployed it at six locations around the Pacific. DART systems obtain high-quality data of tsunami amplitudes in the open ocean. This data can be used to reconstruct the true tsunami source parameters. To achieve that goal, PMEL has created a database of pre-calculated time series of tsunami waves from (unit) sources at subduction zones around Pacific. The solutions from the database can be combined using developed Web interface (called FACTS – Facility for the Analysis and Comparison of Tsunami Simulation), to reconstruct tsunami sources recorded by DART. This approach will be used in the next generation of the real-time tsunami forecasting system, which is under development by the National Tsunami Hazard Mitigation Program (NTHMP).

In order to reconstruct tsunami sources with in automatic mode a special algorithm and a software application has been developed. This module is designed to determine the tsunami source parameters by processing the marigrams, obtained at the DARTS stations. The module effectively determines the amplification coefficients for unit sources, which makes possible to approximate the shape of vertical displacement of the sea surface at the tsunami source area. The algorithm is based on minimization of the calculated difference between the measured marigram(s) and the linear combination of synthetic marigrams, taken from the FACTS database.

This inversion algorithm has been numerically tested with the historical data for the 1996 Andreanov tsunami. The module has been implemented into the FACTS database.

In order to improve the quality of source parameters determination alternative methods have been studied. First, the Inverse Problems approach has been realized. After 1D numerical tests, the cost effective 2D algorithm has been developed and preliminary tested against synthetic and historical data. Finally, the neural network based algorithm has been designed.

Reference

  • Avdeev A. and et. all. Complex analysis of ocean tsunami observation data for solution of the inverse problem // Proc. of the Intern. Tsun. Symp. - USA, Seattle, 2001. - P. 795-808.
  • Gonzalez F.I., Titov V.V., Avdeev A.V., Bezhaev A.Yu., Lavrentiev M.M.(jr), Marchuk An.G. Real-Time Tsunami Forecasting: Challenges and Solutions. Proc. International Conference Mathematical Methods in Geophysics-2003, Novosibirsk, ICM&MG Publisher, 2003, pp. 225-228.

 « Back

High order numerical methods for time dependent Hamilton-Jacobi equations
Chi-Wang Shu, Division of Applied Mathematics, Brown University, USA

In these four lectures we will review high order accurate numerical methods for solving time dependent Hamilton-Jacobi equations. We will start with a brief introduction of the Hamilton-Jacobi equations, the appearence of singularities as discontinuities in the derivatives of their solutions hence the necessity to introduce the concept of viscosity solutions, and first order monotone type numerical schemes on structured and unstructured meshes to approximate such viscosity solutions, which can be proven convergent with error estimates. We then move on to discuss high order accurate methods which are based on the first order monotone schemes as building blocks. We will describe the Essentially Non-Oscillatory (ENO) and Weighted Essentially Non-Oscillatory (WENO) schemes for structured meshes, and the discontinuous Galerkin (DG) and WENO schemes for unstructured meshes.

Lecture One: Introduction to Hamilton-Jacobi Equations, Viscosity Solutions and First Order Monotone Schemes on Structured Meshes

Lecture Two: First Order Monotone Schemes on Unstructured meshes. High Order ENO Schemes on Structured Meshes. High Order TVD Time Discretization

Lecture Three: High Order WENO Schemes on Structured Meshes. High Order WENO Schemes on Unstructured Meshes

Lecture Four: Introduction to Discontinuous Galerkin Method. High Order Discontinuous Galerkin Method on Unstructured Meshes

 « Back

High order discontinuous Galerkin finite element method for solving time dependent Hamilton-Jacobi equations
Chi-Wang Shu, Division of Applied Mathematics, Brown University, USA

In this talk we first describe the joint work with Changqing Hu on designing high order Runge-Kutta discontinuous Galerkin method for solving time dependent Hamilton-Jacobi (HJ) equations. This method is based on the Runge-Kutta discontinuous Galerkin finite element method for solving conservation laws. The method has the flexibility of treating complicated geometry by using arbitrary triangulation, can achieve high order accuracy with a local, compact stencil, and are suited for efficient parallel implementation. The finite element solution approximates the solution of the HJ equation as piecewise polynomials, however the evolution of these piecewise polynomials follows the discontinuous Galerkin framework of the conservation law systems that the gradient of the solution to the HJ equation satisfies. In the original method of Hu and Shu, a least square procedure is performed to obtain a single polynomial approximating the solution of the HJ equation after two polynomials to its gradients are obtained. Recently, in a joint work with Fengyan Li, we utilize a smaller locally divergence-free discontinuous Galerkin space to bypass the need of this least-square procedure, which reduces both the computational cost and the storage requirement. One and two dimensional numerical examples are given to illustrate the capability of the method.

 « Back

Unitary extension principle and applications
Zuowei Shen, Dept. of Mathematics, National University of Singapore

The unitary extension principle makes the construction of tight frame wavelet painless. This gives a great flexibility to design tight frame wavelet filters according to the problems in hands. In this talk, I will briefly review the unitary extension principle and illustrate the power of this principle by showing how it is used to design filters for various high resolution image reconstruction problem.

 « Back

Impulse noise removal by median-type noise detector and edge-preserving regularization
Raymond Chan, Dept. of Mathematics, Chinese University of Hong Kong

This paper proposes a two-phase scheme for impulse noise removal. In the first phase, an adaptive median filter is used to identify pixels which are likely to be contaminated by noise (noise candidates). In the second phase, the image is restored using a specialized regularization method that applies only to those selected noise candidates. In terms of edge preservation and noise suppression, our restored images show a significant improvement compared to those restored by using just nonlinear filters or regularization methods only.

Joint work with Chung-Wa Ho, Chen Hu, and Mila Nikolova.

 « Back

Tutorial on wavelets and PDE-based image processing
Tony Chan, Dept. of Mathematics, University of California, Los Angeles and Jianhong (Jackie) Shen, School of Mathematics, University of Minnesota

Lecture One: A Friendly Tour of Mathematical Image & Vision Analysis (by Jianhong Shen)

This tour explains the vast background, important applications, essential ingredients, and profound connections of Mathematical Image & Vision Analysis in contemporary sciences and technologies. Highlighted are mathematical views of machine and human vision, and various modern mathematical tools in image analysis and processing, including regularization and inverse problems, variational optimization, and geometric and nonlinear PDEs. More details are to come in Tony Chan's tutorials on Thursday.

Lecture Two: Wavelets: Motivation, Construction, and Application (by Jianhong Shen)

Complex information and patterns are built upon simple basic elements, just as the universe is upon basic particles (e.g., quarks and atoms). Wavelets are such a class of building blocks. Across time, space, and scales, like systems of well organized sensors, wavelets efficiently capture and encode multiscale temporal and spatial information and patterns. This tutorial introduces the motivation, construction, implementation, as well as main applications of wavelets in image and signal analysis and processing.

Lecture Three: Total variation based image restoration and inpainting (by Tony Chan)

Total variation based methods have become one of the most popular and powerful PDE-based image processing methodology. Its attractiveness stems from its ability to locate and preserve discontinuities ("edges"). There is also a fundamental connection to a geometric view of images. In this tutorial, we will give an introduction to this methodology, from basic mathematical properties to efficient computational algorithms. We will also give some applications to image denoising and image inpainting.

Lecture Four: Variational image segmentation using level sets (by Tony Chan)

Image segmentation is a fundamental task in image processing and there have been numerous algorithms proposed for it. Over the last 15 years or so, a powerful class of methods have been proposed which is based on a variational formulation in which the desired segmentation is defined as a local minimizer of an energy functional which is constructed to measure the balance between fidelity to the image and a regular (smooth) representation. The most famous of these is the Mumford-Shah model. In the implementation of these models, a representation of the segments or their boundaries must be chosen so as to allow a robust and efficient algorithm. The level set methodology has proven to be a very powerful and versatile tool for this task. Its attractiveness stems from its ability to allow automatic merging and pinching-off of segments during the computation, as well as the ability to capture the segment boundaries on a uniform grid instead of tracking them explicitly. In this tutorial, we will give an introduction to this class of methods.

 « Back

Periodic wavelet frames for processing band-limited signals
Say Song Goh, Dept. of Mathematics, National University of Singapore

Band-limited signals are of interest in various applications. In this talk, we shall discuss about the use of periodic wavelet frames in processing such signals. Results on frame multiresolution analyses and wavelet frames of periodic functions, which provide the theoretical foundation of the approach, will be presented. We shall also see how the approach is realized through its corresponding frequency-based decomposition and reconstruction algorithms. As an illustration, experimental results on using trigonometric polynomial frames to separate shipping noise from snapping shrimp noise and pink noise will be shown.

 « Back

Computational challenges in electron microscopy of macromolecules
Esmond G. Ng, Lawrence Berkeley National Laboratory

Three-dimensional (3-D) density maps of biological molecules can be determined by merging a large number of two-dimensional (2-D) images taken by a low-dose electron microscope. The 2-D images are projections of a collection of randomly oriented 3-D molecules embedded in a thin layer of vitreous ice.

In order to recover the 3-D structure, one must first determine the relative orientations of projection images. Once the orientation parameters are obtained, the 3-D reconstruction can be carried out in the spatial domain by solving a large-scale linear least squares (LLS) problem iteratively. Because the angular coverage by 2-D projection images is limited, and because the projections are typically noisy, the LLS problem is ill-posed.

The noisy nature of the 2-D projections makes it difficult to determine orientation parameters accurately in advance. As a result, an iterative procedure that simultaneously refines the orientation parameters and the 3-D structure of the molecule is often used to render the final structure of macromolecule.

In this talk, we will give an overview on the mathematical formulation of the image reconstruction problem, and discuss numerical techniques that can be employed to solve this problem. In particular, we will point out the important roles numerical linear algebra play in producing a high resolution 3-D images of biological macromolecules.

 « Back

Weberized Mumford-Shah model with Bose-Einstein photon noise: Light adapted segmentation inspired by vision psychology, retinal physiology, and quantum statistics
Jianhong (Jackie) Shen, School of Mathematics, University of Minnesota

Human vision works equally well in a large dynamic range of light intensities, from only a few photons to typical midday sunlight. Contributing to such remarkable flexibility is a famous law in perceptual (both visual and aural) psychology and psychophysics known as Weber's Law. There has been a great deal of efforts in mathematical biology to model and interpret the law in the cellular level as well as in terms of linear (with feedbacks) and nonlinear biological systems (for the two retinas). In terms of image and vision analysis, it is the current author who has first emphasized the significance of the law in faithfully modelling both human and computer vision, and attempted to integrate it into important visual processors such as image denoising (Physica D, 175, pp. 241-251, 2003).

The current work develops and studies a new segmentation model based on the integration of both Weber's Law and the celebrated Mumford-Shah segmentation model (Comm. Pure Applied Math., 42, pp. 577-685, 1989). Explained in details are issues like why the classical Mumford-Shah model lacks light adaptivity, and why its "weberized" version proposed here more faithfully reflects human vision's superior capability in segmenting retinal images in a variety of lighting environments from dawn to dusk. It is also claimed that the popular Gaussian noise model is inappropriate for the weberized Mumford-Shah prior, and instead, the intrinsic thermal noise of photon ensembles is introduced based on Bose-Einstein's distributions in Quantum Statistics, which turns out to be compatible with the weberization idea both analytically and computationally.

The current work also focuses on both the theory and computation of the weberized Mumford-Shah model with Bose-Einstein noise. In particular, Ambrosio-Tortorelli's Γ-convergence approximation theory is adapted (Boll. Un. Mat. Ital., 6-B, pp. 105-123,1992), and its stable numerical algorithms are developed for the associated pair of nonlinear Euler-Lagrange PDEs. Comparison with the classical Mumford-Shah model is made through typical numerical examples.

This is a joint work freshly completed with my Ph.D. student Joon-Mo Jung.

 « Back

Fast summation at nonequispaced knots by NFFTs
Daniel Potts, Institut fur Mathematik, Medizinische Universit at Lubeck, Germany

Download/View PDF

 

 « Back

A fast sweeping method with application in image processing
Hongkai Zhao, Dept. of Mathematics, University of California

In this talk I will present a fast sweeping method for computing the numerical solutions for a class of Hamilton Jacobi equations. The method is an iterative method which uses upwind difference for discretization and uses Gauss-Seidel iterations with alternating sweeping ordering to solve the the discretized system. The method is extremely simple to implement in any number of dimensions. Convergence and error estimates of the algorithm will be discussed. Applications to finding distance on manifolds and image reconstruction will be presented.

 « Back

Using geometry and iterated refinement for inverse problems: Total variation image restoration
Stanley Osher, University of California, Los Angeles, USA

Total Variation based regularization and image restoration was developed by Rudin-Osher-Fatemi in the late 80's. Recently, Yves Meyer characterized textures as elements of the dual of BV and did some extremely interesting analysis on the original ROF model. This led to practical algorithms to decompose images into structure plus texture. Very promising results involving processing image gradients simultaneously with images were obtained by Lysaker-Osher-Tai, based on earlier work on processing surfaces by Tasdizen-Whitaker-Burchard-Osher. This has now led to a new way of refining and enhancing the solutions to a wide class of inverse problems. I will discuss all this and present image restoration results which appear to be state-of-the-art.

Joint work with Jinjun Xu, Wotao Yin and Donald Goldfarb

 « Back

A fast and adaptive surface reconstruction algorithm based on implicit B-spline surfaces
Falai Chen, Dept. of Mathematics, University of Science and Technology of China

Based on implicit B-spline surfaces, we present a fast and adaptive algorithm to reconstruct surfaces from point-clouds. Our algorithm is driven by a surface fitting model, which amounts to solve some qudratic optimization problem. We explore the matrix form of the surface fitting model, and put forward a fast and low memory consumption algorithm to solve the optimization problem. In addition, for non-uniform sampling point sets, methods to adaptively choose the knot vectors of the implicit B-spline surface are presented. Some examples are provided to illustrate effectiveness of the algorithm.

 « Back

Improved splitting-integrating methods for image geometric transformations
Zi-Cai Li, Dept. of Applied Mathematics, National Sun Yat-Sen University

Download/View PDF

 

 « Back

Logic operators for active contours on multi-channel images
Tony Chan, Dept. of Mathematics, University of California, Los Angeles

We present a mathematical framework for active contour object detection in multi-channel images. Our framework allows the user to define arbitrary logic operations to combine object information from the different channels. Specific models are derived based on the single-channel region based "Active contour without edges" models. Numerical examples will be presented to show that the models can robustly detect objects in both synthetic and realistic images defined by general logical combinations.

 « Back

A hierarchical Hough transform for fingerprint matching
Tao Xia, CWAIP, National University of Singapore

This talk addresses the improvement on the generalized Hough transform for the biometric identification applications. The errors in generalized Hough transform for fingerprint matching are investigated and a new hierarchical Hough transform algorithm is proposed, which is faster and more accurate compared to conventional generalized Hough transform.

 « Back

LBIC imaging of semiconductor arrays
Weifu Fang, Dept. of Mathematics, West Virginia University, USA

Laser beam induced current (LBIC) is a non-destructive technique that has been used for a number of years to qualitatively examine large arrays of p-n junctions, especially in HgCdTe infrared focal plane arrays. In this talk, we will present PDE models for the LBIC imaging technique, and employ homogenization method to derive approximations of the LBIC images of large arrays. Such approximations reduce the computational burden in simulations of these LBIC images. We will then illustrate the application of our approximations for the purpose of recovering array parameters from the LBIC images.

 « Back

Improving learning machines with optimization strategies
Juan Manuel Górriz Sáez Escuela Politécnica Superior de Algeciras, Spain

In this work we establish new bounds on for the actual risk functional of a new on-line parametric machine for time series forecasting based on Vapnik-Chervonenkis (VC) theory. Using the strong connection between support vector machines (SVM) and Regularization theory (RT), we propose a regularization operator in order to obtain a suitable expansion of radial basis functions (RBFs) with the corresponding expressions for updating neural parameters. This operator seeks for the "flattest" function in a feature space, minimizing the risk functional and controlling the capacity of the learning machine. The performance of the learning machine can be improved using processing techniques such as Principal Components Analysis or Independent Co mponent Analysis. Finally we mention some modifications and extensions that can be applied to control neural resources (complexity control) and select relevant input space (suitable expansion) to avoid high computational effort i.e solving convex optimization problem.

 « Back

New idea of segmentation method by using Riemannian geometry
Shen Chun-li, Dept. of Mathematics, East China Normal University, China

In the image processing, there are many segmentation methods which can take out the useful contour from the image, such as the well-known geometric active contour method, geodesic active contour method, or various isotropic diffusion method and anisotropic diffusion method. Now we have proposed a new viewpoint which could deduce the above methods by a unified idea from Riemannian geometry, and those diffusion coefficients or diffusion tensor in the above methods could be obtained intrinsically by the geometrical properties of Riemannian geometry.

 « Back

Unitary matrix functions, wavelet algorithms, and structural properties of wavelets
Palle Jorgensen, University of Iowa, USA

Some connections between operator theory and wavelet analysis will be presented in three lectures. Since the mid eighties, it has become clear that key tools in wavelet analysis rely crucially on operator theory. While isolated variations of wavelets, and wavelet constructions had previously been known, since Haar in 1910, it was the advent of multiresolutions, and sub-band filtering techniques which provided the tools for our ability to now easily create efficient algorithms, ready for a rich variety of applications to practical tasks. Part of the underpinning for this development in wavelet analysis is operator theory. This will be presented in the lectures, and we will also point to a number of developments in operator theory which in turn derive from wavelet problems, but which are of independent interest in mathematics. Some of the material will build on chapters in a new wavelet book, co-authored by the speaker and Ola Bratteli, see http://www.math.uiowa.edu/~jorgen/.

  • Part I. Scaling identities in wavelet analysis, and systems of operators in Hilbert space

    Aspects of wavelet theory may be formulated in terms of operator Systems, scaling operator, the translations, and the transition operator. We outline this approach using Hilbert space geometry, and we give a number of examples. We show how the pyramid algorithms may be given a geometric formulation in this framework, and we recall the fundamentals of wavelet algorithms and of wavelet packet constructions.
     
  • Part II. Wavelet algorithms computed with unitary matrix functions (also called poly-phase matrices)

    Frequency sub-band filters may be assembled into a single function, but it is then operator valued. If the number of masking coefficients is finite, it takes values in the unitary matrices, where the size of the matrix depends on the number of masking coefficients. We show how factorization theory for these matrix functions may be used in the analysis and the classification of wavelets. This is joint work with Ola Bratteli.
     
  • Part III. Some special wavelets with general multiplicity configurations

    We recall some fundamentals from the theory of translation invariant subspaces, as well as the notion of multiply generated resolution subspaces. We outline results from two recent joint papers with L.Baggett, K. Merrill, and J. Packer. Multiplicity theory is needed for the study of frequency-localized wavelets, including wavelet sets. Our formulation allows new wavelet construction; for example, we sketch wavelet constructions in Hilbert spaces other than L2(Rd). A number of examples are given, including wavelet constructions for fractals; the fractal theory is joint work with Dorin Dutkay.

References:

[1]  math.CA/0407330 Title: Martingales, endomorphisms, and covariant systems of operators in Hilbert space
       Authors: Dorin Ervin Dutkay, Palle E.T. Jorgensen
[2]  math.CA/0403117 Lecture Notes for the Singapore workshop. Title: Unitary matrix functions, wavelet
       algorithms, and structural properties of wavelets Authors: Palle E. T. Jorgensen
[3]  math.CA/0305443 Title: Wavelets on Fractals Authors: Dorin E. Dutkay, Palle E. T. Jorgensen
[4]  math.CA/0405301 Title: Construction of Parseval wavelets from redundant filter systems Authors: L. W.
       Baggett, P. E. T. Jorgensen, K. D. Merrill, J. A. Packer
[5]  math.OA/0308131 An analogue of Bratteli-Jorgensen loop group actions for GMRA's Authors: L. W. Baggett,
       P. E. T. Jorgensen, K. D. Merrill, J. A. Packer
[6]  Ola Bratteli, Palle E. T. Jorgensen, “Wavelets through a looking glass”; Applied and Numerical Harmonic
      Analysis, Birkhäuser, Boston, 2002.

 « Back

Greedy algorithms in stable recovery of sparse overcomplete representations
Vladimir Temlyakov, Department of Mathematics, University of South Carolina

Overcomplete representations are attracting interest in signal processing theory, particularly due to their potential to generate sparse representations of signals. However, in general, the problem of finding sparse representations must be unstable in the presence of noise. We prove the possibility of stable recovery under a combination of sufficient sparsity and favorable structure of the overcomplete system.

We consider the situation where an ideal underlying signal indeed has a sparse representation but we can observe only a noisy version. We assume that the overcomplete system has a property of {\it mutual incoherence}, and that the ideal underlying signal has a sufficiently sparse representation, according to a sparsity threshold defined using the mutual coherence of the overcomplete system. Under these conditions, we show that the optimally--sparse approximation to the noisy data, to within the noise level, differs from the optimally--sparse decomposition of the ideal noiseless signal by at most a constant multiple of the noise level. It turns out that the Orthogonal Greedy Algorithm is a good method for approximate recovery of signals that have sparse representation.

 « Back

Highly sparse expansions from redundant dictionaries are unique and independent of the sparseness measure
Remi Gribonval, IRISA-INRIA, France

A series of recent results shows that if a signal admits a sufficiently sparse representation (in terms of the number of nonzero coefficients) in an“incoherent” dictionary, this solution is unique and can be recovered as the unique solution of a linear programming problem. We generalize these results to a large class of sparsity measures which includes -but is not restricted to- the p-sparsity measures for 0 p 1. We give sufficient conditions on a signal such that the simple solution of a linear programming problem simultaneously solves all the non-convex (and generally hard combinatorial) problems of sparsest representation w.r.t. arbitrary admissible sparsity measures. Our results should have a practical impact on source separation methods based on sparse decompositions, since they indicate that a large class of sparse priors can be efficiently replaced with a Laplacian prior without changing the resulting solution.

 « Back

Subdivision scheme tuning around extraordinary vertices
Leif Kobbelt, RWTH Aachen University

We extend the standard method to derive and optimize subdivision rules in the vicinity of extraordinary vertices (EV). Starting from a given set of rules for regular control meshes, we tune the extraordinary rules (ER) such that the necessary conditions for C1 continuity are satisfied along with as many necessary C2 conditions as possible. As usually done, our approach sets up the general configuration around an EV by exploiting rotational symmetry and reformulating the subdivision rules in terms of the subdivision matrix' eigencomponents. The degrees of freedom are then successively eliminated by imposing new constraints which allows us, e.g., to improve the curvature behavior around EVs. The method is flexible enough to simultaneously optimize several subdivision rules, i.e. not only the one for the EV itself but also the rules for its direct neighbors. Moreover it allows us to prescribe the stencils for the ERs and naturally blends them with the regular rules that are applied away from the EV. All the constraints are combined in an optimization scheme that searches in the space of feasible subdivision schemes for a candidate which satisfies some necessary conditions exactly and other conditions approximately. The relative weighting of the constraints allows us to tune the properties of the subdivision scheme according to application specific requirements. We demonstrate our method by tuning the ERs for the well-known Loop scheme and by deriving ERs for a √3-type scheme based on a 6-direction Box-spline.

 « Back

The Bramble-Hilbert Lemma and Whitney estimates for convex domains: applications to multivariate piecewise polynomial approximation
Dany Leviatan, School of Mathematics, Tel Aviv University

Download/View PDF

 

 « Back

Support vector machine soft margin classifiers
Ding-Xuan Zhou, Department of Mathematics, City University of Hong Kong

Learning Theory studies learning objects from random samples. Classification (binary) algorithms learn from samples ways of dividing an input space into two classes.

In this talk we discuss the q-norm soft margin classifier, a support vector machine classification algorithm. Our purpose is to provide a PAC error analysis for these algorithms. The misclassification error is bounded by the generalization error associated with a general class of loss functions. The difficulty of bounding the offset is overcome. The convergence rate is then provided by means of the regularization error and the approximation error for reproducing kernel Hilbert spaces. A projection operator is introduced to have better sample error estimates. As a corollary we give almost optimal error bounds for deterministic probability distributions. Even in the well known strictly separable case, our analysis yields the best error bound concerning the influence of the margin. Polynomial kernels and universal kernels are used to demonstrate the main results.

 « Back

Can the Fama-French three-factor model explain the cross-section of stock returns using wavelet analysis?
Francis In, Department of Accounting and Finance, Monash University, Australia

This paper examines whether the Fama and French three-factor model can explain the cross-section of stock returns over various time scales using a new approach. The new approach is based on a wavelet multiscaling method that decomposes a given time series on a scale-by-scale basis. Fama and French’s three factors are the market excess return, SMB, and HML. The empirical results provide strong support for the role of the market factor in explaining the cross-sectional variation of average stock return over all time scales. SMB is shown to play a role in explaining the cross-section of average returns over all time scales for small stocks, but not large stocks, while HML is effective in the short run, but not in the intermediate and long run.

 « Back

The analysis of intrinsic mode functions and instantaneous phase
Robert C. Sharpley, Industrial Mathematics Institute, University of South Carolina

Intrinsic Mode Functions (IMF's) were introduced by N. Huang and his collaborators as elementary monocomponents signals which arise through the Empirical Mode Decomposition (EMD) process for 1-D signals. These monocomponents have been highly successful in analyzing instantaneous phase for signals in a number of disciplines. The EMD method produces highly nonlinear redundant decompositions which are tailored to the signal, but currently lacks a theoretical base.

This presentation will describe recent work of the speaker and V. Vatchev on

  1. characterization of IMF's as solutions to Sturm-Liouville equations,
  2. comparison of the Hilbert transform method (i.e. analytic signals) for computing instantaneous phase against quadrature methods. In particular, extremely nice signals (both discrete and in closed form) will be presented for which the phase is non-monotone.
  3. mode classification of the IMF's of a signal through a Prufer quadrature method.

 « Back

Ultrasonic testing based on wavelet and JTFA
Xiufeng Zhang, Dept. of Mechanical Engineering, Tsinghua University, Beijing, China

Materials with coarse grains are utilized widely in industries, for example, austenitic steel is employed in the pressure vessels and pipes of nuclear, chemical industry. In ultrasonic non-destructive evaluation of these strong sound scattering materials the serious backscattering noise may attain peak values greater than the searched flaw pulse, and distort the flaw echoes. Theory and practice confirm that due to the time-varying characteristic of defect signals, it is more reliable to analysis it within a joint time-frequency representation than pure time-domain or frequency-domain. The joint time-frequency analysis (JTFA) of ultrasonic signals instantaneously reveals the frequency and time of arrival of target echoes which help to characterize the target. But varying results are obtained by applying different time-frequency (t-f) algorithms to ultrasonic signals. Performance evaluation of t-f algorithms for ultrasonic testing is useful for the choice of an algorithm. In this paper we present a performance comparison of t-f algorithms for ultrasonic applications. In particular, we compare the t-f algorithms based on the detection of the time of arrival and frequency of target echoes, estimation of instantaneous frequency of multiple target echoes, cross-terms attenuation and concentration, and signal-to-noise ratio (SNR) enhancement. These characteristics are studied by applying Wigner-Ville distribution (WVD), exponential distribution (ED), Gabor transform (GT), and wavelet transform (WT) to simulated and experimental ultrasonic data. The results indicate that WT tracks the frequency trend of the grain echoes and detects the flaw echoes better than all the other distributions discussed.

 « Back

Second generation wavelets for GIS data compression
Biswajeet Pradhan, Faculty of Engineering & Computer Technology, Asian Institute of Medicine, Science & Technology, Malaysia

Optimization of the size of terrain data set for generation of geographical surface images on computer with different level of details is a challenging task. The wavelet technique which allows to present information at different level of resolution is used in this problem. Triangulated Irregular Network (TIN), generated by using Delaunay Triangulation technique, is used to form the terrain from an arbitrary set of data. The data with the least effect in the representation of surface are selected and eliminated by using second generation wavelets. The quality of terrain representation after using proposed technique is compared with the original representation. The results show that this method can be used for significant reduction of data set.

 « Back

Stability of algorithms and generalization
Sayan Mukherjee, Center for Biological and Computational Learning, Massachusetts Institute of Technology

Supervised learning can be summarized as given follows: given a set of input-ouput pairs a function is learned that hopefully accurately predicts future observations of inputs. Thus the problem of learning can be thought of as a map from a dataset S to a function f. In classical learning theory the algorithm used to learn the function is empirical risk minimization (ERM): select a function from a class of functions that minimizes the error on the observed dataset. Classic results for this algorithm state that under appropriate constraints on the function class the function learnt will generalize, that is the error rate of future predictions will be the same as the obserevd error rate. In this talk we introduce a constraint on the map from the data to the function that ensures generalization. For the case of ERM we show that this condition is equivalent to the classical constraints on the function class. This formulation allows us to analyze functions from ERM to support vector machines to k-nearest neighbors using the same mathematical tool.

 « Back

Adaptive wavelet methods for modeling and computing turbulent flows
Kai Schneider, L3M-CNRS & CMI, Université de Provence, Marseille

This work is joint work with Marie Farge (ENS, Paris) and Giulio Pellegrino (L3M, Marseille).

Recently we have introduced a new wavelet-based method, called Coherent Vortex Simulation (CVS), to compute turbulent flows. The main idea is to split the flow into two orthogonal parts, a coherent flow and an incoherent background flow, using nonlinear filtering of vorticity. We have shown that the coherent flow is responsible for the nonlinear dynamics, and thus its evolution is deterministically computed in an adaptive wavelet basis, while the incoherent flow is noise-like, structureless and decorrelated. Therefore its influence is statistically modeled. Since the coherent part is described by only few wavelet modes, it is possible to reduce the computational cost, both in terms of memory requirement and CPU time. The developed adaptive wavelet solver for the 2D Navier-Stokes equations uses a semi-implicit time discretization and a Petrov-Galerkin scheme in space with wavelets as trial-functions and vaguelettes (operator adapted wavelets) as test-functions. To take into account no-slip boundary conditions and complex geometries we coupled the adaptive scheme with a volume penalization technique. In the talk we shall present different applications of the Coherent Vortex Simulation to 2D and 3D turbulent flows like 2D flows past different bluff bodies and a 3D turbulent mixing layer.

 « Back

Ideal interpolation
Carl de Boor, University of Wisconsin - Madison

Starting off with G. Birkhoff's definition of `ideal interpolation', multivariate polynomial interpolation and its error is explored.

 « Back

Adaptive multivariate approximation using binary space partitions and geometric wavelets
Shai Dekel, RealTimeImage

The Binary Space Partition (BSP) technique is a simple and efficient method to adaptively partition an initial given domain to match the geometry of a given input function. Here, we offer an enhancement to the BSP algorithm, which is in the spirit of geometric wavelets. This new approach is based on recent development in the theory of multivariate nonlinear piecewise polynomial approximation. We provide numerical examples of n-term geometric wavelet approximations of known test images and discuss applications to image denoising and compression. Preprints can be downloaded from: https://turbo.turboimage.com/~sd/default.html

 « Back

Generalized shift-invariant frames and duals for subspaces
Ole Christensen, Department of Mathematics, Technical University of Denmark

Download/View PDF

 

 « Back

Foldable figures, Fuchsian groups, and multiscale structures
Peter Massopust, TIPS - Tuboscope Integrated Pipeline Solutions

When defining a multiscale structure one often uses a dilation and a translation operator to pass information between and within the multiscales. The translation operator can be thought of as a means of tiling a multiscale level, starting with a fundamental domain that usually is a lattice cell. However, other tilings of multiscale levels such as those obtained be reflection instead of translation work as well. In this lecture we will illustrate this approach by considering foldable figures in Euclidean space and Fuchsian groups in the complex upper half plane or, equivalently, the Poincaré disk.

 « Back

On the existence of multiresolution analysis for framelets
Marcin Bownik, University of Oregon

We show that a compactly supported tight framelet comes from an MRA if the intersection of the dilations of the space of negative dilates, which is defined as the shift invariant space generated by the negative scales of a framelet, is trivial. We also construct examples of non-tight framelets, which are nevertheless arbitrarily close to tight frame framelets, such that the corresponding space of negative dilates is equal to the entire space L2(R). This talk is based on a joint work with Ziemowit Rzeszotnik.

 « Back

Construction of Parseval wavelets from a redundant system of filters
Larry Baggett, University of Colorado, USA

We generalize the classical infinite product construction of Mallat and Meyer of a multiresolution analysis and an orthonormal wavelet from a quadrature mirror filter $h.$ Our generalized construction is based on an extended notion of a filter to a matrix $H$ of periodic functions, which satisfies a more complex higher dimensional filter equation. The proof is more abstract than in the classical case and involves the analysis of a pair of partial isometries that satisfy a kind of Cuntz algebra relation.

 « Back

Non-linear subdivision schemes for curves
Nira Dyn, School of Mathematical Sciences Tel-Aviv University, Israel

This talk is concerned with several nonlinear subdivision schemes for curves, which are extensions of linear curve subdivision schemes. A class of curve subdivision schemes on manifolds is defined in terms of geodesic averages, replacing the usual averages between points, with which it is possible to represent in an alternative way affine invariant linear subdivision schemes with masks of finite support. In particular, explicit such presentation will be given for the quadratic and cubic B-spline schemes and for the 4-point interpolator-scheme. Another class of nonlinear schemes on manifolds is obtained by projecting the affine averages to the manifold. These two constructions of nonlinear schemes from linear ones, apply also to Lie groups and in particular to matrix groups as the Euclidean motion group, with geodesics replaced by exponential mappings, and to abstract manifolds such as the hyperbolic plane. The analysis of such a non-linear schemes is done by its proximity to the corresponding linear Euclidean scheme from which it was derived. The proximity condition, holds for the above mentioned two constructions. Two other nonlinear extensions of the 4-point interpolatory subdivision scheme will be discussed. One extension is for control points in Euclidean space. The refinement of the control points is driven by their geometry, a nonlinear feature which makes it possible to control the shape of the generated curve. The other extension is confined to functional data, and is motivated by the idea.

This talk reports on three joint works, one with J. Wallner, one with D. and M. Marinov and one with A. Cohen and B. Matei.

 « Back

MRA framelets and their connections to Riesz wavelets
Bin Han, Department of Maths and Stat Sciences, University of Alberta

Download/View PDF

 

 « Back

Riesz wavelets from the loop scheme in computer graphics
Bin Han, Department of Maths and Stat Sciences, University of Alberta

Download/View PDF

 

 « Back

Wavelet transforms and admissible group representations
Eric Scott Weber, Iowa State University, USA

Most continuous wavelet transforms arise from square-integrable group representations, and discrete wavelet transforms arise from discretizing, or sampling, the continuous transforms. We will talk about necessary and sufficient conditions for there to be either a continuous or a discrete wavelet transforms in terms of a generalization of square-integrability. We will also discuss the relationship of these conditions to several notions of multiresolution analysis.

 « Back

Image compression by linear splines over adaptive triangulations
Nira Dyn, Tel Aviv University, Israel

A new method for image compression is presented. The method is based on the approximation of an image, regarded as a function, by a linear spline over an adapted triangulation, D(Y), which is the Delaunay triangulation of a small set Y of significant pixels. The linear spline minimizes the mean square error to the image, among all linear splines over D(Y). The significant pixels in Y are selected by an adaptive thinning algorithm, which recursively removes less significant pixels in a greedy way, using a sofisticated measure of the significance of a pixel. The proposed compression method combines the approximation scheme with a customized scattered data coding scheme. We demonstrate that our compression method outperforms JPEG2000 on two geometric images and performs competitively with JPEG2000 on three popular test cases of real images.

This is a joint work with L. Demaret and A. Iske.

 « Back

Wavelets on triangulations
Doug Hardin, Mathematics Department, Vanderbilt University

I will review several constructions of wavelets and prewavelets on triangulations. In particular, I will present a recent construction of continuous, orthogonal, piecewise polynomial wavelets that may be adapted to a semi-regular sequence of triangulations. This construction is joint work with G. Donovan, J. Geronimo, and B. Kessler.

 « Back

Geometric image representation with Bandelets
Erwan LE Pennec, Ecole Polytechnique

The wavelet representation is currently one of the most used representation for images. This representation is optimal for smooth images but not for piecewise smooth image. The bandelet representation is one of the geometric representation that overcomes this limitation of the wavelets.

We present here the construction of this adaptive representation, explain how to choose the geometry defining the best bandelet representation in an non-linear approximation setting and show how to apply this framework to compression and estimation.

 « Back

The Bandelet representation: algorithms and applications
Erwan LE Pennec, Ecole Polytechnique

The wavelet representation is currently one of the most used representation for images. This representation is optimal for smooth images but not for piecewise smooth image. The bandelet representation is one of the geometric representation that overcomes this limitation of the wavelets.

After a brief presentation of the bandelets representation and its properties, we describe here their practical discrete implementation and show how to adapt them to various settings : image compression, seismic data restoration,...

 « Back

Wavelet-based extraction of coherent structures to analyze and compute Navier-Stokes equations
Marie Farge, LMD-CNRS, ENS Paris

Joint work with Kai Schneider, CMI, Universite de Provence, Marseille, France

Turbulent flows are solutions of the Navier-Stokes equations in the limit where the nonlinear advection term strongly dominates the linear dissipation term. Their highly nonlinear dynamics leads to the formation of coherent structures and exhibits a multiscale behaviour, characterized by a power-law energy spectrum.

In the first part we will present the wavelet-based method we have proposed [Phys. Fluids, 11(8), 1999 and PRL 87(55), 2001] to extract coherent vortices out of two and three dimensional turbulent flows. We will show that the coherent structures are responsible for the nonlinear dynamics, as for the non-Gaussian statistics and long-range spatial correlation. In contrast, the remaining incoherent modes exhibit a diffusive behaviour, with quasi-Gaussian statistics and spatial decorrelation.

In the second part, we will present the CVS (Coherent Vortex Simulation) method we have proposed to compute turbulent flows [Flow, Turbulence and Combustion, 66 (4), 2001]. It splits each flow realization into two orthogonal parts, a coherent and an incoherent flow, using the previous wavelet-based method, and computes the evolution of the coherent flow, while the incoherent modes are discarded at each time step to model turbulent dissipation.

To download our papers: //wavelets.ens.fr

 « Back

A recursive algorithm for nonlinear wavelet thresholding : Applications to signal and image processing
Marie Farge, LMD-CNRS, ENS Paris

Download/View PDF

 

 « Back

Pseudo-splines, wavelets and framelets
Dong Bin, National University of Singapore

Download/View PDF

 

 « Back

Reconstructing multivariate functions from scattered data efficiently
Holger Wendland, Institut für Numerische und Angewandte Mathematik, Georg-August Universität Göttingen

Download/View PDF

 

 « Back

Localized frames and sparsity
Karlheinz Gröchenig, Institute of Biomathematics and Biometry, Germany

We present a new concept to measure the quality of frames, namely "localization'' of frames which is defined by the decay properties of the associated Gramian matrix. Using Banach algebra methods, it will be shown that the canonical dual frame possesses the same localization properties. As a consequence, elements in a class of naturally associated Banach spaces possess sparse representations. We will further address the solution of problems on non-uniform Gabor frames and on sampling in shift-invariant spaces. (Part of this is joint work with M. Fornasier)

 « Back

The role of Banach algebra methods in numerical analysis and sampling theory
Karlheinz Gröchenig, Institute of Biomathematics and Biometry, Germany

We will give a survey of results in applied analysis where Banach algebra methods have led to progress and new results. Specifically, we will discuss

(a) several new estimates for the off-diagonal decay of the inverse of an (infinite-dimensional) matrix,
(b) the convergence properties of the finite section method in numerical analysis and give new error estimates,
(c) several measures for the quality of a frame and show that the dual frame and the associated tight frame possess the same quality,
(d) sparse representations with frames, and finally
(e) almost diagonalization of pseudodifferential operators (=time-varying systems) and their symbolic calculus.

 « Back

High resolution image reconstruction: a framelet approach
Sherman D. Riemenschneider, Department of Mathematics, West Virginia University

(joint with Raymond Chan, Lixin Shen and Zuowei Shen)

High resolution image reconstruction refers to the construction of high-resolution images from images from multiple low-resolution, shifted, and degraded samples of a true image. We review the model proposed by Bose and Boo and analyze it in a framelet setting and propose a framelet based iterative algorithm.

 « Back

Dual Gramian analysis
Zuowei Shen, Dept. of Mathematics, National University of Singapore

The dual Gramian analysis has been used in the study of the frame properties of shift invariant systems. It allows us to decompose the frame operator into a collection of simpler operators (fibers) in the Fourier domain. Each one of the `fiber’ operators acts from a sequence space to the same sequence space and its matrix representation can be explicitly described in terms of the Fourier transforms of the generators of the shift invariant system. Application to the Gabor systems which are shift invariant shows the power of the analysis. However, it cannot be directly applied to many other systems such as the wavelet systems, which are not shift invariant. This leads us to consider the dual Gramian analysis in a more general setting, i.e. general shift invariant systems which include wavelet system as a special case. In this talk, I will give the development and the recent results on the dual Gramian analysis.

 « Back

Unitary extension principle and applications
Zuowei Shen, Dept. of Mathematics, National University of Singapore

Since the unitary extension principle was published in 1997, it has led to much theoretic development and has been used in various applications. The unitary extension principle provides a great flexibility of designing tight frame wavelet filters and makes constructions of tight frame wavelets painless. In this talk, I will first briefly review the recent development based on or motivated by the unitary extension principle. Then I will focus on a few applications that use the unitary extension principle to design tight frame based algorithms. In particular, the power of this principle is illustrated by showing how it is used in solving various problems in the area of the high resolution image reconstruction.

 « Back

Tight wavelet frames with symmetry
Say Song Goh, Department of Mathematics National University of Singapore

Symmetric and antisymmetric compactly supported wavelets are much desired in many practical applications. In this talk, we shall provide a general, and yet simple, method to derive a new set of wavelets that are symmetric or antisymmetric from any given set of wavelets. The affine system generated by the new wavelets form a tight frame if the affine system generated by the original wavelets is so. Further, when the original wavelets are generated via a multiresolution analysis, the new wavelets can also be derived from a multiresolution analysis. As illustration, examples of new symmetric and antisymmetric compactly supported wavelets obtained from the method will be shown. This talk is based on joint work with Zuowei Shen and Zhi Yuan Lim.

 « Back

Multidisciplinary research projects of the Centre for Wavelets, Approximation and Information Processing
Say Song Goh, Department of Mathematics National University of Singapore

The Centre for Wavelets, Approximation and Information Processing of the National University of Singapore is a multidisciplinary research centre in the Department of Mathematics that emphasizes the synergy of mathematicians, engineers and computer scientists. Its activities, comprising basic research, applied research and systems development, focus on mathematical signal, image and information processing. In this talk, we shall introduce various multidisciplinary research projects of the centre. These include projects on image compression, document image compression, video compression and summarization, underwater signal denoising and separation, and time-frequency analysis, feature extraction and classification of whale call signals.

 « Back

Adaptive scattered data fitting by spline--wavelets: regularization and robust fitting
Angela Kunoth, Institut für Angewandte Mathematik und Institut für Numerische Simulation, Universität Bonn

Download/View PDF

 

 « Back

On vanishing moments for multivariate wavelets
Maria Skopina, Saint Petersburg University

We consider wavelets with matrix dilation. An explicit formula for masks providing vanishing moments is presented. The class of interpolatary masks with vanishing moments is also described. For an interpolatory mask, a formula for a dual mask which also provides vanishing moments of the same order is given. For this case, wavelet masks can be found explicitly.

 « Back

Spanning and sampling in Lebesgue, Sobolev and Hardy spaces
Huy-Qui Bui, University of Canterbury and Richard S. Laugesen, University of Illinois

The Shannon-Whittaker sampling theorem shows how to reconstruct a bandlimited signal from its sampled values at a discrete set of points, using the generating function sinc x. Wavelet and wavelet-frame theory, and quasi-interpolation theory in the Strang-Fix tradition, together form a collective effort to extend the Shannon-Whittaker sampling theorem to other classes of signals and to other methods of sampling (analyzing) and reconstructing (synthesizing). The former theory assumes the generating function to have vanishing integral, and the latter assumes a "constant periodization" condition.

This series of talks, part 1 by R.S. Laugesen and part 2 by H.-Q. Bui, presents a new development in the Strang-Fix tradition. We show how to explicitly approximate an "arbitrary" Lp function, using average sampling and then reconstructing by an affine system whose generating function has neither vanishing integral nor constant periodization. In other words, the signal, the analyzer and the synthesizer can all be essentially arbitrary. The key idea is automatic cancellation of errors at different scales.

Sobolev spaces can likewise be explicitly spanned, with the Strang-Fix condition being reduced by one degree from previous work.

The Hardy space H1 yields to this approach also.

 « Back

On the analysis of synthetic NMR signals
Jeff Kline, Department of Mathematics, University of Wisconsin - Madison

Nuclear Magnetic Resonance (NMR) describes a physical phenomenon exploited by a wide range of applications. In this talk, I will focus on NMR in the context of protein analysis. The data collected from protein analysis experiments is expected to live in a very "thin" space of signals: linear combinations of exponentially decaying sinusoids. Although each signal in this space has a simple parametric representation, recovery of this representation is challenging. I will provide a brief description of NMR to explain how it is used to analyze complex molecules, and then I will summarize a new approach towards parameter recovery.

 « Back

Constructing wavelet frames with given properties
Alexander Petukhov, Department of Mathematics, University of Georgia

We are going to talk about two algorithms for constructing wavelet bases and frames. The first one gives the way to construct compactly supported (anti)symmetric orthogonal wavelets and tight wavelet frames for MRAs generated by symmetric scaling function with an arbitrary integer dilation factor. The second one generates a compactly supported dual wavelet frame with the maximum number of vanishing moments for a given wavelet frame system whose MRA has a given approximation order. All symbols of obtained framelets and scaling functions are polynomials (not rational functions). This is the main difference of the presented algorithm from the existing ones.

 « Back

Fast algorithms for recovery of data irregularly subsampled from redundant representation
Alexander Petukhov, Department of Mathematics, University of Georgia

Download/View PDF

 

 « Back

Universal algorithms for learning theory
Albert Cohen, Laboratoire Jacques-Louis Lions, University Pierre et Marie Curie, France

Download/View PDF

 

 « Back

Analysis on multi-resolution mosaic images
Wen-Liang Hwang, Academia Sinica, Taiwan

Image mosaicing is the act of combining two or more images and is used in many applications in computer vision, image processing, and computer graphics. It aims to combine images such that no obstructive boundaries exist around over-lapped regions and to create a mosaic image that exhibits as little distortion as possible from the original images. In the proposed techniques, the to-be-combined images are first projected into wavelet subspaces. The images projected into the same wavelet subspace are then blended. Our blending function is derived from an energy minimization model which balances the smoothness around the over-lapped region and the fidelity of the blended image to the original images. Experiment results and subjective comparison with other methods are given.

 « Back

Ergodic dynamics in sigma-delta quantization: tiling invariant sets and spectral analysis of error
Sinan Güntürk, Department of Mathematics, New York University

Sigma-delta quantization is an A/D conversion method in which signals are represented by sequences that lie in a very restricted set of discrete amplitudes, such as 0 and 1 only. These sequences are very special in that for each input signal, local averages of the output bit sequence must track those of the input sample sequence, thus enabling simple convolutional reconstruction. We will be concerned with the approximation of constant functions only, which is a seemingly basic case that presents surprisingly complex behavior.

An m'th order sigma-delta scheme can be translated into a piecewise affine dynamical system on Rm. We first show that stable schemes yield invariant sets that tile the phase space by integer translations. In general these tiles may have a finite multiplicity. However when the multiplicity is one, the dynamics becomes isomorphic to a generalized skew-translation of the m-torus. We use this isomorphism to analyze the asymptotics of the convolutional reconstruction error, employing tools from Diophantine approximation and spectral theory.

This is joint work with Nguyen Thao.

 « Back

Projective modules and Hilbert modules for $C(\mathbb T^n)$
Judith Packer, University of Colorado, USA

The concept of finitely generated projective modules and Hilbert modules for unital C*-algebras will be reviewed. The emphasis will be on the study of these modules for the C*-algebra $C(\mathbb T^n),$ continuous complex-valued functions on the n-torus. We will go over the case $C(\mathbb T^2)$ in detail, with an emphasis on examples, and with an eye towards their use in the theory of projective multi-resolution analyses in Talk 2.

 « Back

Projective multi-resolution analyses for $L^2(\mathbb R^n)$
Judith Packer, University of Colorado, USA

The notion of “projective” multi-resolution is defined, for which, by definition, the initial space corresponds to a finitely generated projective module over the algebra $C(\mathbb T^n)$ of continuous complex-valued functions on the n-torus. The case of ordinary multi-wavelets is that in which the projective module is a free module. The properties of projective multi-resolution analyses, including the frames that they provide for $L^2(\mathbb R^n).$ Examples for n=2 and n=3 with diagonal dilation matrices are then discussed, where the initial module is not free. In these examples we compute the isomorphism class of the wavelet module. This work is joint with Marc A. Rieffel.

 « Back

Some operator relations for normal operators and connections to multi-resolution theory
Ola Bratteli, University of Oslo, Norway

This is joint work in progress with Palle E. T. Jorgensen. The starting point is Baggett and Merrills construction of scaling functions and wavelet functions from the action of dilations and translations on a preferred scaling subspace. The method can be extended to construct wavelet frames on more non-traditional structures by replacing the circle and the map z ---> z^N by more general dynamical systems as, say, a Julia set and the action of the corresponding rational function on the set. On the way we construct solutions of operator relations of the form UT = P(T)U where U is unitary, T is normal and P is a rational or more general function.

 « Back

Caplets: wavelets without wavelets
Amos Ron, Department of Computer Sciences, University of Wisconsin-Madison, USA

Download/View PDF

 

 « Back

The romance of hidden components
David Donoho, Statistics Department, Stanford University, USA

Perhaps the most romantic and seductive idea in all of science is that, hiding behind the enormously complex structures we see in the world around us, there are hidden components that are on the one hand very simple and even elegant and on the other hand easily combine to generate all the variety we see about us. Classical examples include Newton and the spectrum of light, Eugenecists and the idea of IQ; modern examples include wavelets and quarks. I will review some of the classical ideas of hidden components, starting from principal components or even before, and describe some of the most recent notions, such as independent components analysis, sparse components analysis, nonnegative matrix factorizations, and cumulant components. I will try to keep things at an elementary level, communicating the attractiveness of these ideas to scientists and engineers outside of statistics, the wide-ranging impact these ideas are having from high-tech industry to neuroscience and astronomy, and describing what I think is the much greater role that statisticians should be playing in developing and deploying these methods.

 « Back

The Calderon reproducing formula and Brownian motion functionals
Richard Gundy, Rutgers University, USA

Recent work on local solutions of the Navier-Stokes equation by Bhattacharya et al. has involved probability ideas and has renewed interest in the "background radiation" process in the upper half-plane in R4. Stochastic integral calculations with this process give representation formulas for compositions of Riesz transforms. In this lecture we will explain these computations in which a key role is played by the Calderon reproducing formula.

 « Back

Low pass filters, probability, and statistical mechanics
Richard Gundy, Rutgers University, USA

Scaling functions for multiresolution analyses are obtained from periodic functions known as low-pass filters. A. Cohen found the first necessary and sufficient conditions for a trigonometric polynomial to be a low-pass filter. In this lecture, we summarize what can be said about Cohen's condition when the candidate periodic function is not a polynomial. We shift the domain of discussion from R1 (or Rn) to a sequence space with a shift operation, and interpret the basic objects as random variables and probabilities on this sequence space. Cohen's conditions have simple interpretations as conditions on certain martingales. Moreover, this approach shows the limitations of his approach: there are continuous low-pass filters (and scaling functions) that do not satisfy his conditions, contrary to some claims in the literature. This kind of probability analysis is typical in statistical mechanics, and parts of Ergodic theory.

 « Back

Unitary systems, wavelet sets, and operator-theoretic interpolation of wavelets and frames
David R. Larson, Texas A&M University

There are three basic parts:

  • Part I. Unitary systems and wavelet sets

    A wavelet is a special case of a vector in a separable Hilbert space that generates a basis under the action of a collection, or "system", of unitary operators defined in terms of translation and dilation operations. This approach to wavelet theory goes back, in particular, to earlier work of Goodman, Lee and Tang in the context of multiresolution analysis. We will begin by describing the operator-interpolation approach to wavelet theory using the local commutant of a system that was worked out by the speaker and his collaborators a few years ago. This is really an abstract application of the theory of operator algebras, mainly von Neumann algebras, to wavelet theory. The concrete applications of operator-interpolation to wavelet theory include results obtained using specially constructed families of wavelet sets. In fact X. Dai and the speaker had originally developed our theory of wavelet sets specifically to take advantage of their natural and elegant relationships with these wavelet unitary systems. We will discuss some unpublished or partially published results that are due to this speaker and his former students. Some of these are related to published work of the Wutam Consortium. We will also discuss some new results and open questions.
     
  • Part II. Unitary systems and frames

    A frame is a sequence of vectors in a Hilbert space which is a compression of a basis for a larger space. (This is not the usual definition in the frame literature, but it is equivalent to the usual definition. In this spirit, the usual "inequality" definition can be thought of as an abstract characterization of a compression of a basis.) Because of this compression relationship between frames and bases, the unitary system approach to wavelets (and more generally: wandering vectors) is perfectly adaptable to frame theory. This idea was developed into a theory a few years ago by D. Han and the speaker. The use of the local commutant is along the same lines. We will describe some new results, notably those of the speaker in collaboration with W-S. Tang and E. Weber. And we will discuss some open questions.
     
  • Part III. Decompositions of operators and operator-valued frames

    We will discuss some joint work with K. Kornelson and others on construction of frames with targeted properties, and some joint work with V. Kaftal on a theory of operator-valued frames. This is indirectly related with joint work with M. Frank on modular frames. These are all related to targeted decompositions of positive operators.

 « Back

Continuously differentiable wavelets on triangulations
Rong-Qing Jia, Department of Mathematical Sciences University of Alberta, Canada

In this talk we will first review B-net representations of bivariate splines (piecewise polynomials) over triangulated polygonal domains. By using the B-net representation as a convenient tool, we will be able to find a nested family of spaces of continuously differentiable splines, which are suitable for multiresolution analysis. Furthermore, we will construct a basis of continuously differentiable wavelets with short support. The global stability of the wavelet basis will be established by employing certain techniques in the theory of function spaces. Our wavelet basis will be useful for numerical solutions of partial differential equations.

 « Back

Approximation power of refinable spaces
Olga Holtz, Department of Mathematics, University of California-Berkeley

Distributional solutions to a vector refinement equation generate shift-invariant spaces, whose approximation power is of interest in MRA, subdivision, and wavelets. The focus of this talk is on the space of compactly supported distributional solutions, which is always finite-dimensional. By extending the classical L2 notion of approximation order to shift-invariant subspaces of Sobolev spaces it is possible to characterize the approximation order of any shift-invariant space generated by a compactly supported distribution. The interplay among solutions to the same refinement equation can be further explored to improve on these generic results. To this end, the notion of universal supervectors is introduced and connected to the notions of polynomial reproduction, Sum rules, and Condition Zk. Finally, a novel approach is presented to derive approximation power of refinable vectors from their smoothness.

 « Back

Perturbation that does not affect the regularity of a linear subdivision scheme -- not even by a bit
Thomas Yu, Department of Mathematical Sciences, Rensselaer Polytechnic Institute

Nira Dyn will probably talk about nonlinear manifold subdivision schemes T derived from linear subdivision schemes S; and after a proximity condition between S and T is established, one can then conclude that T inherits certain regularity (Ck) properties from S. An open question is whether the nonlinearity in T may actually `take away' a fraction of the regularity from that of S -- it is perceivable that a small nonlinear perturbation may take away a little bit of smoothness.

I report three provable examples and also some computationally observed examples of which certain perturbation perturbation of a linear subdivision does not deteriorate smoothness at all -- not even by a fraction. Here smoothness is measured by critical Holder exponent. These examples (unrelated to the aforementioned subdivision schemes of manifold data) include: -- Lagrange interpolatory subdivision operated on mildly irregularly spaced multigrids (work of Daubechies, Guskov and Sweldens) -- Median Interpolating subdivision -- p-mean interpolating subdivision

I shall outline the techniques for proving these results, and address many open questions that remain.

 « Back

Biorthogonal Eigensystems of scaling operators
Seng Luan Lee, Dept. of Mathematics, National University of Singapore

Download/View PDF

 

 « Back

The dual weighted residual method from a wavelet perspective
Wolfgang Dahmen, Rheinisch Westfälische Technische Hochschule-Aachen, Germany

The dual weighted residual method (DWR) has been attracting increasing attention over the past years as a paradigm for so called goal oriented numerical schemes primarily in connection with Finite Element discretizations. In many complex numerical simulation tasks one is not interested in the whole solution of a PDE but merely in some local functional of the solution like point values, local averages or integrals over a lower dimensional manifold. “Goal oriented” then means to compute this value within some target accuracy tolerance at possibly low numerical cost. In particular, this could mean that the solution is only very poorly approximated in parts of the domain which results in significant savings in comparison with numerical schemes that first determine a global approximation of the solution and then apply the functional to that approximation. The approach is based on error representations derived from duality arguments. Since this representation can typically be bounded by sums of local quantities of residual type the equilibration of these quantities gives rise to adaptive resolution schemes. While this has been observed to work often very well in numerical applications no rigorous error estimates or convergence proofs seem to be known, not even for linear model problems. A principal obstruction is that the error representation involves the unknown dual solution. The relevant information on the dual solution has to be inferred from approximations to the dual problem. The need to understand the effect of the unresolved scales in those approximations indicates the multiscale nature of the problem. This talk reviews some recent joint work with A. Kunoth which aims at deriving rigorous convergence results by recasting the DWR method in a wavelet framework and exploiting recent advances in the development of adaptive wavelet solvers.

 « Back

Discretizing manifolds via minimum energy points
Doug Hardin, Mathematics Department, Vanderbilt University

Download/View PDF

 

 « Back

Subdivision in geometric modeling and computer graphics
Denis Zorin, New York University

In this tutorial I present on overview of multiresolution techniques in geometric modeling and computer graphics, focusing on constructions for surfaces of arbitrary topology. Subdivision for arbitrary meshes is a central element of many such constructions. I will review current research on subdivision schemes of arbitrary meshes, including the basic theory and applications and discuss theoretical problems.

  • Part I: Overview of multiresolution in graphics and geometric modeling.

    Mutliresolution methods are used in many principal areas of computer graphics: rendering, geometric modeling and simulation. In this part of the tutorial I will present a survey of existing applications and multiresolution representations used in them.
     
  • Part II: Subdivision surfaces: overview

    Subdivision surfaces are the basis of one of the common approaches to multiresolution geometric modeling. This part will include a survey subdivision schemes for irregular meshes, including the most common subdivision surface constructions, multiresolution surfaces based on subdivision and applications.
     
  • Part III: Subdivision surfaces: theory

    In this part I will discuss the most important results of subdivision surface theory, primarily related to smoothness of limit surfaces.
     
  • Part IV: Approximation on surfaces; alternatives and approaches

    The final part of the tutorial will be devoted to approximation of functions on surfaces, approximation by subdivision bases in particular. In conclusion I will discuss open problems and some alternative approaches.

 « Back