ترغب بنشر مسار تعليمي؟ اضغط هنا

Post-Processing of High-Dimensional Data

179   0   0.0 ( 0 )
 نشر من قبل Alexander Litvinenko
 تاريخ النشر 2019
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

Scientific computations or measurements may result in huge volumes of data. Often these can be thought of representing a real-valued function on a high-dimensional domain, and can be conceptually arranged in the format of a tensor of high degree in some truncated or lossy compressed format. We look at some common post-processing tasks which are not obvious in the compressed format, as such huge data sets can not be stored in their entirety, and the value of an element is not readily accessible through simple look-up. The tasks we consider are finding the location of maximum or minimum, or minimum and maximum of a function of the data, or finding the indices of all elements in some interval --- i.e. level sets, the number of elements with a value in such a level set, the probability of an element being in a particular level set, and the mean and variance of the total collection. The algorithms to be described are fixed point iterations of particular functions of the tensor, which will then exhibit the desired result. For this, the data is considered as an element of a high degree tensor space, although in an abstract sense, the algorithms are independent of the representation of the data as a tensor. All that we require is that the data can be considered as an element of an associative, commutative algebra with an inner product. Such an algebra is isomorphic to a commutative sub-algebra of the usual matrix algebra, allowing the use of matrix algorithms to accomplish the mentioned tasks. We allow the actual computational representation to be a lossy compression, and we allow the algebra operations to be performed in an approximate fashion, so as to maintain a high compression level. One such example which we address explicitly is the representation of data as a tensor with compression in the form of a low-rank representation.

قيم البحث

اقرأ أيضاً

Low-rank approximations of original samples are playing more and more an important role in many recently proposed mathematical models from data science. A natural and initial requirement is that these representations inherit original structures or pr operties. With this aim, we propose a new multi-symplectic method based on the Lanzcos bidiagonalization to compute the partial singular triplets of JRS-symmetric matrices. These singular triplets can be used to reconstruct optimal low-rank approximations while preserving the intrinsic multi-symmetry. The augmented Ritz and harmonic Ritz vectors are used to perform implicit restarting to obtain a satisfactory bidiagonal matrix for calculating the $k$ largest or smallest singular triplets, respectively. We also apply the new multi-symplectic Lanczos algorithms to color face recognition and color video compressing and reconstruction. Numerical experiments indicate their superiority over the state-of-the-art algorithms.
High order implicit-explicit (IMEX) methods are often desired when evolving the solution of an ordinary differential equation that has a stiff part that is linear and a non-stiff part that is nonlinear. This situation often arises in semi-discretizat ion of partial differential equations and many such IMEX schemes have been considered in the literature. The methods considered usually have a a global error that is of the same order as the local truncation error. More recently, methods with global errors that are one order higher than predicted by the local truncation error have been devised (by Kulikov and Weiner, Ditkowski and Gottlieb). In prior work we investigated the interplay between the local truncation error and the global error to construct explicit and implicit {em error inhibiting schemes} that control the accumulation of the local truncation error over time, resulting in a global error that is one order higher than expected from the local truncation error, and which can be post-processed to obtain a solution which is two orders higher than expected. In this work we extend our error inhibiting with post-processing framework introduced in our previous work to a class of additive general linear methods with multiple steps and stages. We provide sufficient conditions under which these methods with local truncation error of order p will produce solutions of order (p+1), which can be post-processed to order (p+2), and describe the construction of one such post-processor. We apply this approach to obtain implicit-explicit (IMEX) methods with multiple steps and stages. We present some of our new IMEX methods and show their linear stability properties, and investigate how these methods perform in practice on some numerical test cases.
Data sites selected from modeling high-dimensional problems often appear scattered in non-paternalistic ways. Except for sporadic-clustering at some spots, they become relatively far apart as the dimension of the ambient space grows. These features d efy any theoretical treatment that requires local or global quasi-uniformity of distribution of data sites. Incorporating a recently-developed application of integral operator theory in machine learning, we propose and study in the current article a new framework to analyze kernel interpolation of high dimensional data, which features bounding stochastic approximation error by a hybrid (discrete and continuous) $K$-functional tied to the spectrum of the underlying kernel matrix. Both theoretical analysis and numerical simulations show that spectra of kernel matrices are reliable and stable barometers for gauging the performance of kernel-interpolation methods for high dimensional data.
104 - Adi Ditkowski , Sigal Gottlieb , 2019
High order methods are often desired for the evolution of ordinary differential equations, in particular those arising from the semi-discretization of partial differential equations. In prior work in we investigated the interplay between the local tr uncation error and the global error to construct error inhibiting general linear methods (GLMs) that control the accumulation of the local truncation error over time. Furthermore we defined sufficient conditions that allow us to post-process the final solution and obtain a solution that is two orders of accuracy higher than expected from truncation error analysis alone. In this work we extend this theory to the class of two-derivative GLMs. We define sufficient conditions that control the growth of the error so that the solution is one order higher than expected from truncation error analysis, and furthermore define the construction of a simple post-processor that will extract an additional order of accuracy. Using these conditions as constraints, we develop an optimization code that enables us to find explicit two-derivative methods up to eighth order that have favorable stability regions, explicit strong stability preserving methods up to seventh order, and A-stable implicit methods up to fifth order. We numerically verify the order of convergence of a selection of these methods, and the total variation diminishing performance of some of the SSP methods. We confirm that the methods found perform as predicted by the theory developed herein.
79 - Bin Wu , Xue-Cheng Tai , 2020
A complete multidimential TV-Stokes model is proposed based on smoothing a gradient field in the first step and reconstruction of the multidimensional image from the gradient field. It is the correct extension of the original two dimensional TV-Stoke s to multidimensions. Numerical algorithm using the Chambolles semi-implicit dual formula is proposed. Numerical results applied to denoising 3D images and movies are presented. They show excellent performance in avoiding the staircase effect, and preserving fine structures.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا