ترغب بنشر مسار تعليمي؟ اضغط هنا

A critical analysis of self-supervision, or what we can learn from a single image

67   0   0.0 ( 0 )
 نشر من قبل Yuki Asano
 تاريخ النشر 2019
  مجال البحث الهندسة المعلوماتية
والبحث باللغة English




اسأل ChatGPT حول البحث

We look critically at popular self-supervision techniques for learning deep convolutional neural networks without manual labels. We show that three different and representative methods, BiGAN, RotNet and DeepCluster, can learn the first few layers of a convolutional network from a single image as well as using millions of images and manual labels, provided that strong data augmentation is used. However, for deeper layers the gap with manual supervision cannot be closed even if millions of unlabelled images are used for training. We conclude that: (1) the weights of the early layers of deep networks contain limited information about the statistics of natural images, that (2) such low-level statistics can be learned through self-supervision just as well as through strong supervision, and that (3) the low-level statistics can be captured via synthetic transformations instead of using a large image dataset.

قيم البحث

اقرأ أيضاً

We investigate the effects of multi-task learning using the recently introduced task of semantic tagging. We employ semantic tagging as an auxiliary task for three different NLP tasks: part-of-speech tagging, Universal Dependency parsing, and Natural Language Inference. We compare full neural network sharing, partial neural network sharing, and what we term the learning what to share setting where negative transfer between tasks is less likely. Our findings show considerable improvements for all tasks, particularly in the learning what to share setting, which shows consistent gains across all tasks.
Learning problems form an important category of computational tasks that generalizes many of the computations researchers apply to large real-life data sets. We ask: what concept classes can be learned privately, namely, by an algorithm whose output does not depend too heavily on any one input or specific training example? More precisely, we investigate learning algorithms that satisfy differential privacy, a notion that provides strong confidentiality guarantees in contexts where aggregate information is released about a database containing sensitive information about individuals. We demonstrate that, ignoring computational constraints, it is possible to privately agnostically learn any concept class using a sample size approximately logarithmic in the cardinality of the concept class. Therefore, almost anything learnable is learnable privately: specifically, if a concept class is learnable by a (non-private) algorithm with polynomial sample complexity and output size, then it can be learned privately using a polynomial number of samples. We also present a computationally efficient private PAC learner for the class of parity functions. Local (or randomized response) algorithms are a practical class of private algorithms that have received extensive investigation. We provide a precise characterization of local private learning algorithms. We show that a concept class is learnable by a local algorithm if and only if it is learnable in the statistical query (SQ) model. Finally, we present a separation between the power of interactive and noninteractive local learning algorithms.
We discuss the features of instabilities in binary systems, in particular, for asymmetric nuclear matter. We show its relevance for the interpretation of results obtained in experiments and in ab initio simulations of the reaction between $^{124}Sn+^{124}Sn$ at 50AMeV.}
The constraints on the value of the CKM phase gamma that may be achieved by prospective measurements of sin(2 beta) and sin(2 beta + gamma) are discussed. Significant constraints require quite small errors, and may depend on assumptions about strong phases. The measurement of sin(2 beta + gamma) combined with other experiments could provide valuable limits on new physics in Bd-Bdbar mixing.
Different from other multiple top-quark productions, triple top-quark production requires the presence of both flavor violating neutral interaction and flavor conserving neutral interaction. We describe the interaction of triple top-quarks and up-qua rk in terms of two dimension-6 operators; one can be induced by a new heavy vector resonance, the other by a scalar resonance. Combining same-sign top-quark pair production and four top-quark production, we explore the potential of the 13 TeV LHC on searching for the triple top-quark production.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا