Do you want to publish a course? Click here

Solar stereoscopy - where are we and what developments do we require to progress?

176   0   0.0 ( 0 )
 Added by Thomas Wiegelmann
 Publication date 2009
  fields Physics
and research's language is English




Ask ChatGPT about the research

Observations from the two STEREO-spacecraft give us for the first time the possibility to use stereoscopic methods to reconstruct the 3D solar corona. Classical stereoscopy works best for solid objects with clear edges. Consequently an application of classical stereoscopic methods to the faint structures visible in the optically thin coronal plasma is by no means straight forward and several problems have to be treated adequately: 1.)First there is the problem of identifying one dimensional structures -e.g. active region coronal loops or polar plumes- from the two individual EUV-images observed with STEREO/EUVI. 2.) As a next step one has the association problem to find corresponding structures in both images. 3.) Within the reconstruction problem stereoscopic methods are used to compute the 3D-geometry of the identified structures. Without any prior assumptions, e.g., regarding the footpoints of coronal loops, the reconstruction problem has not one unique solution. 4.) One has to estimate the reconstruction error or accuracy of the reconstructed 3D-structure, which depends on the accuracy of the identified structures in 2D, the separation angle between the spacecraft, but also on the location, e.g., for east-west directed coronal loops the reconstruction error is highest close to the loop top. 5.) Eventually we are not only interested in the 3D-geometry of loops or plumes, but also in physical parameters like density, temperature, plasma flow, magnetic field strength etc. Helpful for treating some of these problems are coronal magnetic field models extrapolated from photospheric measurements, because observed EUV-loops outline the magnetic field. This feature has been used for a new method dubbed magnetic stereoscopy. As examples we show recent application to active region loops.



rate research

Read More

67 - F. Nicastro 2016
In this article we first review the past decade of efforts in detecting the missing baryons in the Warm Hot Intergalactic Medium (WHIM) and summarize the current state of the art by updating the baryon census and physical state of the detected baryons in the local Universe. We then describe observational strategies that should enable a significant step forward in the next decade, while waiting for the step-up in quality offered by future missions. In particular we design a multi-mega-second and multiple cycle XMM-Newton legacy program (which we name the Ultimate Roaming Baryon Exploration, or URBE) aimed to secure detections of the peaks in the density distribution of the Universe missing baryons over their entire predicted range of temperatures.
This paper presents a review of the topic of galaxy formation and evolution, focusing on basic features of galaxies, and how these observables reveal how galaxies and their stars assemble over cosmic time. I give an overview of the observed properties of galaxies in the nearby universe and for those at higher redshifts up to z~10. This includes a discussion of the major processes in which galaxies assemble and how we can now observe these - including the merger history of galaxies, the gas accretion and star formation rates. I show that for the most massive galaxies mergers and accretion are about equally important in the galaxy formation process between z = 1-3, while this likely differs for lower mass systems. I also discuss the mass differential evolution for galaxies, as well as how environment can affect galaxy evolution, although mass is the primary criteria for driving evolution. I also discuss how we are beginning to measure the dark matter content of galaxies at different epochs as measured through kinematics and clustering. Finally, I review how observables of galaxies, and the observed galaxy formation process, compares with predictions from simulations of galaxy formation, finding significant discrepancies in the abundances of massive galaxies and the merger history. I conclude by examining prospects for the future using JWST, Euclid, SKA, and the ELTs in addressing outstanding issues.
We review the current state of automatic differentiation (AD) for array programming in machine learning (ML), including the different approaches such as operator overloading (OO) and source transformation (ST) used for AD, graph-based intermediate representations for programs, and source languages. Based on these insights, we introduce a new graph-based intermediate representation (IR) which specifically aims to efficiently support fully-general AD for array programming. Unlike existing dataflow programming representations in ML frameworks, our IR naturally supports function calls, higher-order functions and recursion, making ML models easier to implement. The ability to represent closures allows us to perform AD using ST without a tape, making the resulting derivative (adjoint) program amenable to ahead-of-time optimization using tools from functional language compilers, and enabling higher-order derivatives. Lastly, we introduce a proof of concept compiler toolchain called Myia which uses a subset of Python as a front end.
72 - Norbert Wermes 2018
Pixel detectors have been the working horse for high resolution, high rate and radiation particle tracking for the past 20 years. The field has spun off into imaging applications with equal uniqueness. Now the move is towards larger integration and fully monolithic devices with to be expected spin-off into imaging again. Many judices and prejudices that were around at times were overcome and surpassed. This paper attempts to give an account of the developments following a line of early prejudices and later insights.
In the present paper, we investigate the cosmographic problem using the bias-variance trade-off. We find that both the z-redshift and the $y=z/(1+z)$-redshift can present a small bias estimation. It means that the cosmography can describe the supernova data more accurately. Minimizing risk, it suggests that cosmography up to the second order is the best approximation. Forecasting the constraint from future measurements, we find that future supernova and redshift drift can significantly improve the constraint, thus having the potential to solve the cosmographic problem. We also exploit the values of cosmography on the deceleration parameter and equation of state of dark energy $w(z)$. We find that supernova cosmography cannot give stable estimations on them. However, much useful information was obtained, such as that the cosmography favors a complicated dark energy with varying $w(z)$, and the derivative $dw/dz<0$ for low redshift. The cosmography is helpful to model the dark energy.
comments
Fetching comments Fetching comments
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا