Applications of Particle Physics to the Early Universe by Leonardo Senatore Submitted to the Department of Physics in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Physics at the MASSACHUSETTS INSTITUTE OF TECHNOLOGY June 2006 ( Massachusetts Institute of Technology 2006. All rights reserved. Author........................................-. Department of Physics May 5, 2006 Certified by. ............................................. Alan Guth Victor F. Weisskopf Professor of Physics ;ghegis Supqrvisor Certifiedby................ V!im`n ArkaM-Hamed Professor of Physics Thesis Supervisor Acceptedby............... .......... V-.' ' .f. ..'. . ""'Thoma ' Greytak Associate Department Head ff Education MASSACHU i i $IIThT NO OF TECHNOLOGY Al lt. nr Lzz-%00I uu . LIBRARIES ARCHIVES Applications of Particle Physics to the Early Universe by Leonardo Senatore Submitted to the Department of Physics on May 5, 2006, in partial fulfillment of the requirements for the degree of Doctor of Philosophy in Physics Abstract In this thesis, I show some of the results of my research work in the field at the crossing between Cosmology and Particle Physics. The Cosmology of several models of the Physics Beyond the Standard Model is studied. These range from an inflationary model based on the condensation of a ghost-like scalar field, to several models motivated by the possibility that our theory is described by a landscape of vacua, as probably implied by String Theory, which have influence on the theory of Baryogenesis, of Dark Matter, and of Big Bang Nucleosynthesis. The analysis of the data of the experiment WMAP on the CMB for the search of a non-Gaussian signal is also presented and it results in an upper limit on the amount on non-Gaussianities which is at present the most precise and extended available. Thesis Supervisor: Alan Guth Title: Victor F. Weisskopf Professor of Physics Thesis Supervisor: Nima Arkani-Hamed Title: Professor of Physics 2 Acknowledgments I would like to acknowledge my supervisors, Prof. Alan Guth, from whom I learnt the scientific rigor, and Prof. Nima Arkani-Hamed, whose enthusiasm is always an inspiration; the other professors with whom I had the pleasure to interact with: Prof. Matias Zaldarriaga, Prof. Max Tegmark, and in particular Prof. Paolo Creminelli, with whom I ended up spending a large fraction of my time both at work and out of work; all the many Post-Docs and students who helped me, in particular Alberto Nicolis; and eventually all the friends I have met in the journey which brought me here. Often I was in the need of help, and I have always found someone that, smiling, was happy to give it to me. It is difficult to imagine, being a foreign far from home, if without all these people I would have been able to arrive to this point. At the end, a special thank must go to my parents and to my girlfriend, who have always been close to me. 3 Contents 1 Introduction 7 2 Tilted Ghost Inflation 14 2.1 Introduction. . . . . . . . . . . . . . . . . . . . . 14 2.2 Density Perturbations. . . . . . . . . . . . . . . . . . . . . 16 . . . . . . . . . . . . . . . . . . . . 18 2.3 Negative Tilt. 2.3.1 2-Point Function ..... . . . . . . . . . . . . . . . . . . . . 18 2.3.2 3-Point Function ..... . . . . . . . . . . . . . . . . . . . . 21 2.3.3 Observational Constraints . . . . . . . . . . . . . . . . . . . . 27 2.4 Positive Tilt ............ . . . . . . . . . . . . . . . . . . . . 29 2.5 . . . . . . . . . . . . . . . . . . . . 33 Conclusions ............ 3 How heavy can the Fermions in Spliit Susy be? A study on Gravitino and Extradimensional LSP. 36 3.1 Introduction. . . . . . . . . . . . . . . . . . . . . 37 3.2 Gauge Coupling Unification . . . . . . . . . . . . . . . . . . . . . . . 39 3.2.1 Gaugino Mass Unification 42 3.2.2 Gaugino Mass Condition froom Anomaly Mediation . . . . . . . . . . . . . . . . . . . . ...... 45 3.3 Hidden sector LSP and Dark Matt,er Abundance 3.4 Cosmological Constraints ..... . . . . . . . . . . . . . . . . . . . . 51 3.5 Gravitino LSP .......... . . . . . . . . . . . . . . . . . . . . 56 3.5.1 Neutral Higgsino, Neutral AVino, 3.5.2 Photino NLSP ..... 4 . . . . . . . . .... 48 and Chargino NLSP ..... 58 . . . . . . . . . . . . . . . . . . . . 61 3.5.3 ............. .. 63 Bino NLSP. 3.6 Extradimansional LSP ..... . . . . . . . . . . . . . . 64 3.7 Conclusions ........... . . . . . . . . . . . . . . 68 3.7.1 Gravitino LSP ...... . . . . . . . . . . . . . . 69 3.7.2 ExtraDimensional LSP . . . . . . . . . . . . . . . 69 4 Hierarchy from Baryogenesis 4.1 Introduction 4.2 Hierarchy 76 . . . . . . . . . . . . . . . . from Baryogenesis . . . . . . . . ...... . . . . . . . . . . . . . . . 4.2.1 Phase Transition more in detail 4.3 Electroweak Baryogenesis . . .... 80 .......... 85 .................... 4.3.1 CP violation sources .............. 4.3.2 Diffusion Equations . 77 89 91 . . . . . . . . . . . . . . . ...... 4.4 Electric Dipole Moment 4.5 Comments on Dark Matter and Gauge Coupling Unification 4.6 Conclusions 4.7 Appendix: Baryogenesis in the Large Velocity Approximation 100 ......................... . . . . . . . . . . . . . . . . 106 . ..... 111 . . . . . . . ...... 112 .... 114 5 Limits on non-Gaussianities from WMAP data 123 5.1 Introduction. ....... 124 5.2 A factorizable equilateral shape .............. ....... 127 5.3 Optimal estimator for fNL ....... 133 5.4 Analysis of WMAP 1-year data ....... 137 5.5 Conclusions ......................... ....... 144 ................. .............. 6 The Minimal Model for Dark Matter and Unification 154 ....... ....... 155 158 6.1 Introduction. 6.2 The Model . 6.3 Relic Abundance. . . . . 159 .. 6.3.1 .... . 161 .. .... . 163 .. Higgsino Dark Matter. 6.3.2 Bino Dark Matter . 5 Direct Detection 6.5 Electric Dipole Moment 6.6 Gauge Coupling Unification 6.6.1 and matching Running 165 ............................. 6.4 ......................... 166 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ... 6.7 Conclusion 6.8 Appendix A: The neutralino mass matrix ................................ .. 168 171 178 . ............... 6.9 Appendix B: Two-Loop Beta Functions for Gauge Couplings ..... 7 Conclusions 179 180 187 6 Chapter 1 Introduction During my studies towards a PhD in Theoretical Physics, I have focused mainly on the connections between Cosmology and Particle Physics. In the very last few years, there has been a huge improvement in the field of Observational Cosmology, in particular thanks to the experiments on the cosmic microwave background (CMB) and on large scale structures. The much larger precision of the observational data in comparison with just a few years ago has made it possible to make the connection with the field of Theoretical Physics much stronger. This has occurred at a time in which the field of Particle Physics is experiencing a deficiency of experimental data, so that the indications coming from the observational field of Cosmology have become very important, or even vital. This is particularly true, for example, for the possible Physics associated with the ultra-violet (UV) completion of Einstein's General Relativity (GR), as, for example, String Theory. Because of this, Cosmology has become both a field where new ideas from Particle Physics can be tested, potentially even verified or ruled out with great confidence, and finally applied in search for a more complete understanding of the early history of the universe, and also a field which, with its observations, can motivate new ideas on the Physics of the fundamental interactions. Some scientific events which occurred during my years as a graduate student have particularly influenced my work. From the cosmological side, the results produced by the Wilkinsons Microwave 7 Anisotropy Probe (WMAP) experiment on the CMB have greatly improved our knowledge of the current cosmological parameters [1]. These new measurements have gone into the direction of verifying the generic predictions of the presence of a primordial inflationary phase in our universe, even though not yet at the necessary high confidence level to insert inflation in the standard paradigm for the history of the early universe. The first careful study of the non-Gaussian features of the CMB has been of particular importance for my work, as I will highlight later. Further, these new data have shown with high confidence the detection of the presence of a form of energy, called Dark Energy, very similar to the one associated with a cosmological constant, which at present dominates the whole energy of the universe. This last very unexpected discovery from Cosmology is an example of how cosmological observations can affect the field of Theoretical Physics. In fact, this observation has been of great influence on the development of a new set of ideas in Theoretical Physics which have strongly influenced my work. In fact, the problem of the smallness of the cosmological constant is one of the biggest unsolved problems of present Theoretical Physics, and, until these discoveries, it had been hoped that a UV complete theory of Quantum Gravity, once understood, would have given a solution. However, from the theoretical side, String Theory, which is, up to now, the only viable candidate for being a UV completion of GR, has shown that there exist consistent solutions with a non-null value of the cosmological constant, and that there could be a huge Landscape of consistent solutions, with different low energy parameters. Therefore, the theoretical possibility that the cosmological constant could be non zero, and that it could have many different values according to the different possible histories of the universe, together with the observational fact that the cosmological constant in our universe appears to be non zero, has led to a revival of a proposal by Weinberg [2] that the environmental, or anthropic, argument that structures should be present in our universe in order for there to be life, could force the value of the cosmological constant to be small. Recently, this argument has then been extended to the possibility that also other parameters of the low energy theory might be determined by environmental arguments [3, 4]. For obvious reasons, this general class of ideas is 8 usually referred to as Landscape motivated, and here I will do the same. In this thesis, I will illustrate some of the works I did into these directions. I consider useful to explain them in the chronological order in which they were developed. In chapter 2, I will develop a model of inflation where the acceleration of the universe is driven by the ghost condensate [5] . The ghost condensate is a recently proposed mechanism through which the vacuum of a scalar field can be characterized by a constant, not null, speed for the field [6]. The inflationary model based on this has the interesting features of generically predicting a detectable level of non-Gaussianities in the CMB, and of being the first example which shows that it is possible to have an inflationary model where the Hubble rate grows with time. The studies performed in this work have resulted in the publication in the journal Physical Review D of the paper in [7]. In chapter 3, in the context of Split Supersymmetry [3], which is a recently proposed supersymmetric model motivated by the Landscape where the scalar superpartners are very heavy, and the higgs mass is finely tuned for environmental reasons, I will study the possibility that the dark matter of the universe is made up of particles which interact only gravitationally, such as the gravitino. This possibility, if realized in nature, would have deep consequences on the possible spectrum of Split Supersymmetry and its possible detection at the next Large Hadron Collider (LHC). I will conclude that observational bounds from Big Bang Nucleosynthesis strongly constrain the possibility that the gravitino is the dark matter in the context of Split Supersimmetry, while other hidden sector candidates are still allowed. The studies performed in this work have resulted in the publication in the journal Physical Review D of the paper in [8]. In chapter 4, I will study a model developed in the context of the Landscape, in which the hierarchy between the Weak scale and the Plank scale is explained in terms of requiring that baryons are generated in our universe through the mechanism of Electroweak Baryogenesis. While the signature of this model at LHC could be 9 very little, the constraint from baryogenesis implies that the system is expected to show up within next generation experiments on the electron electric dipole moment (EDM). The studies performed in this work have resulted in the publication in the journal Physical Review D of the paper in [9]. In chapter 5, I will show a work in which I performed the analysis of the actual data on the CMB from the WMAP experiment, in search of a signal of non-Gaussianity in the temperature fluctuations. I will improve the analysis done by the WMAP team, and I will extend it to other physically motivated models. At the moment, the limit I will quote on the amplitude of the non-Gaussian signal is the most stringent and complete available. The studies performed in this work have resulted in the publication of the paper in [10], accepted for publication by the Journal of Cosmology and Astroparticle Physics in May 2006. In chapter 6, again inspired by the possibility that there is a Landscape of vacua in the fundamental theory, and that the Higgs particle mass might be finely tuned for environmental reasons, I will study the minimal model beyond the Standard Model that would account for the dark matter in the universe as well as gauge coupling unification. I will find that this minimal model can allow for exceptionally heavy dark matter particles, well beyond the reach of LHC, but that the model should however reveal itself within next generation experiments on the EDM. I will embed the model in a Grand Unified Extradimenesional Theory with an extra dimension in an interval, and I will study in detail its properties. The studies performed in this work has resulted in the publication in the journal Physical Review D of the paper in [11]. This will conclude a summary of some of the research work I have been doing during these years. Both the field of Cosmology and of Particle Physics are expecting great results from the very next years. Improvement of the data from WMAP experiment, as well as the launch of Plank and CMBPol experiments are expected to give us enough data to understand with 10 confidence the primordial inflationary era of our universe. A detection in the CMB of a tilt in the power spectrum of the scalar perturbations, or of a polarization induced by a background of gravitational waves, or of a non-Gaussian signal would shed light into the early phases of the universe, and therefore also on the Physics at very high energy well beyond what at present is conceivable to reach at particle accelerators. Combined with a definite improvement in the data on the Supernovae, expected in the very near future, these new experiment might shed light also on the great mystery of the present value of the Dark Energy, testing with great accuracy the possibility that it is constituted by a cosmological constant. The turning on of LHC will probably let us understand if the solution to the hierarchy problem is to be solved in some natural way, as for example with TeV scale Supersymmetry, or if there is evidence of a tuning which points towards the presence of a Landscape. It is a pleasure to realize that by the next very few years, many of these deep questions might have an answer. 11 Bibliography [1] D. N. Spergel et al. [WMAP Collaboration], "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Determination of Cosmological Parameters," Astrophys. J. Suppl. 148, 175 (2003) [arXiv:astro-ph/0302209]. [2] S. Weinberg, "Anthropic Bound On The Cosmological Constant," Phys. Rev. Lett. 59 (1987) 2607. [3] N. Arkani-Hamed and S. Dimopoulos, "Supersymmetric unification without low energy supersymmetry and signatures for fine-tuning at the LHC," JHEP 0506, 073 (2005) [arXiv:hep-th/0405159]. [4] N. Arkani-Hamed, S. Dimopoulos and S. Kachru, "Predictive landscapes and new physics at a TeV," arXiv:hep-th/0501082. [5] N. Arkani-Hamed, P. Creminelli, S. Mukohyama and M. Zaldarriaga, "Ghost inflation," JCAP 0404, 001 (2004) [arXiv:hep-th/0312100]. [6] N. Arkani-Hamed, H. C. Cheng, M. A. Luty and S. Mukohyama, "Ghost con- densation and a consistent infrared modification of gravity," JHEP 0405, 074 (2004) [arXiv:hep-th/0312099]. [7] L. Senatore, "Tilted ghost inflation," Phys. Rev. D 71, 043512 (2005) [arXiv:astro-ph/0406187]. [8] L. Senatore, "How heavy can the fermions in split SUSY be? A study on gravitino and extradimensional LSP," Phys. Rev. D 71, 103510 (2005) [arXiv:hepph/0412103]. 12 [9] L. Senatore, "Hierarchy from baryogenesis," Phys. Rev. D 73, 043513 (2006) [arXiv:hep-ph/0507257]. [10] P. Creminelli, A. Nicolis, L. Senatore, M. Tegmark and M. Zaldarriaga, "Limits on non-Gaussianities from WMAP data," arXiv:astro-ph/0509029. [11] R. Mahbubani and L. Senatore, "The minimal model for dark matter and unification," Phys. Rev. D 73, 043510 (2006) [arXiv:hep-ph/0510064]. 13 Chapter 2 Tilted Ghost Inflation In a ghost inflationary scenario, we study the observational consequences of a tilt in the potential of the ghost condensate. We show how the presence of a tilt tends to make contact between the natural predictions of ghost inflation and the ones of slow roll inflation. In the case of positive tilt, we are able to build an inflationary model in which the Hubble constant H is growing with time. We compute the amplitude and the tilt of the 2-point function, as well as the 3-point function, for both the cases of positive and negative tilt. We find that a good fraction of the parameter space of the model is within experimental reach. 2.1 Introduction Inflation is a very attractive paradigm for the early stage of the universe, being able to solve the flatness, horizon, monopoles problems, and providing a mechanism to generate the metric perturbations that we see today in the CMB [1]. Recently, ghost inflation has been proposed as a new way for producing an epoch of inflation, through a mechanism different from that of slow roll inflation [2, 3]. It can be thought of as arising from a derivatively coupled ghost scalar field 14 which condenses in a background where it has a non zero velocity: (X)= 2 (q) = M2 t (2.1) where we take M 2 to be positive. Unlike other scalar fields, the velocity (X) does not redshift to zero as the universe expands, but it stays constant, and indeed the energy momentum tensor is identical of that of a cosmological constant. However, the ghost condensate is a physical fluid, and so, it has physical fluctuations which can be defined as: = M 2t + 7r (2.2) The ghost condensate then gives an alternative way of realizing De Sitter phases in the universe. The symmetries of the theory allow us to construct a systematic and reliable effective Lagrangian for 7r and gravity at energies lower than the ghost cut-off M. Neglecting the interactions with gravity, around flat space, the effective Lagrangian for r has the form: a S d4X*2 2 21~2 (V27r)2 2i *(V7r)2 + 2M 2 (2.3) where a and 3 are order one coefficients. In [2], it was shown that, in order for the ghost condensate to be able to implement inflation, the shift symmetry of the ghost field p had to be broken. This could be realized adding a potential to the ghost. The observational consequences of the theory were no tilt in the power spectrum, a relevant amount of non gaussianities, and the absence of gravitational waves. The non gaussianities appeared to be the aspect closest to a possible detection by experiments such as WMAP. Also the shape of the 3-point function of the curvature perturbation Cwas different from the one predicted in standard inflation. In the same paper [2], the authors studied the possibility of adding a small tilt to the ghost potential, and they did some order of magnitude estimate of the consequences in the case the potential decreases while X increases. 15 In this chapter, we perform a more precise analysis of the observational consequences of a ghost inflation with a tilt in the potential. We study the 2-point and 3-point functions. In particular, we also imagine that the potential is tilted in such a way that actually the potential increases as the value of X increases with time. This configuration still allows inflation, since the main contribution to the motion of the ghost comes from the condensation of the ghost, which is only slightly affected by the presence of a small tilt in the potential. This provides an inflationary model in which H is growing with time. We study the 2-point and 3-point function also in this case. The chapter is organized as follows. In section 2.2, we introduce the concept of the tilt in the ghost potential; in section 2.3 we study the case of negative tilt, we compute the 2-point and 3-point functions, and we determine the region of the parameter space which is not ruled out by observations; in section 2.4 we do the same as we did in section 2.3 for the case of positive tilt; in section 2.5 we summarize our conclusions. 2.2 Density Perturbations In an inflationary scenario, we are interested in the quantum fluctuations of the r field, which, out of the horizon, become classical fluctuations. In [3], it was shown that, in the case of ghost inflation, in longitudinal gauge, the gravitational potential 1Idecays to zero out of the horizon. So, the Bardeen variable is simply: H -- H= - (2.4) and is constant on superhorizon scales. It was also shown that the presence of a ghost condensate modifies gravity on a time scale r -1 , with F M3 /Mpl, and on a length scale m-l 1, with m - M 2 /Mpl. The fact that these two scales are different is not a surprise since the ghost condensate breaks Lorentz symmetry. Requiring that gravity is not modified today on scales smaller than the present 16 Hubble horizon, we have to impose < Ho, which implies that gravity is not modified during inflation: r <<m < H (2.5) This is equivalent to the decoupling limit Mpl --+ o, keeping H fixed, which implies that we can study the Lagrangian for r neglecting the metric perturbations. Now, let us consider the case in which we have a tilt in the potential. Then, the zero mode equation for r becomes: +3H + V' = 0 (2.6) which leads to the solution: V' r= 3H (2.7) We see that this is equivalent to changing the velocity of the ghost field. In order for the effective field theory to be valid, we need that the velocity of 7rto be much smaller than M 2 , so, in agreement with [2], we define the parameter: 62 V' =-3HM 22 for V' < 0 3HM V' 62 = + for V' > 0 3HM 2 (2.8) to be 62 < 1. We perform the analysis for small tilt, and so at first order in 62. At this point, it is useful to write the 0-0 component of the stress energy tensor, for the model of [3]: To = -M 4P(X) + 2M4P'(X)q 2 + V(0) (2.9) where X = 09,Oq50u. The authors show that the field, with no tilted potential, is attracted to the minimum of the function P(X), such that, P(Xmi,) = M 2. So, adding a tilt to the potential can be seen as shifting the ghost field away from the minimum of P(X). Now, we proceed to studying the two point function and the three point function 17 for both the cases of a positive tilt and a negative tilt. 2.3 Negative Tilt Let us analyze the case V' < 0. 2.3.1 2-Point Function To calculate the spectrum of the 7r fluctuations, we quantize the field as usual: (2.10) rk(t) = Wk(t)ak + W (t)at k The dispersion relation for Wk is: 4 W2 = a-k + j2k2 k M2 (2.11) (2.11) ± Note, as in [3], that the sign of p is the same as the sign of < X >= M 2 . In all this chapter we shall restrict to P > 0, and so the sign of 3 is fixed. We see that the tilt introduces a piece proportional to k2 in the dispersion relation. This is a sign that the role of the tilt is to transform ghost inflation to the standard slow roll inflation. In fact, w2 - k 2 is the usual dispersion relation for a light field. Defining Wk(t) = uk(t)/a, and going to conformal time dr = dt/a, we get the following equation of motion: k4H22 272 up, + (/6 2 k2 + a kH 2 2)u k = 0 (2.12) If we were able to solve this differential equation, than we could deduce the power spectrum. But, unfortunately, we are not able to find an exact analytical solution. Anyway, from (2.12), we can identify two regimes: one in which the term dominates at freezing out, w - H, and one in which it is the term in - k4 k2 that dominates at that time. Physically, we know that most of the contribution to the 18 shape of the wavefunction comes from the time around horizon crossing. So, in order for the tilt to leave a signature on the wavefunction, we need it to dominate before freezing out.There will be an intermediate regime in which both the terms in k2 and k4 will be important around horizon crossing, but we decide not to analyze that case as not too much relevant to our discussion. So, we restrict to: 62 > a1/2 H >2 (2.13) _ where cr stays for crossing. In that case, the term in k 2 dominates before freezing out, and we can approximate the differential equation (2.12) to: 2 "2 Uk+ (k2 - (2.14) )uk= 0 where k = 3 1/ 2 5k. Notice that this is the same differential equation we would get for the slow roll inflation upon replacing k with k. Solving with the usual vacuum initial condition, we get: Wk =-H7 e-i0 /2 i (1- ) (2.15) which leads to the power spectrum: P= and, using = k3 2 Iwk( 0)12 Wk (q2 = 4' H2 2 /33 / 2 63 (2.16) 7T, H4 PC = 47r2o3/253M 4 (2.17) This is the same result as in slow roll inflation, replacing k with k. Notice that, on the contrary with respect to standard slow roll inflation, the denominator is not suppressed by slow roll parameters, but by the 62 term. 19 The tilt is given by: 2M 2V' = HV dink V"/ + H2 3 dln 2 dlnk dln(H) dln(P¢) dln(k) 1 2 262 9 3 dln52 2 dink - 2 ddln) )Ik= aH =(2.18) dink p1/26 + 4M4(1- 2P"M8)) where k = -r7/25is the momenta at freezing out, and where P and its derivatives are evaluated at Xmin. Notice the appearance of the term - , which can easily be the dominant piece. Please remind that, anyway, this is valid only for 62 > 6c,. Notice also that, for the effective field theory to be valid, we need: V' (2.19) 3H so, Mvi < M . This last piece is in general << 1 if the ghost condensate is present today. In order to get an estimate of the deviation from scale invariance, we can see that the larger contribution comes from the piece in - vH2. From the validity of the effective field theory, we get: 62 M 2 H IV'I IV"/A) = IV"I(M 2 /IH)Ne = V"l < 62 Ne (2.20) where Ne is the number of e-foldings to the end of inflation. So, we deduce that the deviation of the tilt can be as large as: Ins- -<N Ne This is a different prediction from the exact n8 = 1 in usual ghost inflation. 20 (2.21) 2.3.2 3-Point Function Let us come to the computation of the three point function. The leading interaction term (or the least irrelevant one), is given by [2]: eHt Lint = -,3 2M2 (*(V702) (2.22) Using the formula in [4]: < ,k(t)l(t)7(t) >=- i dt' < [rk,(t)7k2 (tk(t)l(t), d3xHt(t')] > (2.23) we get [2]: <7rkl7rk2 7rk3 >= 2 (27r35( M2(2)6( ki) k) (2.24) wl(O)w2 (O)w3 (O)((k2 .k3 )I(1, 2, 3) + cyclic + c.c) where cyclic stays for cyclic permutations of the k's, and where f0 I(1,2,3) = 1 1 l()'w2* (v)w3(n) (2.25) and the integration is performed with the prescription that the oscillating functions inside the horizon become exponentially decreasing as --+-oo. We can do the approximation of performing the integral with the wave function (2.15). In fact, the typical behavior of the wavefunction will be to oscillate inside the horizon, and to be constant outside of it. Since we are performing the integration on a path which exponentially suppresses the wavefunction when it oscillates, and since in the integrand there is a time derivative which suppresses the contribution when a wavefunction is constant, we see that the main contribution to the three point function comes from when the wavefunctions are around freezing out. Since, in that case, we are guaranteed that the term in k2 dominates, then we can reliably approximate the 21 that ( = - H7r,weget: wavefunctions in the integrand with those in (2.15). Using H8 ~~~~~~ < (k(k H8 3 6 sM ki)4 2 ,k3 >= (27r)363(Z (2.26) 3) ((k 2 + k3)kt + k2+ 2k 3 k2) + cyclic) k3 i~=~lk (kl2( where ki = ki . Let us define F(k, k 2 , k3 ) = 3 3 which, apart for the (k(k 2 .k3 ) ((k 2 + k 3)kt + kt2+ 2k3k2) + cyclic) (2.27) function, holds the k dependence of the 3-point function. The obtained result agrees with the order of magnitude estimates given in [2]: <__3_> (< (2 >)3/2 1 1 H (< 1_(H) ¢3 >)I()/ 8 M (1(H)2)3/2 6 el, 1 H H 67/2 (M -()2 (2.28) (2 The total amount of non gaussianities is decreasing with the tilt. This is in agreement with the fact that the tilt makes the ghost inflation model closer to slow roll inflation, where, usually, the total amount of non gaussianities is too low to be detectable. The 3-point function we obtained can be better understood if we do the following observation. This function is made up of the sum of three terms, each one obtained on cyclic permutations of the k's. Each of these terms can be split into a part which is typical of the interaction and of scale invariance, and the rest which is due to the wave function. For the first cyclic term, we have: Interaction (k 2 .k 3 ) (2.29) while, the rest, which I will call wave function, is: ((k + k3 )kt + k + 2k 2 k 3) 2 Wavefunction =(2.30) The interaction part appears unmodified also in the untilted ghost inflation case. While the wave function part is characteristic of the wavefunction, and changes in 22 the two cases. Our 3-point function can be approximately considered as a function of only two independent variables. The delta function, in fact, eliminates one of the three momenta, imposing the vectorial sum of the three momenta to form a closed triangle. Because of the symmetry of the De Sitter universe, the 3-point function is scale invariant, and so we can choose [ki[ = 1. Using rotation invariance, we can choose k1 = 1, and impose k2 to lie in the 61, 62 plane. So, we have finally reduced the 3-point function from being a function of 3 vectors, to be a function of 2 variables. From this, we can choose to plot the 3-point function in terms of xi - ki, i = 1, 2. The result is shown in fig.2-1. Note that we chose to plot the three point function with a measure equal to x2x2. The reason for this is that this results in being the natural measure in the case we wish to represent the ratio between the signal associated to the 3-point function with respect to the signal associated to the 2-point function [5]. Because of the triangular inequality, which implies 3 < 1 - x 2, and in order to avoid to double represent the same momenta configuration, we set to zero the three point function outside the triangular region: 1 - x2 _ 3 < x 2. In order to stress the difference with the case of standard ghost inflation, we plot in fig.2-2 the correspondent 3-point function for the case of ghost inflation without tilt. Note that, even though the two shapes are quite similar, the 3-point function of ghost inflation without tilt changes signs as a function of the k's, while the 3-point function in the tilted case has constant sign. An important observation is that, in the limit as 3 -- 0 and x2 - 1, which corresponds to the limit of very long and thin triangles, we have that the 3 point function goes to zero as roll inflation result - 1 . This is expected, and in contrast with the usual slow . 3 The reason for this is the same as the one which creates the same kind of behavior in the ghost inflation without tilt [2]. The limit of x 3 -- 0 * corresponds to the physical situation in which the mode k 3 exits from the horizon, freezes out much before the other two, and acts as a sort of background. In this limit, let us imagine a spatial derivative acting on 7r3, which is the background in the interaction Lagrangian. The 2-point function < 7r7r2 > depends on the position 23 1 0.8 Figure 2-1: Plot of the function F(l, X2, x3)x~x5 for the tilted ghost inflation 3-point function. The function has been normalized to have value 1 for the equilateral configuration X2 = X3 = 1, and it has been set to zero outside of the region 1 - X2 :S X3 :S X2 0.5 F(x2,x3) o 1 0.8 Figure 2-2: Plot of the similarly defined function F(l, X2, x3)x~x5 for the standard ghost inflation 3-point function. The function has been normalized to have value 1 for the equilateral configuration X2 = X3 = 1, and it has been set to zero outside of the region 1 - X2 :S X3 :S X2 [5] on the background wave, and, at linear order, will be proportional variation of the 2-point function along the 1r3 24 to 8(Tr3. The wave is averaged to zero in calculating the 3-point function < 7rk 7rk2 irk3 >, because the spatial average < 7r3 i7r3 > vanishes. So, we are forced to go to the second order, and we therefore expect to receive a factor of k, which accounts for the difference with the standard slow roll inflation case. In the model of ghost inflation, the interaction is given by derivative terms, which favors the correlation of modes freezing roughly at the same time, while the correlation is suppressed for modes of very different wavelength. The same situation occurs in standard slow roll inflation when we study non-gaussianities generated by higher derivative terms [6]. The result is in fact very similar to the one found in [6]. In that case, in fact, the interaction term could be represented as: Lint 2(_,2 + e-2Ht(Oip)2) (2.31) where one of the time derivative fields is contracted with the classical solution. This interaction gives rise to a 3-point function, which can be recast as: < k(k2(k3 >a ((ki )) ((k2 + k3)kt+ k2 +2k2k3) + cyclic)+ (2.32) 12 H f(ka)ka(k + + k3) We can easily see that the first part has the same k dependence as our tilted ghost inflation. That part is in fact due to the interaction with spatial derivative acting, and it is equal to our interaction. The integrand in the formula for the 3-point function is also evaluated with the same wave functions, so, it gives necessarily the same result as in our case. The other term is due instead to the term with three time derivatives acting. This term is not present in our model because of the spontaneous breaking of Lorentz symmetry, which makes that term more irrelevant that the one with spatial derivatives, as it is explained in [2]. This similarity could have been expected, because, adding a tilt to the ghost potential, we are converging towards standard slow roll inflation. Besides, since we have a shift symmetry for the ghost field, the interaction term which will generate the non gaussianities will be a higher 25 derivative term, as in [6]. We can give a more quantitative estimate of the similarity in the shape between our three point function and the three point functions which appear in other models. Following [5],we can define the cosine between two three point functions F1 (k1 , k 2, k 3 ), F2(kl, k2, k3), as: 2 F2)1/ (F. F cos(Fi, F2) (2.33) where the scalar product is defined as: Fl(kl, k2, k3) . F2(k, k2, k3 ) d /2 2 -x2 dx2 dx F(1,x 2 ,x 3)F2 (1,x 2 ,X 3 ) (2.34) where, as before, xi = k. The result is that the cosine between ghost inflation with tilt and ghost inflation without tilt is approximately 0.96, while the cosine with the distribution from slow roll inflation with higher derivatives is practically one. This means that a distinction between ghost inflation with tilt and slow roll inflation with higher derivative terms , just from the analysis of the shape of the 3-point function, sounds very difficult. This is not the case for distinguishing from these two models and ghost inflation without tilt. Finally, we would like to make contact with the work in [7], on the Dirac-BornInfeld (DBI) inflation. The leading interaction term in DBI inflation is, in fact, of the same kind as the one in (2.31), with the only difference being the fact that the relative normalization between the term with time derivatives acting and the one with space derivatives acting is weighted by a factor 2 = (1 - v2)- 1 , where v, is the gravity-side proper velocity of the brane whose position is the Inflaton. This relative different normalization between the two terms is in reality only apparent, since it is cancelled by the fact that the dispersion relation is w k. This implies the the relative magnitude of the term with space derivatives acting, and the one of time derivatives acting are the same, making the shape of the 3-point function in DBI inflation exactly equal to the one in slow roll inflation with higher derivative couplings, as found in [6]. 26 2.3.3 Observational Constraints We are finally able to find the observational constraints that the negative tilt in the ghost inflation potential implies. In order to match with COBE: 1 H H4 152 )43M (4.8 10-5) 2 PC = 47r233/2 = 0.018l 83/ 6 3 /4 (2.35) From this, we can get a condition for the visibility of the tilt. Remembering that 62 - al/2 _3 ( ), we get that, in order for 6 to be visible: Mh 62 62 = 208 0.018 1/263/4 5/8 /~5/8 62 > =4/5= 0.0016 2ibility 13 (2.36) In the analysis of the data (see for example [8]), it is usually assumed that the non-gaussianities come from a field redefinition: =( 3 2 - 5fNL(C- < >) (2.37) where Cgis gaussian. This pattern of non-gaussianity, which is local in real space, is characteristic of models in which the non-linearities develop outside the horizon. This happens for all models in which the fluctuations of an additional light field, different from the inflaton, contribute to the curvature perturbations we observe. In this case the non linearities come from the evolution of this field into density perturbations. Both these sources of non-linearity give non-gaussianity of the form (2.37) because they occur outside the horizon. In the data analysis, (2.37) is taken as an ansatz, and limits are therefore imposed on the scalar variable fNL. The angular dependence of the 3-point function in momentum space implied by (2.37) is given by: < (klk2k3 > = (27r)363( ki)(2W)4(--fNL PR) Hi k (2.38) In our case, the angular distribution is much more complicated than in the previous expression, so, the comparison is not straightforward. In fact, the cosine between the 27 two distributions is -0.1. We can nevertheless compare the two distributions (2.27) and (2.37) for an equilateral configuration, and define in this way an "effective" fNL for kl = k2 = k3. Using COBE normalization, we get: fNL = -2 0.29 (2.39) The present limit on non-gaussianity parameter from the WMAP collaboration [8] gives: -58 < fNL < 138 at 95% C.L. (2.40) and it implies: 62 > 0.005 (2.41) which is larger than 52ivility (which nevertheless depends on the coupling constants a,O0. Since for 62 >> usiibility we do see the effect of the tilt, we conclude that there is a minimum constraint on the tilt: 62 > 0.005. In reality, since the shape of our 3-point function is very different from the one which is represented by fNL, it is possible that an analysis done specifically for this shape of non-gaussianities may lead to an enlargement of the experimental boundaries. As it is shown in [5], an enlargement of a factor 5-6 can be expected. This would lead to a boundary on 62 of the order 62 > 0.001, which is still in the region of interest for the tilt. Most important, we can see that future improved measurements of Non Gaussianity in CMB will immediately constraint or verify an important fraction of the parameter space of this model. Finally, we remind that the tilt can be quite different from the scale invariant result of standard ghost inflation: ]n, 1< 28 - Ne -1 (2.42) 2.4 Positive Tilt In this section, we study the possibility that the tilt in the potential of the ghost is positive, V' > 0. This is quite an unusual condition, if we think to the case of the slow roll inflation. In this case, in fact, the value of H is actually increasing with time. This possibility is allowed by the fact that, on the contrary with respect to what occurs in the slow roll inflation, the motion of the field is not due to an usual potential term, but is due to a spontaneous symmetry breaking of time diffeomorphism, which gives a VEV to the velocity of the field. So, if the tilt in the potential is small enough, we expect to be no big deviance from the ordinary motion of the ghost field, as we already saw in section one. In reality, there is an important difference with respect to the case of negative tilt: a positive tilt introduces a wrong sign kinetic energy term for r. The dispersion relation, in fact, becomes: - p 2 k2 W2 = (2.43) The k2 term is instable. The situation is not so bad as it may appear, and the reason is the fact that we will consider a De Sitter universe. In fact, deep in the ultraviolet the term in k4 is going to dominate, giving a stable vacuum well inside the horizon. As momenta are redshifted, the instable term will tend to dominate. However, there is another scale entering the game, which is the freeze out scale w(k) - H. When this occurs, the evolution of the system is freezed out, and so the presence of the instable term is forgotten. So, there are two possible situations, which resemble the ones we met for the negative tilt. The first is that the term in k 2 begins to dominate after freezing out. In this situation we would not see the effect of the tilt in the wave function. The second case is when there is a phase between the ultraviolet and the freezing out in which the term in k 2 dominates. In this case, there will be an instable phase, which will make the wave function grow exponentially, until the freezing out time, when this growing will be stopped. We shall explore the phase space allowed for this scenario, 29 which occurs for 3262 > a 1/ 2 H r= a:/ H (2.44) and we restrict to it. Before actually beginning the computation, it is worth to make an observation. All the computation we are going to do could be in principle be obtained from the case of positive tilt, just doing the transformation -62 in all the results we 62 _- obtained in the former section. Unfortunately, we can not do this. In fact, in the former case, we imposed that the term in k2 dominates at freezing out, and then solved the wave equation with the initial ultraviolet vacua defined by the term in k 2 , and not by the one in k4 , as, because of adiabaticity, the field remains in the vacua well inside the horizon. On the other hand, in our present case, the term in k2 does not define a stable vacua inside the horizon, so, the proper initial vacua is given by the term in k4 which dominates well inside the horizon. This leads us to solve the full differential equation: u" + (_ 3 2k2+ a k4H2 7 2 2 2 = (2.45) Since we are not able to find an analytical solution, we address the problem with the semiclassical WKB approximation. The equation we have is a Schrodinger like eigenvalue equation, and the effective potential is: V = p2k 2 _ a k 4 H2 2 M2 + -2 r2 (2.46) Defining: 2 - /332 M 2 oh=to Mr (2.47) we have the two semiclassical regions: for << r7o,the potential can be approximated to: 4H2r2 ¢ ~ -aM M22 30 (2.48) (2.48) while, for > r70: V= d,32 k2+ 2 (2.49) 772 The semiclassical approximation tells us that the solution, in these regions, is given by: for r << 770: U (p( Al (2.50) ei fcr p(')d' while, for 7r> r70: U (p e-)crP(7)d7' (2.51) 1 /2 where P(77)= (IV(77)) The semiclassical approximation fails for r7 ro. In that case, one can match the two solution using a standard linear approximation for the potential, and gets A 2 = Al e -i 62 7r/4 [9]. It is easy to see that the semiclassical approximation is valid when 62. Let us determine our initial wave function. In the far past, we know that the solution is the one of standard ghost inflation [2]: 7r12 U=( 1 r(1) H k 2 a- (-)1/2( 77)1/2H() (2.52) r2) We can put this solution, for the remote past, in the semiclassical form, to get: 1 = U (2--k /e e (8+H e 2M 77 (2.53) (-77))1/2 So, using our relationship between Al and A 2 , we get, for r > ro, the following wave function for the ghost field: w = u/a 1 2/2 1/2ei(- i _7 * 2He-") 2H e aH (Hrek" + H (2.54) e-kr/) (2.54) Notice that this is exactly the same wave function we would get if we just rotated 5 - i in the solutions we found in the negative tilt case. But the normalization 31 would be very different, in particular missing the exponential factor, which can be large. It is precisely this exponential factor that reflects the phase of instability in the evolution of the wave function. From this observation, the results for the 2-point ant 3-point functions are immediately deduced from the case of negative tilt, paying attention to the factors coming from the different normalization constants in the wave function. So, we get: 2162M 1 e al 4P13/2H H (2.55) 4 (H Notice the exponential dependence on a, 3, HIM, and 2. The tilt gets modified, but the dominating term n-1 = V2M 2 n=V'(HV V"/ 1 2 (- + 4M + 1-'is not modified: 27r3 62M + a H2M (2.56) M 4 + 2 + ) (1 - 2P r 62 H 3 3 3H M2 MM (2 - 4P"M 8 )) For the three point function, we get: H8 < k3 ki) 433a8M8 kl(k 2 (k3 >= (27r)363(Z 1 2 k3 (,62(,Z262 (kl2(.k3) ((k2 + k 3 )kt + kt + 2k3k2) + cyclic) e 6 (2.57) H which has the same k's dependence as in the former case of negative tilt. Estimating the fNL as in the former case, we get: fNL - 0.29 6M 62 e aH (2.58) (2.58) Notice again the exponential dependence. Combining the constraints from the 2-point and 3-point functions, it is easy to see that a relevant fraction of the parameter space is already ruled out. Anyway, because of the exponential dependenceon the parameters 62,- , and the coupling constants a, and 3, which allows for big differences in the observable predictions, there are many 32 configurations that are still allowed. 2.5 Conclusions We have presented a detailed analysis of the consequences of adding a small tilt to the potential of ghost inflation. In the case of negative tilt, we see that the model represent an hybrid between ghost inflation and slow roll inflation. When the tilt is big enough to leave some signature, we see that there are some important observable differences with the original case of ghost inflation. In particular, the tilt of the 2-point function of ( is no more exactly scale invariant n, = 1, which was a strong prediction of ghost inflation. The 3-point function is different in shape, and is closer to the one due to higher deriva- tive terms in slow roll inflation. Its total magnitude tends to decrease as the tilt increases. It must be underlined that the size of these effects for a relevant fraction of the parameter space is well within experimental reach. In the case of a positive tilt to the potential, thanks to the freezing out mechanism, we are able to make sense of a theory with a wrong sign kinetic term for the fluctuations around the condensate, which would lead to an apparent instability. Consequently, we are able to construct an interesting example of an inflationary model in which H is actually increasing with time. Even though a part of the parameter space is already excluded, the model is not completely ruled out, and experiments such as WMAP and Plank will be able to further constraint the model. 33 Bibliography [1] A. H. Guth, "The Inflationary Universe: A Possible Solution To The Horizon And Flatness Problems," Phys. Rev. D 23 (1981) 347. A. D. Linde, "A New Inflationary Universe Scenario: A Possible Solution Of The Horizon, Flatness, Homogeneity, Isotropy And Primordial Monopole Problems," Phys. Lett. B 108 (1982) 389. J. M. Bardeen, P. J. Steinhardt and M. S. Turner, "Spontaneous Creation Of Almost Scale - Free Density Perturbations In An Inflationary Universe," Phys. Rev. D 28 (1983) 679. [2] N. Arkani-Hamed, P. Creminelli, S. Mukohyama and M. Zaldarriaga, "Ghost inflation," JCAP 0404 (2004) 001 [arXiv:hep-th/0312100]. [3] N. Arkani-Hamed, H. C. Cheng, M. A. Luty and S. Mukohyama, "Ghost conden- sation and a consistent infrared modification of gravity," arXiv:hep-th/0312099. [4] J. Maldacena, "Non-Gaussian features of primordial fluctuations in single field inflationary models," JHEP 0305 (2003) 013 [arXiv:astro-ph/0210603]. See also V.Acquaviva, N.Bartolo, S.Matarrese and A.Riotto,"Second-order cosmological perturbations from inflation", Nucl.Phyis. B 667, 119 (2003) [astro-ph/0207295] [5] D. Babich, P. Creminelli and M. Zaldarriaga, "The shape of non-Gaussianities," arXiv:astro-ph/0405356. [6] P. Creminelli, On non-gaussianities in single-field inflation, JCAP 0310 (2003) 003 [arXiv:astro-ph/0306122]. 34 [7] M. Alishahiha, E. Silverstein and D. Tong, "DBI in the sky," arXiv:hep- th/0404084. [8] E. Komatsu et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Tests of Gaussianity," Astrophys. J. Suppl. 148 (2003) 119 [arXiv:astro-ph/0302223]. [9] See for example: Ladau, Lifshitz, Quantum Machanics, Vol.III, Butterworth Heinemann ed. 2000. 35 Chapter 3 How heavy can the Fermions in Split Susy be? A study on Gravitino and Extradimensional LSP. In recently introduced Split Susy theories, in which the scale of Susy breaking is very high, the requirement that the relic abundance of the Lightest SuperPartner (LSP) provides the Dark Matter of the Universe leads to the prediction of fermionic superpartners around the weak scale. This is no longer obviously the case if the LSP is a hidden sector field, such as a Gravitino or an other hidden sector fermion, so, it is interesting to study this scenario. We consider the case in which the NextLightest SuperPartner (NLSP) freezes out with its thermal relic abundance, and then it decays to the LSP. We use the constraints from BBN and CMB, together with the requirement of attaining Gauge Coupling Unification and that the LSP abundance provides the Dark Matter of the Universe, to infer the allowed superpartner spectrum. As very good news for a possible detaction of Split Susy at LHC, we find that if the Gravitino is the LSP, than the only allowed NLSP has to be very purely photino like. In this case, a photino from 700 GeV to 5 TeV is allowed, which is difficult to test 36 at LHC. We also study the case where the LSP is given by a light fermion in the hidden sector which is naturally present in Susy breaking in Extra Dimensions. We find that, in this case, a generic NLSP is allowed to be in the range 1-20 TeV, while a Bino NLSP can be as light as tens of GeV. 3.1 Introduction Two are the main reasons which lead to the introduction of Low Energy Supersymmetry for the physics beyond the Standard Model: a solution of the hierarchy problem, and gauge coupling unification. The problem of the cosmological constant is usually neglected in the general treat- ment of beyond the Standard Model physics, justifying this with the assumption that its solution must come from a quantum theory of gravity. However, recently [1], in the light of the landscape picture developed by a new understanding of string theory, it has been noted that, if the cosmological constant problem is solved just by a choice of a particular vacua with the right amount of cosmological constant, the statistical weight of such a fine tuning may dominate the fine tuning necessary to keep the Higgs light. Therefore, it is in this sense reasonable to expect that the vacuum which solves the cosmological constant problem solves also the hierarchy problem. As a consequence of this, the necessity of having Susy at low energy disappears, and Susy can be broken at much higher scales (106 - 109 GeV). However, there is another important prediction of Low Energy Susy which we do not want to give up, and this is gauge coupling unification. Nevertheless, gauge coupling unification with the same precision as with the usual Minimal Supersymmetric Standard Model (MSSM) can be achieved also in the case in which Susy is broken at high scales. An example of this is the theories called Split Susy [1, 2] where there is an hierarchy between the scalar supersymmetric partners of Standard Model (SM) particles (squarks, sleptons, and so on) and the fermionic superpartners of SM particles (Gaugino, Higgsino), according to which, the scalars can be very heavy at 37 an intermediate scale of the order of 109 GeV, while the fermions can be around the weak scale. The existence for this hierarchy can be justified by requiring that the chiral symmetry protects the mass of the fermions partners. While the chiral symmetry justifies the existence of light fermions, it can not fix the mass of the fermionic partners precisely at the weak scale. As a consequence, this theory tends to make improbable the possibility of finding Susy at LHC, because in principle there could be no particles at precisely 1 TeV. In this chapter, for Split Susy, we do a study at one-loop level of the range of masses allowed by gauge coupling unification, finding that these can vary in a range that approximately goes up to 20 TeV. A possible way out from this depressing scenario comes from realizing that cosmological observations indicate the existence of Dark Matter (DM) in the universe. The standard paradigm is that the Dark Matter should be constituted by stable weakly interacting particles which are thermal relics from the initial times of the universe. The Lightest Supersymmetric Partner (LSP) in the case of conserved Rparity is stable, and, if it is weakly interacting, such as the Neutralino, it provides a perfect candidate for the DM. In particular, an actual calculation shows that in order for the LSP to provide all the DM of the universe, its mass should be very close to the TeV scale. This is the very good news for LHC we were looking for. Just to stress this result, it is the requirement the the DM is given by weakly interacting LSP that forces the fermions in Split Susy to be close to the weak scale, and accessible at LHC. In three recent papers [2, 3, 4], the predictions for DM in Split Susy were investigated, and revealed some regions in which the Neutralino can be as light as 200 GeV (Bino-Higgsino), and some others instead where it is around a 1 TeV (Pure Higgsino) or even 2 TeV (Pure Wino). As we had anticipated, all these scales are very close to one TeV, even though only the Bino-Higgsino region is very good for detection at LHC. Since the Dark Matter Observation is really the constraint that tells us if this kind of theories will be observable or not at LHC, it is worth to explore all the possibilities for DM in Split Susy. In particular, a possible and well motivated case which had been not considered in the literature, is the case in which the LSP is a very weakly 38 interacting fermion in a hidden sector. In this chapter, we will explore this possibility in the case in which the LSP is either the Gravitino, or a light weakly interacting fermion in the hidden sector which naturally appears in Extra Dimensional Susy breaking models of Split Susy [1, 5]. We will find that, if the Gravitino is the LSP, than all possible candidates for the NLSP are excluded by the combination of imposing gauge coupling unification and the constraint on hadronic decays coming from BBN. Just the requirement of having the Gravitino to provide all the Dark Matter of the univese and to still have gauge coupling unification would have allowed weakly interacting fermionic superpartneres as heavy as 5 TeV, with very bad consequences on the detactibility of Split Susy at LHC. This means that these constraints play a very big role. The only exception to this result occurs if the NLSP is very photino like, avoiding in this way the stringent constraints on hadronic decays coming from BBN. However, as we will see, already a small barionic decay branching ratio of 10- 3 is enough to rule out also this possibility. For the Extradimensional LSP, we will instead find a wide range of possibilities, with NLSP allowed to span from 30 GeV to 20 TeV. The chapter is organized as follows. In section 4.2, we study the constraints on the spectrum coming from the requirement of obtaining gauge coupling unification. In section 4.3, we briefly review the relic abundance of Dark Matter in the case the LSP is an hidden sector particle. In section 4.4, we discuss the cosmological constraints coming from BBN and CMB. In section 4.5, we show the results for Gravitino LSP. In section 4.6, we do the same for a dark sector LSP arising in extra dimensional implementation of Split Susy. In section 4.7, we draw our conclusions. 3.2 Gauge Coupling Unification Gauge coupling unification is a necessary requirement in Split Susy theories. Here we investigate at one loop level how heavy can be the fermionic supersymmetric partner for which gauge coupling unification is allowed. We will consider the Bino, Wino, and Higgsino as degenerate at a scale M 2, while we will put the Gluinos at a different 39 scale M 3 . Before actually beginning the computation, it is interesting to make an observation about the lower bound on the mass of the fermionic superpartners. Since the Bino is gauge singlet, it has no effect on one-loop gauge coupling unification. In Split Susy, with the scalar superpartners very heavy, the Bino is very weakly interacting, its only relevant vertex being the one with the light Higgs and the Higgsino. This means that, while for the other supersymmetric partners LEP gives a lower bound of - 50-100 GeV [11], for the Bino in Split Susy there is basically no lower limit. Going back to the computation of gauge coupling unification, we perform the study at -loop level. The renormalization group equations for the gauge couplings are given by: A dgi 1 dA (4ir) 2 bi(A)g 3 (3.1) where bi(A) depends of the scale, keeping truck of the different particle content of the theory according to the different scales, and i = 1, 2, 3 represent respectively /-5/3g',g, g,. We introduce two different scales for the Neutralinos, M 2, and for the Gluinos M3 , and for us M3 > M2. In the effective theory below M2 , we have the SM, which implies: 41 19 bM =(10' -- 7 -7) (3.2) Between M 2 and M3: bsplitl = ( 9 7 2' 6' 7) (3.3) Between M3 and mh,which is the scale of the scalars: 2 = ( bsplit 9 7 2' 6 , -5) (3.4) and finally, above ri we have the SSM: bssm(33 bssm 33( 1, -3) 5 40 (3.5) The way we proceed is as follows: we compute the unification scale MGUT and aGUT as deduced by the unification of the SU(2) and U(1) couplings. Starting from this, we deduce the value of as at the weak scale Mz, and we impose it to be within the 2 experimental result aS(Mz) = 0.119 ± 0.003. We use the experimental data: sin2 (Ow(MIz))= 0.23150± 0.00016and a- 1(Mz) = 128.936± 0.0049[12]. A further constraint comes from Proton decay p -+ lrOe+,which has lifetime: 8fM 2M4 T((1 + GrmP DUT F)AN) 2 T(p -- roe+) = ( MGUT 1/35\ 10 16 GeV aGUT 2 0.15GeV \ (3.6) \ 1.3 x 1035 yr CaN where we have taken the chiral Lagrangian factor (1 + D + F) and the operator renormalization A to be (1 + D + F)A - 20. For the Hadronic matrix element aN, we take the lattice result [13] acN = 0.015GeV3 . From the Super-Kamiokande limit [14], r(p - r°e+ ) > 5.3 x 1033 yr, we get: MGUT> 0.01 V3) 1/2 12GUT GUT/35 4 x 1015 GeV (3.7) An important point regards the mass thresholds of the theory. In fact, the spectrum of the theory will depend strongly on the initial condition for the masses at the supersymmetric scale im. As we will see, in particular, the Gluino mass M3 has a very important role for determining the allowed mass range for the Next-Lightest Supersymmetric Particle (NLSP), which is what we are trying to determine. In the light of this, we will consider M2 as a free parameter, with the only constraint of being smaller than mn. M 3 will be then a function of M2 and fm, and its actual value will depend on the kind of initial conditions we require. In order to cover the larger fraction of parameter space as possible, we will consider two distinct and well motivated initial conditions. First, we will require gaugino mass unification at m. This initial condition is the best motivated in the approach of Split Susy, where unification plays a fundamental role. Secondarily, we will require anomaly mediated gauigino mass initial conditions at the scale m. This second kind of initial conditions will give results 41 quite different from those of Gaugino mass unification, and, even if in this case the Gravitino can not be the NLSP, the field Ix, which will be a canditate LSP from extradimensions that we will introduce in the next sections, could be still the LSP. 3.2.1 Gaugino Mass Unification Here we study the case in which we apply gaugino mass unification at the scale 7i. In [2], a 2-loop study of the renormalization group equations for the Gaugino mass starting from this initial condition was done, and it was found that, according to mn and M2, the ratio between M 3 and M2 can vary in a range - 3 - 8. We shall use their result for M3 , as the value of M3 will have influence on the results, tending to increase the upper limit on the fermions' mass. At one loop level, we can obtain analytical results. After integration of eq.(3.1), we get the following expressions: MGUT = (e (M(b -bn 9 ) M((bMPvit1 -blm)_(bsplitl-bS)) 2 1-- 2 In M- (1b~pitl+b2m) (3.8) (38) 1pit M(bit2 split23 bMbi) valuesm 2 (bSsm +bplit2) -" turns loops out-(b that are 8w important to determbine the predicted 2 gTbItltwo g 2 effect (Mzu) (bssm+bsPt \bssm (bsPit 2 +bsplitl) of aoe(Mz). Since our main purpose is to have a rough idea of the maximum scale for the fermionic masses allowed by Gauge Coupling Unification, we proceed in the 42 following way. In [2], 2-loop gauge coupling unification was studied for M 2 = 300 GeV and 1 TeV. Since the main effect of the 2-loop contribution is to raise the predicted value of a,(Mz), we translate our predicted value of as(Mz) to match the result in [2] for the correspondent values of M 2. Having set in this way the predicted scale for a,(Mz), we check what is the upper limit on fermion masses in order to reach gauge coupling unification. The amount of translation we have to do is: 0.008. In fig.3-1, we plot the prediction for a,(Mz) for M2 = 300 GeV, 1 TeV, and 5 TeV. We see that for 5 TeV, unification becomes impossible. And so, 5 TeV is the upper limit on fermionic superpartner allowed from gauge coupling unification. Note that the role of the small difference between M3 and M2 is to raise this limit. as (Mz) U.1 0 .125 IVe 0.115 Tev 6 8 10 12 14 Log(ffi/GeV) Figure 3-1: In the case of gaugino mass unification at scale ms,we plot the unification prediction for a,(Mz). The results for M 2 = 300 GeV, 1 TeV and 5 TeV are shown. The horizontal lines represent the experimental bounds In fig.3-2, and fig.3-3,we plot the predictions for CYGUT(MGUT)and for MGUT, for the same range of masses. We see that unification is reached in the perturbative regime, with unification scale large enough to avoid proton decay limits. Note, however, that for M2 = 5 TeV, the limit is close to a possible detection. Finally, note that with this Gaugino mass initial conditions, the Wino can not be the NLSP if the Gravitino is the LSP, as shown in [2]. 43 a (UT) 0 0 0 0 Log (m/GeV) 0 0 Figure 3-2: In the case of Gaugino mass unification at scale h, we plot the prediction for ca,(MGUT). The results for M2 = 300 GeV, 1 TeV and 5 TeV are shown. "GUT - - 16 Z-1U 1.75.1016 1.5 1016 1.25 .1016 1.1016 7.5 -56 - 1 1-44, 8 10 12 14 16 Log Cn/GeV) MI imit Figure 3-3: In the case of Gaugino mass unification at scale mi, we plot the prediction for MGUT. The results for M2 = 300 GeV, 1 TeV and 5 TeV are shown, together with the lower bound on MGUTfrom Proton decay. As we will see later, a particular interesting case for the LSP in the hidden sector is given by a Bino NLSP. For this case, we need to do a more accurate computation, splitting the mass of the Gauginos, from that of the Higgsinos,and taking the Wino mass roughly two times larger than the Bino mass, as inferred from [2] for gaugino mass unification initial conditions. In fig.3-4, we show what is the allowed region for 44 the mass of the Bino and the ratio of the Hissino mass and Bino mass, such that gauge coupling unification is attained with a mass for the scalars, m, in the range 105 GeV-10 8 GeV. Raising the Higgsino mass with respect to the Bino mass has the effect of lowering the maximum mass for the fermionic superpartners. This is due to the fact that, raising the Higgsino mass, the unification value for the U(1) and SU(2) couplings is reduced, so that the prediction for a,(Mz) is lowered. MB MS Log(M/GeV) Figure 3-4: Shaded is the allowed region for the Bino mass and the ratio of the Higgsino mass and the Bino mass, in order to obtain Gauge Coupling Unification with a value of the scalar mass rh in the range 105 GeV-101 8 GeV. We take M2 ~- 2M1 as inferred from gaugino mass unification at the GUT scale [2] 3.2.2 Gaugino Mass Condition from Anomaly Mediation Of the possible initial conditions for the Gaugino mass which can have some influence on the upper bound on fermions mass, there is one which is particularly natural, and which is coming from Anomaly Mediated Susy breaking, and according to which the initial conditions for the gaugino masses are: Mi = gi 3/23/2 ' i-' M, g~ia lxm 45 (3.11) (3.11) where 3i is the beta-function for the gauge coupling, and ci is an order one number. These initial conditions are not relevant for the Gravitino LSP, as in this case the Neutralinos are lighter than the Gravitinos; but they can be relevant in the case the LSP is given by a fermion in the hidden sector, as we will study later. Further, the study of this case is interesting on its own, as it gives an upper bound on the fermionic superpartners which is higher with respect to the one coming from gaugino mass unification initial conditions. The study parallels very much what done in the former section, with the only difference being the fact that in this case, as computed in [2], the mass hierarchy between the Gluinos and the Gauginos is higher ( a factor - 10 - 20 instead of 3 - 8). This has the effect of raising the allowed mass for the fermions. We do the same amount of translation as before for the predicted a,(Mz). The result is shown in fig.3-5, and gives, as upper limit, M2 = 18 TeV. aS(M ) tx 19a V . 1 300 GeV 0.125 1TeV 0.12 1 TeV 0.115 . . . 6 I8T eV.................... ....... 8 10 12 14 Log(ffGeV) Figure 3-5: In the case of Gaugino mass condition from anomaly mediation at scale fi, we plot the unification prediction for a(Mz). The results for M2 = 300 GeV, 1 TeV and 18 TeV are shown. The horizontal lines represent the experimental bounds In fig.3-7, and fig.3-6, we plot the predictions for aGuT and MGUT for the same range of masses, and we see that unification is reached in the perturbative regime, 46 and that the unification scale is large enough to avoid proton decay limits, but it is getting very close to the experimental bound for large values of the mass i. ( MGUT ) ,.~ ,. A9- b 8 1U Log (m/GeV) J . Figure 3-6: In the case of Gaugino mass from Anomaly Mediation, we plot the prediction for CCUT.The results for M 2 = 300 GeV, 1 TeV and 18 TeV are shown. MGUT , 16 Z I'IU - 1.75.1016 18 TeV 1.5 1016 1.25 1016 1 TeV 1 1016 7.5.1015 = I :),I Z. y 45 ' 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 11.1.1.1 6 12 14 ...... 16 Log (~/GeV) MLimit Figure 3-7: In the case of Gaugino mass from Anomaly Mediation, we plot the prediction for MGUT. The results for M2 = 300 GeV, 1 TeV and 18 TeV are show, together with the lower bound on MGUT from Proton decay. As we can see, in the case of Gaugino Mass from Anomaly Mediation, the upper limit on fermion mass is raised to 18 TeV. This last one can be interpreted as a sort 47 of maximum allowed mass for fermionic superpartners. It is important to note that, as pointed out in [2], in this case the Bino can not be the NLSP. 3.3 Hidden sector LSP and Dark Matter Abundance An hidden sector LSP which is very weakly interacting can well be the DM from the astrophysical and cosmological point of view. Its present abundance can be given by two different sources: it can be a thermal relic, if in the past the temperature was so high that hidden sector particles were in equilibrium with the thermal bath, or it can be present in the universe just as the result of the decay of the other supersymmetric particles. We concentrate in the case in which the thermal relic abundance is negligible, which is generically the case for not too large reheating temperatures, and the abundance is given by the decaying of the other supersymmetric particles into the LSP. A discussion on the consequences of a thermal relic abundance of Gravitinos is discussed in [6]. In our case, the relic abundance of the heavier particles is what determines the final abundance of the LSP, and so it is the fundamental quantity to analyze. In the very early universe, the typical time scale of the cosmic evolution H - 1 is much larger than the time scale of interaction of a weakly interacting particle, and so a weakly interacting particle is in thermal equilibrium. Therefore, its abundance is given by the one of a thermal distribution. As the temperature of the universe drops down, the interaction rate is not able anymore to keep the particle in thermal equilibrium, and so the particle decouples from the thermal bath, and its density begins to dilute, ignoring the rest of the thermal bath. We say in this case that the particle species freezes out. In the case of weakly interacting particles around the TeV scale, the freeze out 48 temperature is around decades of GeV, and so they are non relativistic at the moment of freezing out. In this case, the relic abundance of these particles is given by the following formula [7, 8, 9]: (1019GeV) J (30) 0 0i5 ( 10-9GeV-2) (15 < Uv > (3.12) where < av > is the thermally averaged cross section at the time of freeze out, Xf = mNLsP Tf where Tf is the freeze out temperature, g is the effective number of degrees of freedom at freeze out, and h is the Hubble constant measured in units of 100Km/(sec Mpc). It is immediate to see that, for weakly interacting particle at 1 TeV, the resulting Q is of order unity, and this has led to the claim that the Dark Matter bounds some supersymmetric partners to be at TeV scale. In this chapter, we shall check this claim for an LSP in the hidden sector. Once the weakly interacting particles are freezed out, they will rapidly decay to the NLSP, which, being lighter, will be in general still in thermal equilibrium. So, it will be the NLSP the only one to have a relevant relic abundance, determined by the freeze out mechanism, and so it will be the NLSP that, through its decay, will generate the present abundance of the LSP. In Split Susy, the NLSP can either be the lightest Neutralino, or the lightest Chargino. The Neutralino is a mixed state of the interaction eigenstates Bino, Wino, and neutral Higgsinos, and is the lightest eigenstate of the following matrix[3]: _0kM1v kI v 0 \/- kv M2 k2 v 0 \/~ klv V8-v k1 v -/i 8(3.13) 0 which differs from the usual Neutralino matrix in low energy Susy for the Yukawa coupling, which have their Susy value at the Susy breaking scale m1,but then run differently from that scale to the weak scale. The Chargino is a mixed eigenstate of charged Higgsino and charged Wino, and 49 is the lightest eigenstate of the following matrix: (M2 kLv 2 (3.14) j/ The actual and precise computation of the thermally averaged cross section of the NLSP at freeze out, which determines the QNLSP, is very complicated, because there are many channels to take care of, which depend on the abundance of the particles involved, and on the mixing of states, creating a very complicated system of differential equations. A software called DarkSusy has been developed to reliably compute the relic abundance [10], and, in a couple of recent papers [2, 3], it has been used to compute the relic abundance of the Neutralino NLSP in Split Susy. In this kind of theories, in particular, this computation is a bit simplified, since the absence of the scalar superpartners makes many channels inefficient. Nevertheless, in most cases, the computation is still too complicated to be done analytically. In the study in this chapter, we consider both the possibility that the NLSP is a Neutralino and a Chargino. In the case of Neutralino NLSP, we modify the Dark Susy code [10] and adapt it to the case of Split Susy. We consider the cases of pure Bino, pure Wino, pure Higgsino, and Photino NLSP. In the case of Chargino NLSP, we consider the case of charged Higgsino and Charged Wino as NLSP, and we estimate their abundance with the most important diagram. We will see, in fact, that in this case a more precise determination of the relic abundance is not necessary. Once the NLSP has freezed out, it will dilute for a long time, until, at the typical time scale of 1 sec, it will decay gravitationally to the LSP, which will be stable, and will constitute today's Dark Matter. It's present abundance is connected to the NLSP "would be" present abundance by the simple relation: QLSP - L'LSP m s NLSP mNLSP (3.15) Already from this formula, we may get some important information on the masses of the particles, just comparing with the case in which the Neutralino or the Chargino 50 is the LSP. In fact, since mLsP < 1, QNLSP has to be greater than what it would have to have if the NLSP was the LSP, in order for the LSP to provide all the DM. The abundance of the NLSP is inversely proportional to < av >, and this means that we need to have a typical cross section smaller than the one we would obtain in the case of a weakly interacting particle at TeV scale. This result can be achieved in two ways: either raising the mass of the particles, since or 2, or by choosing some particle which for some reason is very low interacting. The direction in which the particle become very massive is not very attractive from the LHC detection point of view, but still, in Split Susy, is in principle an acceptable scenario. The other direction instead immediately lets a new possible candidate to emerge, which could be very attractive from the LHC detection point of view. In fact, in Split Susy, a pure Bino NLSP is almost not interacting, the only annihilation channel being the one into Higgs bosons in which an Higgsino is exchanged. In this case the relic abundance has QNLSP > 1, and this was the reason why a pure Bino could not be the DM in Split Susy [2, 3]. In the case of a gravitationally interacting LSP, as we are considering, this over abundance would go exactly into the right direction, and it could open a quite interesting region for detection at LHC. 3.4 Cosmological Constraints Since we wish the LSP to be the Dark Matter of the universe, so, we impose its abundance to cope with WMAP data [15, 16]. In general, for low reheating temperature, only the weakly interacting particles are thermally produced, and only the NLSP will remain as a thermal relic in a relevant amount. Later on, it will decay to the LSP. This decay will give the strongest cosmological constraints. In fact, concentrating on the Gravitino, which interacts only gravitationally, we 51 can naively estimate its lifetime as: r 3 MNLS (3.16) where the mNLSP term comes from dimensional analysis. In reality, we can easily do better. In fact, as we have Goldstone bosons associated to spontaneous symmetry breaking, the breaking of supersymmetry leads to the presence of a Goldstino, a massless spinor. Then, as usually occurs in gauge theories, the Goldstino is eaten by the massless Gravitino, which becomes a massive Gravitino with the right number of polarization. Therefore, the coupling of the longitudinal components of the Gravitino to the LSP will be determined by the usual pattern of spontaneous symmetry braking, and in particular will be controlled by the scale of symmetry breaking. This means that the coupling constant may be amplified. In fact, if we concentrate on the Gauginos for simplicity, we can reconstruct their coupling to the Goldstino simply by looking at the symmetry braking term in the lagrangian in unitary gauge, then reintroducing the Goldstino performing a Susy transformation, and promoting the transformation parameter to a new field, the Goldstino. The actual coupling is then obtained after canonical normalization of the Goldstino kinetic term, which is obtained performing a Susy transformation of the mass term of the Gravitino, and remembering that in the case of Sugra, the Susy transformation of the Gravitino contains a piece proportional to the vacuum energy. In formulas, the Gaugino Susy transformation is given by: X = o4"Fu, where (3.17) is the Goldstino. This implies that the mass term of the Gaugino sources the following coupling between the Gaugino and the Goldstino: 6(mAA) D m\XAu"LF,,v 52 (3.18) The Goldstino kinetic term cames from the Gravitino tranformation, which is: 6Pl = mppl + ifae (3.19) so, the Gravitino mass term produces the Goldstino kinetic term: 5(mgr,,ut, vbv>)D mgrfmpl Z = f2maucr (3.20) where in the last expression we used that mgr = f/rmpl. So, after canonical normalization, we get the following interaction term: LI = f2 .A1F.C f2 where c = (3.21) f is the canonically normalized Goldstino. After all this, we get an enhanced decay width like this: M2 mgr) rMNLSP (3.22) Note that this is independent on the particle species, as it must be by the equivalence principle [17]. Plugging in some number, we immediately see that, for particles around the TeV scale, without introducing a big hierarchy with the Gravitino, the time of decay is approximately 1-sec, and this is right the time of Big Bang Nucleosynthesis (BBN). This is the origin of the main cosmological bound. In fact, the typical decay of the LSP will be into the Gravitino and into its SM partner. The SM particle will be very energetic, especially with respect to a thermal bath which is of the order of 1 MeV, and so it will create showers of particles, which will destroy some of the existent nuclei, and enhance the formation of others, with the final result of alterating the final abundance of the light elements [17]. There is also another quantity which comes into play, and it is what we can call the "destructive power". In fact, the alteration of the light nuclei abundance will 53 be proportional to the product of the abundance of the decaying particle and to the energy release per decay. This information is synthesized in an upper limit on the variable ~ defined as: =BEY B (3.23) where B is the branching ratio for hadronic or electromagnetic decays (it turns out that hadronic decays impose constrains a couple of order of magnitude more stringent that electromagnetic decays), is the energy release per decay, and finally Y = n where n is the number of photons per comoving volume, and nx the number of decaying particles per comoving volume. Again, it is easy to see what will be the lower limit on the upper limit on 6. For the moment, we will neglect the dependence on the branching ratio B, because, clearly, one of the two branching ratios must be of order one. Then, we understand that the most dangerous particles for BBN will be those particles that decay when BBN has already produced most of the nulcei we have to see today. A particle which decays earlier than this time, will in general have its decay products diluted and thermalized with an efficiency that depends on the kind of decay product of the particle: either baryonic or electromagnetic, and it turns out that the dilution for electromagnetic decays is much more efficient. So, it is clear that the upper limit on the "desctructive power" will be lower for particles which decay after BBN. For these late decaying particles, we can estimate what the upper limit on should be with the following argument. will become dangerous if the energy release is bigger than I MeV, in order for the decay product to be able to destroy nuclei, and also if Y is greater than ni which represent the number of baryons per comoving volume opportunely normalized. Plugging in the numbers, with, again, B - 1, we get (dangerous > 10- 14 GeV (3.24) This values of dangerous is more or less where the limit seems to apply in numerical simulations for late decaying particles, and it is in fact independent on the particular kind of decay, as in this case there are not dilution issues [18, 193. For early decaying particles the limits do depend on the kind of decay, and they get more and more 54 relaxed as the decay time becomes shorter and shorter, until there is practically no limit on particles which decay earlier than 10-2 sec. Notice however that, from the estimates above on the decay time, the particles which we will be interested in will tend to decay right in the region where these limits apply. The limit in eq.(3.24) translates into another useful parameter: Qdangerous 10 - 7 (3.25) for the contribution of the NLSP around the time of nucleosynthesis. An easy computation shows that, imposing tQDM 1 today, we get that the contribution of NLSP goes as: QNLSP 1 0 -7MNLSP (3.26) (MeV This estimates are obviously very rough, but they are useful to give an idea of the physics which is going by, and they are, at the end of the day, quite accurate. They nevertheless tell us that we are really in the region in which these limits are effective, with two possible consequences: on one hand, a big part of the parameter region might be excluded, but also, on the other hand, this tells us that a possible indirect detection through deviations from the standard picture nucleosynthesis might reveal new physics. In two recent papers [18, 19], numerical simulation were implemented to determine the constraints on , both for the hadronic and the electromagnetic decays, and we shall use their data. (See also [20, 21] where a similar discussion is developed.) Cosmological constraints come also from another observable. A late decaying particle can in fact alter the thermal distribution of the photons which then will form the CMB, introducing a chemical potential in the CMB thermal distribution bigger than the one which is usually expected due to the usual cosmic evolution, or even bigger than the current experimental upper bound [22, 23]. Analytical formulas for the produced effect are given in [24, 25]. Nevertheless, it is useful to notice that the induced chemical potential is mostly and hugely dependent on the time of decay of 55 the particles. In particular, we see that: e where NLSP e (3.27) TNLSP is the lifetime of the NLSP, and rdc, 106 S is the time at which the double Compton scattering of the photons is no more efficient. The quantity is dangerous, dangerous 10 - 9 GeV. So, we conclude that basically, for there are no limits, while for rNLSP TNLSP for this < Td , > rd, the limit from nucleosynthesis is stronger. We easily see that this constraint never comes into play in our work. From formula (3.22), we can already extract an idea of what will be the final result of the analysis. In fact, we can avoid the limits from nucleosynthesis by decaying early. This means that, according to (3.22), we need to let the ratio consequently QNLSP mNLSP to grow, and has to grow as well. This leads to two directions: either a very massive LSP or a a very weakly interacting LSP. The first direction goes in agreement with one of the directions we had found in order to match the constraint from QDM, and tells us that, in general, a massive NLSP will be acceptable from the cosmological point of view. However, it will have chanches to encounter the constraints coming from gauge couplings unification. The other direction is to have an NLSP whose main annihilation channel is controlled by another particle, which can be made heavy. As an example, this is the case for the Bino, whose channel is controlled by the Higgsino: so, we might have a light Bino, if the Higgsino will be heavier. 3.5 Gravitino LSP In this section, we concentrate in detail on the possibility that the LSP in Split Susy is the Gravitino, and that it constitute the Dark Matter of the universe. We shall consider the mass of the Gravitino as a free parameter, and we shall try to extract information on the mass and the nature of the NLSP. However, an actual lower limit on the Gravitino mass can be expected in the case Susy is broken directly, as in that case the mass of the Gravitino should be: mgr 56 - M , where h is the Susy breaking MEl' scale. Since, roughly, in Split Susy in is as light as 100 TeV, we get the lower limit mgr > 10-8 GeV, which, as we will see, is lower than the region we will concentrate on. As we learnt in the former two sections, there are two fundamental quantities to be computed: the lifetime of the NLSP, and QNLSP. As we said before, we shall consider both the Neutralino and the Chargino as LSP. The decaying amplitude of a Neutralino into Gravitino plus a Standard Model particle was computed in [26, 27, 28]. For decay into Photons: where X = Nnl(-iB) + N1 2 (-iW) + Nl 3Hd + N14 H is the NLSP. As we see, eq.(3.22) reproduces the right behavior in the limit mNLSp/mLSp > 1 . This decay will contribute only to ElectroMagnetic (EM) energy. The leading contribution from hadronic decays comes from the decay into Z, gr and h, gr. These decays will contribute to EM or Hadronic energy according to the branching ratios of the SM particles. The decay width to Z boson is given by: r(X Zgr) -, - 2 Zogr)) I~i sin() N2CO(6)1 G(mwhere =Z, gr)= (X X --)mgrn(-iz) + 5x-F(mx7mgrrz) (3.29) 3 4 2 / M(-iT + 1id + N1 - 2 is see, 2\, the NLSP. As ,) we (331" 48r M m gr 2r) 21 + 3Msg) rnZ G(mx, mgr mz)) - - where F(m mgrm )= ((1-(m 9 r+MZ)t x x 57 (3.30) (1 ( mgr( -mz- x x x The decay width to h boson is given by: h, gr) (X +N s _m ~ ~ ~- N1 ~ sin(0)qM r(~~X 3 - m x )m where h =-H x (3.32) F(m, gr , mT) (3.32) mr2 x Fcos(/3)mx 14 m2 21 487 mgr ) sin(i3) + H°Cos(3) is the fine tuned light Higgs, and 2 H(mxMgrm h) =3+4 3 r+ 2 mgr+ mx m m 2 r+ gr+ m 3 mx mx 2 2\ gr+3 g mx mx (3.33) EM Bx Had F( - E 2 mx mr- 1(3.35) 2m h, gr)Bhad+ F(X --+q, q, gr) Z, g)Bad + (X , + r(x Fr(x -> gyr) x 22 2 mX + rF( h, gr) 2 m gr -m 2 Z,h Z, gr) (3.36) (3.37) and Bx is the branching ratio in the EM and Hadronic channel. We use Bhad 0.9, Bzaa 0.7. Since it will not play an important role, we just estimate the channel (X - q, , gr), and do not perform a complete computation. This channel provides the hadronic decays when mNLSP - mgr is less than the mz or ma. The leading diagram in this case is given by the tree level diagram in which there is a virtual Z boson or a virtual Higgs that decays into quarks. 3.5.1 Neutral Higgsino, Neutral Wino, and Chargino NLSP An Higgsino NLSP will be naturally muchsyinteracting in Split Susy, quite independently of the other partners mass. In fact, there are gauge interaction and Yukawa 58 coupling to the other particles. While the coupling to the Z vanishes for > mz, in that case neutral Higgsino and Charged Higgsino are almost degenerate, and so the interaction with the Higgs become relevant. This means that the annihilation rate will never be very weak, and so QNLSP will be large only for large u. An analytical computation is too complicated for our necessities, even with the simplifications of Split Susy, so, we modify the DarkSusy code [10] to adapt it to the Split Susy case, and we obtain the following relic abundance: Qoh 2 = 0.09 (T-V) (3.38) In order to avoid nucleosynthesis constraints, we need to decay early. This can be achieved either raising the hierarchy between Higgsino and Gravitino, or raising the mass of the Higgsino. Since QLSP =-mLS5pNLSP, we can not grow too much with the hierarchy, and so we are forced to raise the mass of the Higgsino. This is exactly one of the two directions to go in the parameter space we had outlined in the first sections, and it is the one which is less favourable for detection at LHC. The results of an actual computation are shown in fig.3-8, where we plot the allowed region for the Higgsino NLSP, in the plane mg,, quantity 3m 3m = mNLSP - mr. The well represents the available energy for decay, and, obviously, can not be negative. The Hadronic and the Electromagnetic constraints we use come from the numerical simulations done in [18, 19]. There, constraints are given as upper limit on the quantity = B(EM,Had)E(EM,Had)Y as a function of the time of decay. We then apply this limit to our NLSPs computing both the time of decay and the quantity 5 with the formulas given in the former section. As we had anticipated, the cosmologically allowed region is given by: mgr < 4 x 102 GeV mf > 20TeV 59 (3.39) (3.40) Log(MLsp / G eV) 4 1 0 5 5 \\ \-I\. .I.\~~~~~~~~~~~ I .I. Had -: .I. . 1. 4 \ M,=5 n TeV 4 I I ' I, Xs r \' \ > 3 % 3 I sec \\ ---- ------ '0 re m --- -- I < bD 2 2 2 -lO EM 10 D-i 1 1 DM o 0 0 · · · · 1 2 3 4 Log(MLsp/GeV) Figure 3-8: Constraints for the Higgsino NLSP, Gravitino LSP. There is no allowed region. The long dashed contour delimitats from above the excluded region by the hadronic constraints from BBN, the dash-dot-dot contour represents the same for EM constraints from BBN, the dash-dot lines represent the region within which QDM is within the experimental limits; finally, we show the short dashed countors where the Higgsino decays to Gravitino at 1 sec, 105 sec, and 1010sec, and the solid line where the Higgsino is 5 TeV heavy, which represents the upper bound for Gauge Coupling Unification. For Neutral Wino and Chargino NLSP, the result is very similar. This mass range is not allowed by gauge coupling unification, as, for Gravitino LSP, we have to use the upper bound on NLSP coming from gaugino mass unification initial conditions. So we conclude that the Higgsino NLSP is excluded. This is a very nice example of how much we can constraint physics combining particle physics data, and cosmological observations. Finally, note how the hadronic constraints raise the limit on the Higgisino mass 60 of approximately one order of magnitude. In the case of the Neutral Wino NLSP, and Chargino NLSP, there are basically no relevant differences with respect to the case of the Neutral Higgsino, the main reason being the fact that the many annihilations channels lead to an high mass for the NLSP, exactly in parallel to the case of the Higgsino. We avoid showing explicitly the results, and simply say that, for all of them, the cosmologically allowed parameter space is very similar in shape and values to the one for Higgsino, with just this slight correction on the numerical values: mgr < 5 102 GeV (3.41) mWo > 30TeV (3.42) for the Wino case. Notice that the mass is a little higher than in the Higgsino case, as the Wino is naturally more interacting. In fact, its relic abundance is given by (again, using a on porpuse modified version of Dark Susy [10] ): ,h = 0.02 wo (TeV) (3.43) For the Chargino, the mass limit is even higher: mgr < 103GeV (3.44) mW+ > 40TeV (3.45) All of these regions are excluded by the requirement of gauge coupling unification. 3.5.2 Photino NLSP As we saw in the former section, hadronic constraints pushed the mass of the NLSP different from a Bino one above the 10 TeV scale, with the resulting conflict with gauge coupling unification. Anyway, just looking at the electromagnetic constraints in fig.3-8 , one can see that a particle that will not decay hadronically at the time 61 of Nucleosynthesis will be allowed to be one order of magnitude lighter. This makes the photino, A = cos(0w)B + sin(Ow)W3 , a natural candidate to be a NLSP with Gravitino LSP. Computing the relic abundance of a Photino is rather complicated, even in Split Susy. The reason is that co-hannilation channels with the charged Winos makes a lot of diagrams allowed. In order to estimate the Photino relic abundance, we then observe that, because of the fact that the Bino is very weakly interacting in Split Susy, Photino annihilation channels will be dominated by the contribution of the channels allowed by the Wino component. We then quite reliably estimate the Photino relic abundance starting from the formula for the relic abundance of a pure Wino particle we found before: Q2WinoNLSPh = 0.02 (3.46) TeV and consider that the Photino has a Wino component equal to sin(Gw) So, for the Photino case we will have: QPhotinoNLSP (w)40.02 sini MTeV ) (M) 2 A 037TeV (34) (3.47) We then obtain the allowed region shown in fig.(3-9). The graph is very similar to the Higgsino case, with the difference that the Electromagnetic constraints are less stringent than the Hadronic ones. This allows to have the following region: 700 GeV < MA < 5 TeV (3.48) the lightest part of which might be reachable at LHC. However, already if we allow an hadronic branching ratio of 10- 3 , we see that the Photino NLSP becomes excluded. So, we conclude that a Photino NLSP is in principle allowed, but only if we fine tune it to be extremily close to a pure state of Photino. 62 o Log(MLSplGeV) 2 1 5 I " " " " I " " " " I " " " , , " /-- " " -- . -- " " " """" " " 4 " o ~ 3 ,,' ,,' i ! _ .. ' __ .~;,;.;:~ /.---"-"- ~ __ --<,:..''''-:': :'\.,\"\.- -;#'",» . J ,.,,;;;:.,.;;-;,,""",.,""., ,. lOsee 2 I ~ , - 1010s~_, __.,-.,- .'- - - - - - - - - - - - -- -- - - - -- -- - - - - -- -- -- - - - - - - - ,-' CMB " .... L _ 2 /\\ ! i DM ", "' \ \ / 1 t --------------. ---- ---- --- \.\,, _,:.:;~., ..;.~ -,;;. ~.;;;.,~ 4 I ,,' ..:. - ... - ... - ... - bO I I EM ::s -- " .._--"- ~~ .. ---.~~~~.~;~"",.,., "<:'><~"~",, MK=5TeV 10 ~, .. ' ."", " " ~ 5 I " " .3 4 3 ,., , ! i ! i !i -i .-t. - -, - - -- - - - - - - - - - - - - -- - - -. 1 . I ! i ! i ! i ! i ! i ! i ! i ~i o o o 3 1 4 Figure 3-9: Shaded is the allowed region for the Photino NLSP, Gravitino LSP. The long dashed contour delimitates from the left the region excluded by CMB, the dashdot-dot contour delimitates from above the region excluded by the EM constraints from BBN in the case Bh rv 0, the dashed-dot-dot-dot countor represents the same for Bh rv 10-3. The region within the dash-dot lines represents the region where nDM is within the experimental limits; finally, we show the short dashed contours which represent where the Photino decays to Gravitino at 1 sec, 105 sec, and 1010 sec, and the solid contour where the Photino is 5 TeV heavy, which represents the upper limit for Gauge Couplig Unification. We see that already for Bh rv 10-3 a Photino NLSP is excluded. 3.5.3 Bino NLSP Bino NLSP is a very good candidate for avoiding all the cosmological constraints. In Split Susy, a Bino NLSP is almost not interacting. 63 For a pure Bino, the only interaction which determines its relic abundance is the annihilation to Higgs bosons through the exchange of an Higgsino. Since this cross section is naturally very small, by eq.(3.12), QNLSP is very big. This means that, in order to create the right amount of DM (see eq.(3.15)), we need to make the Gravitino very light. And this is exactly what we need to do in order to avoid the nucleosynthesis bounds. We conclude then that, of the two directions to solve the DM and the nucleosynthesis problems that we outlined in the former sections, a Bino NLSP would naturally pick up the one which is the most favorable for LHC detection. However, as we saw in the section on Gauge Coupling Unification (see fig.3-4), the Higgsino can not be much heavier than the Bino. This implies that a Bino like NLSP will have to have some Higgsino component unless it is very heavy and the off diagonal terms in the mass matrix are uniportant. As a consequence, new annihilation channels opens up for the Bino NLSP through its Higgsino component. This has the effect of diminishing the relic abundance of a Bino NLSP with respect to the naive thinking we would have done if we neglected the mixing. As a result, the cosmological constraints begin to play an important role in the region of the spectrum we are interested in, and, at the end, considering the upper limit from gauge coupling unification, exclude a Bino NLSP. In order to compute the Bino NLSP relic abundance, we again modify the Dark Susy code[10]. The results are shown in fig.3-10. As anticipated, we see that the relic abundance strongly depends on the ratio beteween the Bino and the Higgsino masses M1 , . The relic abundance of the Bino NLSP becomes large enough to avoid the cosmological constaints only for so large values of the ratio between /A and M1 which are not allowed by Gauge Coupling Unification. We conclude, then, that a Bino NLSP is not allowed. 3.6 Extradimansional LSP When we consider generic possibilities to break Susy, we can have, further than the Gravitino, other fermions in a hidden sector which are kept light by an R symmetry. The implications of these fermions as being the LSP can be quite different with respect 64 to the case of Gravitino LSP, as we will see in this section. Here, we concentrate os Susy breaking in Extra Doimensions, where, as it was shown in [1], it is very generic to expect a light fermion in the hidden sector. In Susy breaking in Extra Dimension, one can break Susy with a radion field, which gets a VEV. Its fermionic component, the Goldstino, is then eaten by the Gravitino which becomes massive. Even though at tree level there is no potential, one sees that at one loop the Casimir Energy makes the radius instable. One can compensate for this introducing some Bulk Hypermultiplets, finding that, however, the cosmological constant is negative. Then, in order to cancel this, one finds that he has to introduce another source of symmetry breaking, a chiral superfield X localized on the brane (see for example [1]). This represents a sort of minimal set to break Susy in Extra Dimensions. If one protects the X field interactions with a U(1) charge, than one finds that the interactions with the SM particles are all suppressed by the 5 dimensional Plank Mass, of the form: d40 1 XtXQtQ (3.49) This induces the following mass spectrum [1]: 1rM3 Mr ;;MSra mgr N iM ~q2 ms8 where M 2 M5 M5 ;miX,A mx M9 5 (3.50) rM53 are the 4 and 5 dimensional Plank constants, and where Mi are the gaugino masses. It is quite natural to use the extradimension to lower the higher dimensional Plank mass to the GUT scale, a la Horawa-Witten [29], M 5 - MGUT ' 3x 1016 GeV. We have this range of scales[1]: mgr 1013GeV; ms - 10 9 GeV; mradion _ 107GeV; M, ,u, m, -x 100GeV (3.51) We notice that we have just reached the typical spectrum of Split Susy, in a very 65 natural way: we break Susy in Extra Dimension, stabilize the moduli, and we introduce a further Susy breaking source to compensate for the cosmological constant. We further notice that there is no much room to move the higher dimensional Plank mass M5 away from the Horawa-Witten value. In fact, the fermion mass scales as so, a slight change of M5 makes the fermions of Split Susy generically either too heavy, making them excluded by gauge coupling unification, or too light, making (-), them conflict with collider bounds. Concerning the study of the LSP, we notice that the fermionic component of the X field we have to introduce in order to cancel the cosmological constant is naturally light, of the order of the mass of the Gauginos. So, it is worth to investigate the case in which this fermion is the LSP, and how this case differs from the case in which the LSP is the Gravitino. Concerning the DM abundance, nothing changes with respect to the case of the gravitino LSP, so, we can keep the former results. Next step it is to evaluate the decay time, to check if the nucleosynthesis and CMB constraints play a role. To be concrete, let us concentrate on the Higgsino NLSP. When the Higgsino is heavier than the Higgs, the leading contribution to the decay of the Higgsino will come from the tree level diagram mediated by the operator[l]: Jd2 _20m2X m2 HUHd (3.52) 128 M 4( m2 The decay time is then given by: 7r +, mH '-M5 - __ mh \ H/ 4mHx m In the limit of mh > mrx, mh this expression simplifies to: T1 8( s(3.54) (MA 128 7r9 ft 66 M5 -1)2 (3.53) Estimating with the number we just used before, we get: T 1014 (Te) m ff sec (3.55) This time is so long before nucleosynthesis, that all the BBN constraints we found in the former case for the Gravitino now disappear. Clearly, this statment is not affected if we vary M5 in the very small window allowed by the restrictions on the fermionic superpartners' spectrum. Basically, in this mass regime, the only constraint which will apply will be the one from QDM. As we can see form the formula for the higgsino relic abundance, the region where Higgsino is lighter that Higgs is not relevant, and is excluded by the constraint on Dark Matter abundance. So, we conclude that nucleosynthesis and CMB constraints do not apply in the case in which the LSP is the field Ox, and the Higgsino is the NLSP, the only constraints which applies are the one coming from QDM and the one coming from gauge coupling unification. In fig.3-11, we show the allowed region. While the full region is quite large, and covers a rather large phase space, there is a region where the Higgsino is rather light, -- 2 TeV, and the mass of the field Ox is constrained quite precisely to be around 2 TeV. The region is bounded from above by the limit on gauge coupling unification at around 18 TeV, as in this case we must allow also for anomaly mediated initial conditions for Gauginos mass at the intermediate scale. This region is not extremely attractive for LHC. For gaugino NLSP, the situation is very similar, as the decay of the gauginos to the field Ox is mediated by the same kind of operator as for the Higgisino case [1]: d 2WW (3.56) where W is the gaugino vector supermultiplet. Clearly, again in this case, the decay time will be way before the time of BBN. In this cases, the curves that delimitate the allowed region are practically identical to the one of the Higgsino, with the only difference that the region where the NLSP is the lightest and it is practically degenerate with the Ox, correspont to an higher mass of 67 3 TeV, more difficult to see at LHC. Similarly occurs for the Bino NLSP, with the only difference that the region allowed by the Dark Matter constraint is a bit different with respect to the case of Higgsino and Wino. We obtain the allowed region shown in fig.3-12, where we see that the spectrum is very light, with Bino and Higgsino starting at tens of GeV , and gluinos at 200 GeV, with ox in the range 101 - 103 GeV . This is a very good region for LHC. Notice that the upper limit on the Bino mass is again 5 TeV, as in the case the Bino is the NLSP, we can not have anomaly mediated initial conditions for Gaugino mass at the intermediate scale. 3.7 Conclusions In Split Susy, the only two motivations to expect new physics at the TeV scale are given by the requirement that gauge coupling unification is achieved, and, mostly, that the stable LSP makes up the Dark Matter of the Universe. This is true in the standard scenario where the LSP is a neutralino. Here we have investigated the other main alternative for the LSP, that is that the LSP is constituted by a hidden sector particle. A natural candidate for this is the Gravitino, which here we studied quite in detail. Nevertheless, it is true that among the different possibilities we have in order to break Susy, one can expect the appearence in the spectrum of another light fermion protected by R symmetry. Here, as an example, we study the case of a light fermion arising in Susy breaking in Extra Dimension. The requirement to obtain gauge coupling unification limits the masses for the fermions to be less than 5 TeV or 18 TeV, according to the different initial conditions for the Gaugino masses at the the intemediate scale i. In this range of masses, we have seen how constraints from Nucleosynthesis put strong limits on the allowed region. In fact, there are two competing effects: in order to avoid Nuclosynthesis constraints, the NLSP must decay to the LSP early, and this is achieved creating a big hierarchy between the NLSP and the LSP. On the other hand, this hierarchy tends to diminish the produced 68 QLSP, and in order to compensate for it, the NLSP tends to be heavy. This goes against the constraints from gauge coupling unification. This explains why a large fraction of the parameter space is excluded. The details depend on the particular LSP and NLSP. 3.7.1 Gravitino LSP Gravitino LSP forces us to consider Gaugino Mass Unification at the intermediate scale as initial condition. This implies that we have to live with the more restrictive upper limit on fermionic masses from gauge coupling unification: 5 TeV. At the same time, the typical decay time of an NLSP to the Gravitino is at around 1 sec, and this goes exactly into the region where constraints from Nucleosynthesis on ElectroMagnetic and Hadronic decays apply. The final result is that only if the NLSP is very pure Photino like, then the Gravitino can be the LSP, with a Photino between 700 GeV and 5 TeV. If the NLSP is different by this case, then the Gravitino can not be the LSP. The reason is that a very Photino like NLSP can avoid the constraints on BBN on hadronic decays, which are much more stringent than the ones coming from electromagnetic decays, and so it can be light enough to avoid the upper limit on its mass coming from gauge coupling unification. This is very good news for the detactability of Split Susy at LHC. In fact, if the Gravitino was the LSP, than the NLSP could have been much heavier than around 1 TeV, making detection very difficult. In the study in this chapter we show that this possibility is almost excluded. 3.7.2 ExtraDimensional LSP Following the general consideration that in breaking Susy we might expect to have some fermion other than the Gravitino in the hidden sector which is kept light by an R symmetry, we have studied also the possibility that the fermionic component of a chiral field which naturally arises in Extra Dimensional Susy breaking is the LSP. In 69 this case the time of decay is so early that no Nucleosynthesis bounds apply, so, the only constraints applying are those from Dark Matter and gauge coupling unification. Concerning gauge coupling unification, in this case we must consider the possibility that the Gaugino Mass initial conditions are also those from anomaly mediation. This implies that we have to give the upper limit of 18 TeV to Fermions' mass. Having said this, the lower bound on the mass is given by the DM constraint. In fact, it is clear that QNLSP must be greater than QLSP. As a consequence, the Mass of the NLSP has to be greater than the one found in [2, 3] for the case in which these particles where the LSP. As a consequence, Charginos and Neutralinos NLSP are in general allowed, but they are restricted to be heavier than 1 TeV. It is quite interesting that, in this cases, the LSP is restricted to be in the range 100-1000 GeV. Again, there is an exception: the Bino. In this case, the LSP and NLSP are much lighter than in the case of the others Neutralinos NLSP, with an NLSP as light as a few decades of GeV. 70 Log(M LSP JOeV) -5 -4 -3 -2 -1 o 1 2 3 -5 -4 -3 -2 -1 0 1 2 3 Log(M LSP JOeV) Figure 3-10: Constraints for Bino NLSP. The long dashed countour delimitates from the left the excluded region by the hadronic constraint from BBN, the dash-dotdot contour represents the same for EM constraints from BBN, the dash-dot lines represents the ratio between the Higgsino mass and the Bino mass necessary for f2LSP to be equal to observed DM amount; the solid line represents the upper limit from GUT, while the dash-dot-dot-dot line represents the lower limit from LEP; we also show in short dashed the contours where the Bino decays to Gravitino at 1 see, 105 see, and 1010 see, and in dotted some characterstic contours for the Bino mass. CMB constraint plays no role here. We take M2 ~ 2M1, as inferred from gaugino mass unification at the GUT scale [2], and we see that no allowed region is present. 71 Log(MLSP JGeV) 1 5 3 2 "" 4 5 " ............ ...... ...... ...... ............ " ...... " ............ " ............ ...... Mn=18TeV 4 4 b OQ i: 3~ CI.l "'0 I a:: t""" CI.l / 2 )< '-" DM 1 1 o o 1 2 Log(MLSP JGeV) 3 4 Figure 3-11: Shaded is the allowed region for the Neutral Higgsino NLSP, 'l/Jx LSP. Since there are no constraints from CMB and BBN, the only constraints come from nDM, which delimitates the region within the dash-dot lines, and Gauge Coupling Unification, which set the upper bound of 18 TeV with the solid line. For Neutral Wino and Chargino NLSP the result is very similar. 72 Log(M /GeV) LSP -5 -4 -3 -2 -1 0 1 2 3 5 \ 5 \' \ \\_\ '\\ \ . ,x,~]%A . IA 4 A AY-.R \ 4 S C, 3 3 zEn 0o 2 2 2 1 Z C 1 0 0 -5 -4 -3 -2 -1 Log(M 0 1 2 3 /GeV) LSP Figure 3-12: Shaded is the allowed region for the Bino NLSP, 'x LSP. The dash-dot lines represent the ratio between the Higgsino mass and the Bino mass in order for QLSP to be equal the observed amount of DM. The dotted lines are some characteristic countors for the Bino mass. The solid line is the upper limit from GUT, while the dash-dot-dot-dot line is the lower limit from LEP. We take M2 - 2M1 . 73 Bibliography [1] N. Arkani-Hamed and S. Dimopoulos, arXiv:hep-th/0405159. [2] G. F. Giudice and A. Romanino, arXiv:hep-ph/0406088. [3] A. Pierce, arXiv:hep-ph/0406144. [4] A. Masiero, S. Profumo and P. Ullio, arXiv:hep-ph/0412058. [5] M. A. Luty and N. Okada, JHEP 0304, 050 (2003) [arXiv:hep-th/0209178]. [6] N. Arkani-Hamed, S. Dimopoulos, G. F. Giudice and A. Romanino, arXiv:hepph/0409232. [7] J. Bernstein, L. S. Brown and G. Feinberg, Phys. Rev. D 32 (1985) 3261. [8] R. J. Scherrer and M. S. Turner, Phys. Rev. D 33 (1986) 1585 [Erratum-ibid. D 34 (1986) 3263]. [9] See for example, Kolb and Turner, The Early Universe, Perseus Publishing, 1990 [10] P. Gondolo, J. Edsjo, P. Ullio, L. Bergstrom, M. Schelke and E. A. Baltz, arXiv:astro-ph/0211238. [11] S. Eidelman et al. [Particle Data Group Collaboration], Phys. Lett. B 592 (2004) 1. [12] G. Altarelli and M. W. Grunewald, arXiv:hep-ph/0404165. [13] S. Aoki et al. [JLQCD Collaboration], Phys. Rev. D 62 (2000) 014506 [arXiv:hep- lat/9911026]. 74 [14] Y. Suzuki et al. [TITAND Working Group Collaboration], arXiv:hep-ex/0110005. [15] C. L. Bennett et al., Astrophys. J. Suppl. 148 (2003) 1 [arXiv:astro-ph/0302207]. [16] D. N. Spergel et al. [WMAP Collaboration], Astrophys. J. Suppl. 148 (2003) 175 [arXiv:astro-ph/0302209]. [17] R. H. Cyburt, J. R. Ellis, B. D. Fields and K. A. Olive, Phys. Rev. D 67, 103521 (2003) [arXiv:astro-ph/0211258]. [18] M. Kawasaki, K. Kohri and T. Moroi, arXiv:astro-ph/0402490. [19] M. Kawasaki, K. Kohri and T. Moroi, arXiv:astro-ph/0408426. [20] K. Jedamzik, Phys. Rev. D 70 (2004) 083510 [arXiv:astro-ph/0405583]. [21] K. Jedamzik, Phys. Rev. D 70 (2004) 063524 [arXiv:astro-ph/0402344]. [22] J. Fixsen, E. S. Cheng, J. M. Gales, J. C. Mather, R. A. Shafer and E. L. Wright, Astrophys. J. 473 (1996) 576 [arXiv:astro-ph/9605054]. [23] K. Hagiwara et al. [Particle Data Group Collaboration], Phys. Rev. D 66 (2002) 010001. [24] W. Hu and J. Silk, Phys. Rev. Lett. 70 (1993) 2661. [25] W. Hu and J. Silk, Phys. Rev. D 48 (1993) 485. [26] J. L. Feng, S. Su and F. Takayama, arXiv:hep-ph/0404231. [27] J. L. Feng, A. Rajaraman and F. Takayama, Phys. Rev. Lett. 91 (2003) 011302 [arXiv:hep-ph/0302215]. [28] J. L. Feng, A. Rajaraman and F. Takayama, Phys. Rev. D 68 (2003) 063504 [arXiv:hep-ph/0306024]. [29] P. Horava and E. Witten, Nucl. Phys. B 460 (1996) 506 [arXiv:hep-th/9510209]. 75 Chapter 4 Hierarchy from Baryogenesis We study a recently proposed mechanism to solve the hierarchy problem in the context of the landscape, where the solution of the hierarchy problem is connected to the requirement of having baryons in our universe via Electroweak Baryogenesis. The phase transition is triggered by the fermion condensation of a new gauge sector which becomes strong at a scale A determined by dimensional transmutation, and it is mediated to the standard model by a new singlet field. In a "friendly" neighborhood of the landscape, where only the relevant operators are "scanned" among the vacua, baryogenesis is effective only if the higgs mass mh is comparable to this low scale A, forcing mh - A, and solving the hierarchy problem. A new CP violating phase is needed coupling the new singlet and the higgs field to new matter fields. We study the constraints on this model given by baryogenesis and by the electron electric dipole moment (EDM), and we briefly comment on gauge coupling unification and on dark matter relic abundance. We find that next generation experiments on the EDM will be sensitive to essentially the entire viable region of the parameter space, so that absence of a signal would effectively rule out the model. 76 4.1 Introduction The most problematic characteristic of the Standard Model (SM) seems to be the smallness of its superrenormalizable couplings, the higgs mass m', and the cosmological constant A, with respect to the apparent cutoff of the theory, the Planck mass, Mpl. These give rise respectively to the hierarchy, and cosmological constant problems. According to the naturalness hypothesis, the smallness of these operators should be understood in terms of some dynamical mechanism, and this has been a driving motivation in the last two decades in the high energy theory community. In the last few years, several things have changed. Concerning the hierarchy problem, experimental investigation has shown that already many of the theoretically most attractive possibilities for the stability of the weak scale, such as supersymmetry, begins to be disfavored, or present at least some fine tuning issues [1, 2]. Concerning the cosmological constant problem, there was the hope that some symmetry connected to quantum gravity would have forced the cosmological constant to be zero. However, first, cosmological observations have shown evidence for a non zero cosmological constant in our present universe [3, 4]; second, from the string theory point of view, two main things have occurred: on the one hand, consistent solutions with a non zero cosmological constant have been found [5, 6, 7, 8, 9], and, on the other hand, it has been becoming more and more clear that string theory appear to have a huge landscape of vacua, each one with different low energy parameters [10, 12]. If the landscape is revealed to be real, it would force a big change in the way physics has to be done, and some deep questions may find a complete new answer. In particular, it is conceivable that some characteristics of our low energy theory are just accidental consequences of the vacua in which we happen to be. This is a very big step from the kind of physics we are used to, and its consequences have been explored recently in [11]. There , it is shown that the presence of a landscape offers a new set of tools to address old questions regarding the low energy effective theory of our universe. On one hand, there are statistical arguments, according to 77 which we might explore the statistically favored characteristics for a vacuum in the landscape [12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22]. On the other hand, there are also some selection rules due to anthropic reasoning, which we might have to impose on the vacuum, in order for an observer to be present in it. Pioneering in this, is Weinberg's "structure principle" [23], which predicts the right order of magnitude for the cosmologicalconstant starting from the requirement that structures should have had time to form in our universe in order for life to be present in it. At this point, it is necessary to speak about predictivity for theories formulated with a lot of different vacua, as it occurs in the landscape. There are two important points that greatly affect the predictivity of the theory. First, we must know how the characteristic of the low energy theory change as we scan among the vacua. Second, we must know also how the probability of populate such a vacuum changes among the vacua, also considering the influence of the cosmological evolution. These are very deep questions whose answer in general requires a full knowledge of the UV theory, and we do not address them here. However, there is still something we can do. In fact, as it was pointed out in [11],theories with a large number of vacua can be described in an effective field theory approach. In such a study, it was shown that it is very natural for parameters not to effectively scan in the landscape, unless their average value in the landscape is null. So, it is reasonable to assume that some parameters in the landscape do scan, and some do not. Among the parameters of the theory, the relevant operators have a particular strong impact on the infrared properties of the theory. But exactly because of this, if they are the only ones to scan, their value can be fixed by environmental arguments. This is however true only if the marginal couplings do not scan. For this reason, a neighborhood of the landscape in which only the relevant parameters are scanning, is called a "friendly neighborhood", because it allows to fix the value of the relevant couplings with environmental reasoning, while the marginal parameters do not scan, and so they are analyzed with the usual instruments of physics, judging them on the basis of their simplicity and of their correlations. We do not know if the true string theory landscape have these properties. However, there is some phenomenological motivation to expect that this could well be. In fact, in the 78 physics connected to the standard model, we have been able to successfully address question concerning the marginal parameters with dynamical reasoning, deducing some striking features as gauge coupling unification, chiral symmetry breaking, or the weakly interacting dark matter, while, on the contrary, we have been having big troubles concerning the problems connected to the relevant coupling in the standard model, the cosmological constant, and the higgs scale, both of which arise large fine tuning issues. In this chapter, we want to address the hierarchy problem of the electroweak scale assuming we are living in such a "friendly neighborhood" of the landscape. Because of this, we concentrate a little more on the predictivity of a phenomenological model based on this. As it is clear, predictivity in the landscape is very much enhanced if there is an infrared fragile feature [11] which must be realized in the universe in order for it not to be lethal, and which is however difficult to realize in the scanned landscape. Exactly the fact that a necessary fragile feature is difficult to realize gives a lot of constraints on the possible vacua in which we should be, or, in other words, on the parameters of the theory, making then the model predictive. An example of a fragile feature is the presence of structures in the universe. According to Weinberg's argument [23], if we require the fact that in the universe there should be structures, than the value of the cosmological constant is very much tighten to a value close to the observed one. The presence of structures can then also be used to anthropically require the lightness of the higgs field in the case dark matter particles receive mass from electroweak symmetry breaking [11, 24]. In this chapter, we concentrate on another fragile feature which we consider necessary for the development of any sort of possible life in the universe, and which is a necessary completion on Weinberg's structure principle. In the absence of baryons, the dark matter would just form virilized structures, and not clumped structures, which instead are necessary for the development of life. We can construct a model where baryons are explicitly a fragile feature. Then, we can connect the solutions of the problem of baryons to the hierarchy problem with a simple mechanism which imposes a low energy electroweak scale in order for baryons to be present in the universe. This is the line of thought implemented in [11], which is the model on which we 79 focus in this chapter. As we will explain in detail in the next section, the mechanism is based on the fact that baryogenesis is possible in the model only if the electroweak scale is close to a hierarchical small scale. The hierarchical small scale is naturally introduced setting it equal to the scale at which a new gauge sector of the theory becomes strong through dimensional transmutation. This mechanism naturally pro- vides a small scale. In order to successfully implement baryogenesis, we will also need to add some more CP violation. This will force the presence of new particles and new couplings. The model becomes also predictive: the lightest of these particle will naturally be a weakly interacting particle at the weak scale, and so it will be a very good candidate for dark matter; further, the new CP violating terms will lead to the prediction of a strong electron electric dipole. The main purpose of the study in this chapter is to investigate in detail the mechanism in which baryogenesis occurs in this model, and the consequent predictions of the model. In particular, we will find that next generation experiments of the electron Electric Dipole Moment (EDM), together with the turning on of LHC, are going to explore the entire viable region of the parameter space, constraining it to such a peculiar region of the parameter space, so that absence of a signal would result in ruling out the model. The chapter is organized as follows: in sec. 5.2, we explain in detail the model; in sec. 5.3 we study the amount of baryons produced; in sec. 5.4 we determine the electron EDM; in sec. 5.5 we briefly comment on Dark Matter and Gauge Coupling Unification ; and in sec. 5.6, we draw our conclusions. 4.2 Hierarchy from Baryogenesis In this section, following [11], we show how we can connect the electroweak scale to a scale exponentially smaller than the cutoff. As the presence of baryons is a necessary condition for the formation of clumped structures, it is naturally to require that the vacuum in which we are should allow for the formation of a net baryon number in the universe. 80 This is not a trivial requirement. In fact, in the early hot phase of the universe, before the electroweak phase transition, baryon number violating interactions through weak sphalerons were unsuppressed, with a rate given approximately by [25]: F,, = 6ka5T where k - (4.1) 20, with the consequence of erasing all precedently generated baryon asymmetry (if no other macroscopic charges are present). However, at the electroweak phase transition, all the Sakharov's necessary conditions [26] for generating a baryon asymmetry are satisfied, and so it is possible in principle to generate a net baryon number at the electroweakphase transition, through a processknown in the literature as electroweak baryogenesis (see [27, 28] for two nice reviews). However, it is known that the SM electroweak phase transition can not, if unmodified, generate the right amount of baryons for two separate reasons. On one hand, CP violating interactions are insufficient [29], and, on the other hand, the phase transition is not enough first order for preventing weak sphalerons to be active also after the phase transition [30]. Since this last point is very important for our discussion, let us review it in detail. Let us suppose in some early phase of the universe we have produced some initial baryon number B and lepton number L, and no other macroscopic charge. In partic- ular B - L = 0. The quantum number B + L is anomalous, and the equation for the abundance of particles carrying B + L charge is given by: d(8 .) dtwhere nB+L = r (nB+L) (4.2) is the number of baryons and leptons per unit volume, s is the entropy density - T3 , and F is related to the sphaleron rate [27],r ,1 N Te()(), where v(T) is the temperature dependent higgs vev (v(To) = 246 GeV as measured in our vacuum), Nf is the number of fermionic families, and g92is the SU(2) weak coupling. The reason why reactions destroying baryons are faster than reactions creating them 81 is due to the fact that the relative reaction rate goes as: + -Af IL- (4.3) where Af is the difference in free energy, and F± are the sphaleron rates in the two directions. Now, it is easy to see that the free energy grows with the chemical potential AB, which then grows with the number of baryons nB. In the limit of small difference, we then get eq.(4.2). This differential equation can be integrated to give: (n (nB+L B+L ) - / final S Nfi v2mpIle ( initial (4_ v(Tc) ) g, v(Tc) (4.4) ) where Tc is the critical density, and where g, is the number of effective degrees of freedom ( 55). If in the universe there is no macroscopic lepton number, than clearly at present time we would have no baryons left, unless: Nf g* (mpl e-( 92 v(Tc) < (T,) 1 (4.5) which roughly implies the constraint: v(T) > 1 Tc (4.6) This is the so called "baryons wash out", and the origin of the requirement that the electroweak phase transition should be strongly first order. Note that this is the same condition we would get if we required sphaleron interactions not to be in thermal equilibrium at the end of the phase transition: F <H 82 (4.7) implies roughly: 4 v(Tc) 72 T, e 92S T < 2 c (4.8) and so v(T¢) > 1 (4.9) Also note that the requirement for the sphalerons not to be in thermal equilibrium already just after the phase transition is necessary, as otherwise we would have a baryon symmetric universe, which leads to a far too small residual relic density of baryons. Later on, when we shall study electroweak baryogenesis, we shall get a number for the baryon number. In order to get then the baryon number at, let us say, Big Bang Nucleosynthesis (BBN), we need to multiply that number by the factor in eq.(4.4). In the next sections, we shall consider the requirement in eq.(4.6) fulfilled, and we will consider completely negligible the wash out from the sphalerons coming from eq. (4.4). Now, let us go back to the higgs phase transition, and let us assume that in the neighborhood of the landscape in which we are, all the high energy mechanisms for producing baryons have been shut down. This is easy to imagine if, for example, the reheating temperature is smaller than the GUT scale. We are then left with the only mechanism of electroweak baryogenesis. From the former discussion, it appears clear that there should be some physics beyond the SM to help to make the phase transition strong enough. We may achieve this by coupling the higgs field to a singlet S, with the potential equal to: V = AS4 + Ah(hthwhere we have assumed a symmetry S T, JFC, S2)2 + mS2 + mhth (4.10) -S. We can couple this field to two fermions which are charged under a non-Abelian gauge group through the interaction ksSlqJc 83 (4.11) In order to preserve the symmetry S -+ -S, we give the fields i, oCcharge i. We can then assume that this sector undergos confinement and chiral symmetry breaking at its QCD scale determined by dimensional transmutation < tcV >- A3 (4.12) which is naturally exponentially smaller than the cutoff of the theory. We assume that this phase transition is first order, so that departure from thermal equilibrium is guaranteed. Now, following our discussion in the introduction, suppose that, scanning in the landscape, the only parameters which are effectively scanned are the relevant couplings mh and ms. If then we must have baryons in our universe so that clumped structures can form, than we need to be in the vacuum in which these two parameters allow for a strong enough first order phase transition in the electroweak sector. So, we have to require that this phase transition triggers the electroweak phase transition. It is clear that this can only be if it triggers a phase transition in the S field, which is possible then only if ms is of the order of A. Finally, the phase transition in S can trigger a strong first order phase transition in the higgs field only if again mh - ms - A (for a more detailed discussion, see next subsection). So, summarizing, we see that, the requirement of having baryons in the universe forces the higgs mass to be exponentially smaller than the cutoff, solving in this way the hierarchy problem. In order to produce baryons, we still need to improve the CP violating interactions, that in the SM are not strong enough. We can minimally extend the introduced model to include a singlet s and 2 SU(2) doublets 9I±, with hypercharge ±1/2 (notice that they have the same quantum numbers as higgsinos in the Minimal Supersymmetric Standard Model (MSSM)), with the following Yukawa couplings: kSss+ k'SF+'_ +ght xJ+.s +g'h'_s 84 (4.13) There is then a reparametrization invariant CP violating phase: 0 = arg(kk'g*g'*) (4.14) The mass terms for this new fields can be prohibited giving proper charges to the fields s, 'I±. This implies that, since at the electroweak phase transition these new fermions get a mass of order of the electroweak scale, and also because of the fact that the lightest fermion is stable, we actually have a nice candidate for dark matter. This last point is a connection between the higgs mass, which, up to this point, we have just assured to be exponentially smaller than the cutoff, and the weak scale, but this connection will come out quite naturally later. So, if this model happens to describe our universe, what we should see at LHC should be the higgs, the two new singlets S and s, and the two new doublets 4'±. Since the model is particularly minimal, it is interesting to explore the possibility for generating the baryon number of the universe in more detail. Before doing so, however, let us see in more detail how the vevs of the higgs and of the singlet S are changed by the phase transition. 4.2.1 Phase Transition more in detail Before going on, here we show more in detail how the requirement of having a strong first order phase transition leads to have ms - mh A. In unitary gauge, the equations to minimize the potential are (from here on, we mean by S also the vev of the field S; the meaning will be clear by the contest): 2 4AS3 - 4Ah2S(2 -S 2) -AS 2) where T is the critical temperature. + m2S + ksA 3E(T, - T) = 0 + 2m2h = 0 (4.15) (4.16) The first equation comes from the derivative with respect to S, and we will refer to it as S equation, while for the other we will use the name h equation. 85 We first consider the case > 0 and 0 m > . Before the phase transition, we have the minimum at the symmetric vacua: S = O, v = 0, for T> Tc (4.17) For T < T, the minimum conditions change, and we can not solve them analytically. We can nevertheless draw some important conclusions. Let us consider the minimum equation for S, which is the only equation which changes. Let us first consider the case ms > A. In this case, we can consistently neglect the cubic terms in the S equation, to get: S ksA 3 -2AA 2) (m (4.18) Then, if m2 > 2AAhv2 , we have: S- - ksA 3 2 (4.19) ms v= 0 (4.20) while the other solutions are still unphysical, as: v2 _mh 2 + kA <0 (4.21) ms Ah If m << 2AAhv2 , then: S- 2Aa (4.22) 2XAhv2 For the higgs, the non null solution is: v2 h Ah , ( A) which has some relevant effect for the electroweak phase transition only if But in that case v (4.23) 4A2A2v4a mh - A. A, and from mn<2< 2AAhv 2 we get that m2 <<A 2 , in contradiction with our initial assumption. Then, if mS2hv 86 < 1, in the S equation, we can consistently consider just the cubic term, to get: 1 /3 41/3((A+ AAh) (4.24) then, the non null solution for v 2 becomes: v2 2/3AA2 2 (42/3( + AAh)(A+ Ah Ah)) (4.25) which has some relevant effect only if m~h< A2 . However, in this case the condition Mr-2AAhv s2 2 <1 would imply M < 1, again in contradiction with our assumptions. So, in order to have some effect on the higgs phase transition, we are left with the only possibility of having ms A. Restricting to this, consistently, we can neglect the linear term in the S equation, to have: 4AS3 -4AhS ( 2-AS2) 2Ah AS =-ksA 3 = -2mh2 (4.26) (4.27) which implies the following equation for S: 4AS3 + 4Am2S = -ksA 3 (4.28) For mh > A, we have: ksA 3 S= 4Am2 (4.29) and the non null solution for v is still not physical: v2 mh + sA 2< (4.30) So, finally, we have that in order to have a strong first order phase transition triggered by the new sector, we are forced to have mh < A, which is what we wanted to show. In detail, for mh < A, we implement the following phase transition 87 S: 0 S - (4.31) (4X)1/3 11/21/3 v: 0 21/3 A .. A1/3 Now, the final step is to show that, not only mh < A and ms < A, but that actually mh - A and ms A. The argument for this is that, scanning in the - landscape, it will be generically much more difficult to encounter light scalar masses, as they are fine tuned. So, in the anthropically allowed range, our parameters shall most probably be in the upper part of the allowed range. So, we conclude that this model predicts mh - ms A, as we wanted to show. This phase transition must satisfy the requirement that the sphalerons are ineffective ([c) > 1), after the phase transition, which occurs at, roughly, the critical temperature T - A. So: f1/ 2 (2ks)1 /3 v(T) 11/3 - T, 1 (4.32) "d which can clearly be satisfied for some choices of the couplings. In order to better understand the natural values of the ratio v(Tc)/Tc in terms of the scalar couplings, it is worth to notice that the coupling A appears in the lagrangian as always multiplied by Ah. So, the coupling which naturally tends to be equal to the other ones in the lagrangian is A - AhA. With this redefinition, we get the constraint: v(T) T, (A )1/ 2 Ah (2 1 A r1> (4.33) (4.33) which can clearly be satisfied with some choice of the parameters. It is also worth to write the ratio of the vevs of S and h after the phase transition: -v (=A )(4.34) " We can also analyze the case in which mh < 0. In this case, the electroweak phase transition would occur before the actual strong sector phase transition. 88 However, since we know that in the SM the phase transition is neither enough first order, nor enough CP violating, we still need, in order to have baryon formation, to require that the strong sector phase transition triggers a phase transition in the higgs sector. It is then easy to see that all the former discussion still applies with tiny changes, and we get the same condition m~2,, m2 A2 . There is a further check to make, though, which is due to the fact that, in order for baryogenesis to occur, we need an unsuppressed sphaleron rate in the exterior of the bubble. This translates in the requirement, for m2 < 0: mh lnh< 1 A (4.35) which can be satisfied in some vacua. In the next of this chapter, we will always assume that these conditions are satisfied, and the sphalerons are suppressed in the broken phase. Now, we are ready to treat baryogenesis in detail. 4.3 Electroweak Baryogenesis During the electroweak phase transition, we have all the necessary conditions to fulfill baryogenesis [27, 28]. We have departure from thermal equilibrium because of the phase transition; we have CP and C violation because of the CP and C violating interactions; and finally we have baryon violation because of the unsuppressed sphaleron rate. The sphaleron rate per unit time per unit volume is in fact unsuppressed in the unbroken phase (see eq.(4.1)). There are various different effects that contribute to the final production of baryons. For example, CP violation can be due just to some CP violating Yukawa coupling in the mass matrix, or it can be mainly due to the fact that, in the presence of the wall, the mass matrix is diagonalized with space-time dependent rotation matrixes, which induce CP violation. Further, CP violation can be accounted for by some time dependent effective coupling in some interaction terms. Baryon number production as well can be treated differently. In the contest of electroweak baryogenesis which we are dealing with, there are mainly two ways of approaching the problem, one in which 89 the baryon production occurs locally where the CP violation is taking place, so called local baryogenesis, or one in which it occurs well in the exterior of the bubble, in the unbroken phase, in the so called non local baryogenesis. At the current status of the art, it appears that the non local baryogenesis is the dominant effect, at least for not very large velocity of the wall vw, which however is believed to be not large because the interactions with the plasma tend to slow down the wall considerably [31], and we concentrate on this case (see Appendix A for a brief treatment of baryogenesis in the fast wall approximation). In order to compute the produced baryon abundance, we follow a semiclassical method developed in [32]. A method based on the quantum Boltzmann equation and the closed time path integral formalism was developed in [33], making a part of the method more precise. However, the corrections given by applying this procedure to our case can be expected to be in general not very important once compared to the uncertainties associated to our poor knowledge of certain parameters of the electroweak phase transition, as it will become clear later, which make the computed final baryon abundance reliable only to approximately one order of magnitude [32]. Further, it is nice to note that our general conclusions will be quite robust under our estimated uncertainty in the computation of the baryon abundance. For these reasons, the method in [32] represents for us the right mixture between accuracy and simplicity which is in the scope of our study. Since the method is quite contorted, let us see immediately where the basic ingredients for baryogenesis are. Departure from thermal equilibrium obviously occurs because of the crossing of the wall. C violation occurs because of the V - A nature of the interactions. CP violation occurs because the CP violating phases in the mass matrix are rotated away at two different points by two different unitary matrix such that U(x)tU(x + dx) =! 1. There could be other sources of CP violation, but, in our case, this is the dominant one. Baryon production occurs instead well in the exterior of the wall, in the so called non-local baryogenesis approach, where the weak sphaleron rate is unsuppressed. Let us anticipate the general logic. The calculation is naturally split into two 90 part. In the first part, we compute the sources for CP violating charges which are due to the CP violating interactions of the particles with the incoming wall. This calculation will be done restricting ourself to the vicinity of the wall, and solving a set of coupled Dirac-Majorana equations to determine the transmission and reflection coefficients of the particles in the thermal bath when they hit the wall, which are different for particles and antiparticles. In the wall rest frame, the wall is perceived as a space-time dependent mass term. This will give rise to a CP violating current. The non null divergence of this current will be the input of the second part of the calculation. In this second part, we shall move to a larger scale, and describe the plasma in a fluid approximation, where we shall study effective diffusion equations. The key observation is that, once the charges have diffused in the unbroken phase, thermal equilibrium of the sphalerons will force a net baryon number. In fact, in the presence of SU(2) not neutral charges, the equilibrium value of the baryon number is not zero: (B + L)eq = E cQi (4.36) where ci are coefficients which depends on the different charges . Once produced, the baryon number will diffuse back in the broken phase, where, due to the suppression of the sphaleron rate, it will be practically conserved up to the present epoch. This will end our calculation. In the next two subsections, we proceed to the two parts of the outlined calcula- tion. 4.3.1 CP violation sources We begin by computing the source for the CP violating charges, following [32]. We restrict to the region very close the the wall, so that the wall can be considered flat, and we can approximately consider the problem as one dimensional. We consider a set of particles with mass matrix M(z) where z is the coordinate transverse to the wall, moving, in the rest frame of the wall, with energy-momentum E, k. Taken z 0 as the last scattering point, these particles will propagate freely for a mean free time 91 i, when they will rescatter at the point zo + Tri, where v is the velocity perpendicular to the wall, ktr/E. Now, because of the space dependent CP violating mass matrix, these particles will effectively scatter, and the probability of being transmitted and reflected will be different for particles and antiparticles. This will create a current for some charges, whose divergence will then be the source term in the diffusion equations we shall deal with in the next section. The effect will be particularly large for charges which are explicitly violated by the presence of the mass matrix, and we shall restrict to them. We introduce J as the average current resulting from particles moving towards the positive and the negative z direction, between zo0 and z0o + A, where A = T TV, and is the coherence time of the particles due to interactions with the plasma [32]. J are the CP violating currents associated with each layer of thickness A. J+ receives, for example, contributions from particles originating from the thermal bath at z0 with velocity v, and propagating until z0o+ A, as well as from particles originating at zo+ A with velocity -v, and being reflected back at zo0 + A. The formula for J+ is given by: J+ = ( (z ( (TtQT - TtQT)) Tr (Pzo+ (tQR J_= (Tr (Pzo(RQR- - RtQ)) - Tr (Pzo+A(TtQT RtQR))) (1,0,0, ) (4.37) tQT))) (1,0,0, -v) (4.38) where R(R) and T(T) are reflection and transmission matrices of particles (antiparticles) produced at zo with probability pzo, evolving towards positive z; while T and R are the correspondent quantities for particles produced at z0 + A with probability pzo+ and evolving towards negative z; v is the group velocity perpendicular to the wall at the point zo, 0 and is the same quantity at zo + A; Q is the operator corre- spondent to the chosen charge; and the trace is taken over all the relevant degrees of freedom and averaged over location z0o within a layer of thickness A. When boosted in the plasma rest frame, these currents will become the building block to construct the CP violating source for the charges that, diffusing into the unbroken phase, will let the production of baryon number possible. 92 Now, consider a small volume of the plasma, in the plasma rest frame. As the wall crosses it, it leaves a current density equal to (J+ + J-)pIasmaevery time interval T, where the subscript plasma refers to the quantity boosted in the plasma frame. So, at a time t, the total current density accumulated will be given by: t') + J(X t )))ilasma dt'-(J(I S = TR (4.39) 7 where TR is the relaxation time due to plasma interaction. From this, the rate of change of the charge Q per unit time is given by: VyQ(,t) = O,s"= ~~~~~~~~-1 1 (J+(Zt) + J-_(,t))plasma0 (J+(, t --TR)+ J (t (4.40) - 7-R))plasma -7 -j/ 'TR dz&z(J + J- )plasma Since in the SM CP violation is very small, and already proved to be not enough to account for the baryon number of the universe, we can clearly concentrate in the sector of the neutral particles of the new theory, where the mass matrix is given by: kS(z) M(z)= ( 0o k'S(z) I k'S(z) O The transmission and reflection coefficients can be found by solving the free coupled Dirac-Majorana equations for these particles with the mass matrix given in M(z). We can solve this by a method developed in [32], in a perturbative expansion in mass insertion. As it is explained in [32, 34], the small parameter in the expansion is mA, where m is the typical mass of the particles, in the case wA < 1, or m/w in the case wA > 1. In both cases, the expansion parameter is smaller than one. This can be understood noticing the analogy of our system with the scattering off a diffracting medium with a step potential of order m. In that case, reflection and transmission are comparable (and this is the only case in which we produce a net CP violating 93 charge) only if the wave packet penetrates coherently over a distance of order 1/m, and has few oscillation over that distance. Suppression of the reflection occurs both if mA < 1 and if m/w < 1. In the first case, this is because only a layer of thickness A contributes to the coherent reconstruction of the reflected wave, while in the second case, because fast oscillations tend to attenuate the reconstruction of the reflected wave. Up to sixth order in the mass insertion, we get: T = dzl 1- dzl +| J j dz2 M 2 M*e 2iw(zi-z2) + dz 2 zl o dzl j 1 dz3 '\ dZ 2 j (4.41) dz4 M4 M3*M2Me 2 iw(ZlZ2+Z34) 1Z3 1/ A dZ 3 Z5 dz5 dZ 4 M 6 M5*M 4M3*M2 M1 e 2iw(zl-z2+z3-z4+z5-z6) T= 1- dzl dz2 dz3 dz 2 "A dz - (4.42) A z2+ z3- dz 6 dz 5 - z6 ) + z4+Z5 dzMe 2iw(z)± r+ jridzl +j dziy dZ2y -| + Z4 dz 4 3 M6*M5 M4*M3 M2*Mle- 2 iw(zl- R= dzo| dzJO dz2 | d 6 dz4 M4M 3 M2*Mle-2iw(zl-z2+z3-z4)+ rz2 Ai dzl - dz dz2M*Me- 2iw(zl -z2) + dz1 + + (4.43) A 2 dZ3 M3*M 2M e i(z z2+z3)+ dz3 Z 2 dz3 | | 4'dz dz 94 4 dz5M *M4M M2M e2iw(zl- z2+Z3-z4+z5) 5 5 4 3 R dzlMle - 2iw(z) + =- j dzj + dzl j (4.44) dZ2 d 2iw(zZ2+z3) + ddZ3M3M2Mle dz2 j dz3 j dz4 j4 dz 5M 5 M4M3M2Mle- 2i(zl-z2+z 3- 4+ 5) +... where Mi = M(zi).The analogous quantities for the antiparticles are obtained replacing M -- M* in all the former formulas. We also need to have the density matrices pzo and p+.We can choose these densities as describing thermal equilibrium densities in eigenstates of the unbroken phase: Pzo= Diag(ns(E, v), n,+(E, v), n_ (E, v)) (4.45) where n(E, v) is the Fermi-dirac distribution, boosted in the wall frame: n = and pzo+ = Pzo( - -v). e 1(E-vwktr) TC (4.46) +1 The motivation of this is that particles are produced in interaction eigenstates which differ from mass eigenstates by a unitary rotation; ignoring this, amounts at ignoring small corrections of order (M(z)/Tc) 2 . The choice of thermal distribution is particularly good in the small velocity v, regime, in which we have restricted, where the non thermal contribution is of order v,, and it induces corrections of order v in the final baryon density [32]. Finally, we have to consider the charges which can play a role in generating the baryon number. When choosing such charges, one has to consider that the most important charges are those which are approximately conserved in the unbroken phase, as these are the ones which can efficiently diffuse in the unbroken phase, and induce a large generation of baryon number. Keeping this in mind, it is easy to see that the only relevant charge in our model is the "higgs number" charge, which in the same basis in which we expressed the mass matrix, for the new CP violating sector 95 particles, is given by: (4.47) Qh = Diag (O,0, 1, -1) The name "higgs number" just comes from the fact that the fields I± have the same quantum numbers as higgsinos in the MSSM. Now, we can substitute in the formulas for J+ and J_. In order to keep some analytical expression, we decide to do a derivativeexpansionM(z) = M(zo +M'(zo)(- Zo)+ (r/w)2 and v = + O(r/w)2 . This expansion is justified in the parameter range r < w, which is expected to be fulfilled [34]. approximately Unlike in the case of the analogous calculation for the MSSM [32]where the leading effect occurs at fourth order in mass insertion, here the leading contribution occurs at sixth order. We will explain later the reason of this. For the moment, we get: (4.48) J+= (1,O,O,v) x ( fz1 fA Z3 A +4 A dzl l dz2 dZ3 d3 z sin(2w(zi-4 dz 1 j z 32+ dz2 j r fZ5 dz4 ddz6 Z4 + Z5 - z 6))Im (pZoQhM 6M5*M 4M3*M 2 Mfl)+ dz3 j dz4 j f dz5 J dz6 sin(2w(zi- 2 - z 3 + z4 - Z5 + z 6))Im (poQhMlM2*M + 6M5*M 4 M3*) A A sin(2w(z - z2 + Z3 A - z4 A Z3 Z5 d + z5 -z 6))Im (Zo+aQhMM6 M5*M 4MM 2)) 96 J_ = (1,0,0, -v) x ( A A dZ2odz dzl J -4/ (4.49) A Z2 dz4 3 j A Z4 dz5 dz6 sin(2w(zl - z2 + z3 - z 4 + z5 - z6 ))Im (Pzo+aQhM6 M5 M4 M3M2M) + +4 j dzl / sin(2w(zl r -4 j z2 dz3 dz4 dZ3 dz6 dz 5 j 2 - z 3 + z4 - z 5 + z6 ))Im (pZo+AQhMlM2*M6 M5M4M) + rA dzl j A Z2 dz2 j dz 3 j A Z4 dz4 dz 5 J dz6 sin(2w(z - Z2 + z3 - z 4 + Z5 - z 6))Im (pQhMlM 6 M5M 4M3M2 )) where the z dependence in each mass matrix Mi is to be understood at linearized level. We can substitute the results in eq.(4.40), to get an expression for the higgs charge source. In reality, if we take the relaxation time large enough, and if we keep performing a derivative expansion, only the first term in eq.(4.40) is relevant. At first order in the wall velocity, we get: yQ ^- WVW T7J 7 (Z where J Jt (4.50) x 3 k f(wiA) W+d 3w 7 eEi/T (1 + eE/T,)2 Ei / is a CP violating invariant: 1 2 2 2 - g'2 ) sin(9) , = -ggkk'(g 4 97 (4.51) and fi(wA) is: f(wA) = v(zO) (z) ((S'(Zo)(9 - 24W2A 2 - 14w4 A4)v(zo) - 9v'(zo)S(zo) (4.52) 24Tj -2w 2 A(11v'(zo)A(3 + w2A2) + 3(5 + 2w2A 2)v(zo))S(Zo) + 6cos(6wA)(-S'(zo)v(zo) + v'(zo)S(zo)) + 6cos(4wA)(2S'(zo)(-1 + 3w2zA2)v(Zo) +2v'(zo)S(zo)+ w2 /(-3v'(zo)/A + v(zo))S(zo))+ 3 cos(2wA)(3S'(zo)(1 + 6W 2 A2 )V(ZO) -3v'(zo)S(Zo) + 2w2 A\(3v'(zo) A + 4V(zo))S(zo)) +2w(2S'(zo)A(-3 + 14W2A 2)v(zo)+ (5v'(zo)A(3+ 7w2A 2) + 3(1 + 7w2 A 2 )v(Zo))S(zo)) sin(2wA)3w(16S'(zo)Av(zo) +(-13v'(zo)L + v(zo))S(zo)) sin(4wA) + 6wA) (-S'(zo)v(zo) + v'(zo)S(zo))sin(6wA)))) f,(wA) (zO)3 S (ZO)(S'(Zo)(15+ 12w2A2 - 28w4 4 )v( (453o) -15v'(zo)S(Zo)- 4w2A(v'(zo)A(48 + 11w2A2 ) + 3(5 + 2w2A 2)v(zo))S(Zo) + 12cos(6wA)(S'(zo)v(zo) -v'(zo)S(zo)) + 3 cos(4wA)(S'(zo) (1 - 12w2 A 2 )V(Zo) - v'(zo)S(zo) + 4w 2 A(6v'(zo)A + v(zo))S(zoO) + 6cos(2wA(S'(zo)(-5 + 14W2A 2)v(zo) + 5v'(zo)S(zo)+ 2w2A2(5v'(zo)A+ 4v(zo))S(zo)) + 12w(S'(zo)A(-7 + 6w2 /X2)(Zo) + ((zo) + A(5v'(zo)(2 + 3w2LA2 ) + 7W2Av(zo))S(zo))sin(2wA) -6w(-9S'(zo)Av(zo) + (12v'(zo)A+ v(zo))S(zo)) sin(4wA) + 12wA(S'(zo)v(zo) -v'(zo)S(zo)) sin(6wA))) where zo here is a typical point in the middle of the wall. We can estimate the coherence time as rT (gT) -1 , and so we can take, roughly, the typical interval which is valid also in the MSSM [32]: 15 < rT 98 < 35, and in the approximation of making the particles massless, which is still consistent with our approximations, we can integrate numerically the integral, and fit it linearly in r (which is a good approximation in the range of interest), to obtain: yQ(X, t) _ 24vw-y (gg'kk'(g 2 - g, 2 ) sin(O)) x (4.54) (1 + (TTc- 25)) V3S (S) TTC T3 Even though the reparametrization invariant CP violating phase 0 requires only 4 Yukawa couplings to be present in the theory, it is somewhat puzzling that the leading effect appears at sixth order in the couplings. The reason is as follows. In the case g = g', there is an approximate Z2 symmetry which exchanges the Xi fields. This is only an approximate symmetry, because A± differ by their hypercharge. However, the hypercharge is considered a subleading effect here, and never comes into play in our computation, and so we do not see the breaking of this symmetry. The Z 2 symmetry has, in the basis we have been using up to now, the matrix form: Z2= 0 0 1 We clearly see that: [M(z), Z21 = 0, [p(Z), Z 2] = 0, {Q, Z 2 } = 0 (4.55) which then implies that: Tr (pzoTQTt) = 0 (4.56) and similar for the others terms in J+ and J_. So, in the limit in which g = g', the baryon asymmetry should vanish. Then, since each particle needs to have an even number of mass insertions to be transmitted or reflected, and since the CP violation requires at least 4 mass insertion, we finally obtain that the coupling dependence in the CP violating source must be of the form present in eq. (4.54). From this discussion, 99 it is clear that in the case g = g' the baryon production will be heavily suppressed, and, from what we will see later, it will be clear that the couplings in this case will have to be so large to necessarily hit a Landau pole at very low energies, making the model badly defined. We shall neglect this degeneracy of the couplings for the next of the chapter. Now, we are ready to begin the second part of the computation. 4.3.2 Diffusion Equations Here, we begin the second part of the calculation, still following [32]. We turn to analyze the system at a larger scale, and approximate it to a fluid. We then study the evolution of the CP violating charges due to the presence of the sources and of the diffusion effects in the plasma. To this purpose, we shall write a set of coupled differential equations which include the effects of diffusion, particle number changing reactions, and CP violating source terms, and we shall solve them to find the various densities. We shall be interested in the evolution of particles which carry some charges which are approximately conserved in the unbroken phase. Near thermal equilibrium, which is a good approximation for small velocities, we can approximate the number density as: ni = kiiT,2/6 (4.57) where pi is the chemical potential, and ki is a statistical factor which is equal to 2 for each bosonic degree of freedom, and 1 for each fermionic. The system of differential equations simplifies a lot if we neglect all couplings except the gauge couplings, the top quark Yukawa coupling, and the Yukawa couplings in the new CP violating sector. From the beginning, we take the interactions mediated from these last ones to be fast with respect to the typical timescale of the fluid. We include the effect of strong sphaleron, but neglect the one of the weak sphalerons until almost the end of the computation. This allows us to forget about leptons. We need only to keep track of the following populations: the top left doublet: Q = (tL + bL), the right top: T = tR, the Higgs particle plus our new fields I±:H = ho + + + -I 100 Strong sphalerons will be basically the only process to generate the right bottom quarks B = bR, and the quarks of the first two generations Q(1,2)L,UR, CR, SR, DR. This implies that all these abundance can be expressed in terms of the one of B: Q1L= Q2L= -2UR = -2DR = - 2 SR = -2CR = -2B = 2(Q + T) (4.58) The rate of top Yukawa interaction, Higgs violating process, and axial top number violation are indicated as ry, rh, rm, respectively. We take all the quarks to have the same diffusion equation, and the same for the higgs and the s. The charge abundances are then described by the following set of differential equations: Q= DqV2Q - ry(Q/kQ - H/kH - T/kT) - rm(Q/kQ - T/kT) (4.59) -6r,,(2Q/kQ - T/kT + 9(Q+ T)/kB) = T DqV2T - ry(-Q/kQ + H/kH + T/kT) -rm(-Q/kQ +T/kT)+ 3r.,(2Q/kQ- T/kT+ 9(Q+T)/kB) H= DhV2H - ry(-Q/kQ + T/kT + H/kH) - rhH/kh + YQ We can restrict ourselves to the vicinity of the wall, so that we can neglect the curvature of the surface, and assume we can express everything in a variable z IrF+ vt. = The resulting equations of motions become: vwQ' DqQ" - F(Q/kQ - H/kH - T/kT) - rm(Q/kQ - T/kT) (4.60) -6r,,(2Q/kQ - T/kT + 9(Q+ T)/kB) vWT'= DqT"- r(-Q/kQ + H/kH + T/kT) -rm(-Q/kQ + T/kT) + 3r,,(2Q/kQ- T/kT + 9(Q+ T)/kB) VWH'= DhH" - ry(-Q/kQ + T/kT + H/kH) - rhH/kh + YQ We now assume that ry and r,, are very fast, and we developthe result at O(1/ry, 1/r,,). This allows to algebraically express Q and T in terms of H, to get the following re- 101 lationships: Q=H( kQ(9kT-kB) '=k kH(kB + 9kQ + 9kT)) )(4.61) (4.61) and, find) substituting back,we equation for H: (4.62)differential and, substituting back, we find the following effective differential equation for H: vH' = DH"- rH +5 (4.63) where the effective couplings are given by: D = Dq(9kQkT - 2kQkB - 2kBkT) + DhkH(9kQ + 9kT + kB) 9kqkT - 2kQkB - 2kBkT + kH(9kQ + 9kT + kB) 9kkT- kH(9 kQ + 9kT + kB) ,Q99kQkT - 2kQkB -2kBkT +kH(9kQ +9kT +kB r = (rm +rh) 91cc (9kQ + 9kT + kB) 9kQkT - 2kQkB - 2kBkT + kH(9kQ+ 9kT + kB) (4.64) (4.65) (4.66) We can estimate the relaxation rates for the higgs number and the axial quark number as [32]: (m + Fh) 4M1 ,(T, z) 21g2T A (4.67) where At is the top Yukawa coupling, and Mw is the W boson mass, and, in order to keep analytical control, we approximate the source term and the relaxation term as step functions: = 5', w > > O (4.68) y = O, otherwise and r=r, F=o, 102 z0>o <0O (4.69) For the source term j, we can take the avaraged value of expression (4.54). However, due to our lack of knowledge of the details of the profiles of the fields during the phase transition, we can just approximate that expression with: yQ = 24vwyw (gg'kk'(g 2 2 ) sin(O)) x - g9' (4.70) - 25)) v4S2 rT T3 W (T, (1+ where we have taken S' . S/W and v' v/W, with W the wall width, and we have - assumed, as expected, that no cancellation is occurring. Here S and v are taken to be of the order they are today. In the approximation that D is constant, and with the boundary conditions given by H(±oo) = 0, we have an analytical solution in the unbroken phase [32]: H = AeZVw/D (4.71) where, in the limit that DI < vW, which is in general applicable, A- (4.72) (1 - e-2WV)) Note that diffusion of the higgs field, and so of the other charges, in the unbroken phase, occurs for a distance of order z D/vw. We now turn on the weak sphaleron rate, which is the responsible for the baryon generation. The baryon density follows the following equation of motion: VwPB = Dqp - E(-f)3rw,nL() (4.73) where we have assumed that the weak sphaleron operates only in the unbroken phase, and where nL is the total number density of left fermions. The solution to this equation is given by, at first order in vw: PB = - W w dznL() -oo 103 (4.74) Now, in the approximation in which all the particles in our theory are light, we have: kQ = 6, kT = 3, kB = 3, kH = 8 (4.75) and in the limit as rF --, oo, the resulting baryon abundance is zero [32]. This means that we have to go to the next order in the strong sphaleron rate expansion. Note also that, with this particle content, using the SM quark and higgs diffusion equation Dq 6/T, Dh '. 96/Tc and DFr/lv > 2/v, 110/T [35],we have D so that the assumption that the Yukawa interaction is fast is self-consistent, since ry (27/2)At2asTc [32]. Finally, we take: rF7 = 6k'3-a 4T rws = 6kaoTc, (4.76) where k' is an order one parameter and k - 20 [25]. To go to next order in the expansion in large F,,, we write: = Q - kB) ) Q=H( k( +9kT (k Substituting in (4.60), we get: H ++9k (2kQ B+ )) + kQ Q 5QI=F-v ' JQ = (DH Using nL = Q+ H (4.78) (4.78) (4.79) (1/) 3kH(9kQ + 9kt + kB)2 Q1L + Q2L = 5Q + 4T nL = (4.77) (5kQ+4kT) 6 Q, - we get: (D H -vwH) 76 = - (479) (4.80) Substituting the solution for the Higgs field, we finally get: PB __ (112 s r (A1s 104 )(1 Dq (4.81) where s = (27r2 g,/45)Tg - 55T3 . Substituting our parameters, we get: PB s 5x 10- 5 ( k,p/20 k (gg(g 2 - g'2)kk' sin(O)) 1(T - 25)) (1 + (8T/25) (T,/25) (T (4.82) (S It is worth to make a couple of small comments on the parametric dependence of this expression. The dependence of the coupling terms, and therefore on the vevs of h and S, was explained in the former section. The wall width W has simplified away because the exponential in eq.(4.72) is small and can be Taylor expanded. The presence of s,, in the denominator is due to the particular particle content of this model, for which the leading term in the expansion in 1/r,, is zero. The factor 10- 5 is mainly due to the factor of - 10-2 from the entropy density, and the ratio 20 a 5 /4, times some other factors coming from the diffusion terms. For typical values of the wall velocity, we can take approximately the SM range: v - 0.05 - 0.3, while the mean free time is -T- 20/To - 30/Tc. With these values, the produced baryon number ranges in the regime: PB (2 X 10 - - 3 x 10 - 6) (gg(g ( 2 ) _ g' 2 )kk' sin(O)) (4.83) T1 T, T,) (4.83) This is the number which has to be equal to the baryon density at BBN: = 9 (PB) S BBN X 10" (4.84) For the moment, let us try do draw some preliminary conclusions on what this does imply on the parameters of the model. Clearly, there are still too many parameters which could be varied, so, as a beginning, we can set all the Yukawa couplings in the new CP violating sector to be roughly equal to each other k - '- g g', and v/T - 1, we then get: PB , (2 x 10 S 7 - 4 x 10- 6) (g6 sin(O)) $ 105 - 1010 (4.85) We then get the following constraint: 10- 4 g 6 (-)sin(O) (4.86) It is natural, in this theory, to take the CP violating phase to be of order 1. In this case: g -g' k k' 0.21() (4.87) 1/ From this we see that the assumption of considering the interactions mediated by these Yukawa couplings to be fast is justified for a large fraction of the parameter space. These values become lower bounds for the couplings if we allow for the CP violating phase to be smaller than order 1 (see eq.(4.86)). 4.4 Electric Dipole Moment The same CP violating phase which is responsible for baryogenesis, induces an electric dipole moment through the 2-loop diagram shown in fig.4-1. f' f Figure 4-1: 2 loops diagram contributing to fermion EDM, where Xi are the neutral mass eigenstates, and w+ is the charged one. The situation here is much different than in the MSSM, where a CP violating phase in general introduces EDM at one loop level, the constraints on which generically force very small CP violating phases in the MSSM. It is instead much more similar to the case of split supersymmetry [44, 45, 47], where all the one loop diagrams contributing to the EDM are decoupled. Here the leading diagram is at two loops level, and, as we will soon see, it will induce electron EDM naturally just a little beyond the present 106 constraints, and on the edge of detection by future experiments. The induced EDM is (see [38]): . e ± 2 mf 872sM 2 (r, r, rf,) 1'dy o[1 ±mx o dzo' dy -IjZy [ (R- rf) Im (LOR*)9 (rir, M(4.8) dy (4.88) , 3(z z/2) (zyz(y+ +R) +K i) 3Ki)R + 2(Ki + R)y 4R(Ki - R) 2 Ki(Ki - 2y) Ki 2(Ki- R) 3 R (4.89) with w R = y+ ( - y)rf,, Ki = 1- M , rI OR = N3*i,O = -N 4 iC R where CR = e - i° and NTMNN 2 ° M2 f . (4.90) (4.91) = diag{mX ,MX,m,m 1 3} with real and positive diagonal elements. The plus(minus) sign on the right-hand side of eq. (6.2) corresponds to the fermion f with weak isospin +(-)1/2. and f' is its electroweak partner. We mean by w the charged mass eigenstate, and with Xi, i = 1, 2, 3, the neutral mass eigenstates in order of increasing mass. Diagonalizing the 3 x 3 mass matrix numerically, we see that generically the model predicts EDM very close to detection. The induced EDM for some generic parameters are shown in fig.4-2. It is clear then that improvements of the determination of the EDM are going to explore the most interesting region of the parameter space. Ongoing and next generation experiments plan to improve the EDM sensitivity by several orders of magnitude within a few years. For example, DeMille and his Yale group [39] will use the molecule PbO to improve the sensitivity of the electron EDM to 10-29 e cm within three years, and possibly to 10- 3 1 e cm within five years. Lamoreaux and his Los Alamos group [40] developed a solid state technique that can improve the sensitivity 107 de/(e cm) l . I 1. x 10-27 1. x 10- 28 1. x 10- 29 . - 1. x 10 - 30 100 g=9'=O.1 ..... ' ' ______________________________________ 300 500 700 1000 1500 2000 150 200 Figure 4-2: The predicted EDM in this model. We plot the induced electron EDM as a function of k'S. The solid line represents the induced EDM with g = g' = 1/2 and kS = 100 GeV, while the dashed line represents g = g' = 1/10 and kS = 100 GeV, and we take maximal CP violating phase. The horizontal line represents the present electron EDM constraint de < 1.7 x 10-27 e cm at 95% CL [37]. to the electron EDM by 104to reach 10-31 e cm. By operating at a lower temperature it is feasible to eventually reach a sensitivity of 10- 3 5 e cm, an improvement of eight orders of magnitude over the present sensitivity. The time scale for these is uncertain, as it is tied to funding prospects. Semertzidis and his Brookhaven group [41] plan to trap muons in storage rings and increase the sensitivity of their EDM measurement by five orders of magnitude. A new measurement has been presented by the Sussex group [42]. A number of other experiments aim for an improvement in sensitivity by one or two orders of magnitude, and involve nuclear EDMs. In order to understand the current and future constraints on the model, we can study what is the induced electron EDM, once we have satisfied the constraint from baryogenesis in eq.(4.83): 99(92 _ g'2 )kk' v3 S2 ,, 10- 4 (4.92) c The way we proceed is as follows. First, we fix g' = g/x/. This is a good represen- tative of the possible ratios between g and g', as it is quite far from the region where 108 the approximate symmetry in the case g = g' suppresses baryogenesis, and g' is not too small to suppress baryogenesis on its own. Later on we shall relax this condition. Having done this, we invert eq.(4.92), to get an expression for g in terms of the other parameters of the model. Now, the couplings g, g' are expressed in terms of T, and in terms of kS and k'S which, as it will be useful, in the case of small mixing, can be thought of as respectively the singlet and the doublet mass. We shall impose the constraint on the couplings g, g', k, k' to be less than - 1, in order for these Yukawa couplings not to hit a Landau pole before the unification scale 1015GeV. We decide to restrict our analysis to the case in which the possible ratios between the marginal couplings of the same sector of the theory are smaller than 2 orders of magnitude. Here, by same sector of the theory we mean either the CP violating sector, or the sector of the scalar potential and the strong gauge group. The justification of this relies in the fact that we wish to explore the most natural region of the parameter space. For this reason, we expect that there is no large hierarchy between the marginal couplings of the same sector. We think that 2 order of magnitudes is a threshold large enough to delimitate this natural region. However, since marginal couplings are radiatively relatively stable, and large hierarchies among them does not give rise to fine tuning issues, in principle, large hierarchies among the marginal couplings are acceptable. We think that, however, such a hierarchy would require the addition of further structure to the model to justify its presence, and we decide to restrict to the simplest realization of the model. As a consequence of this, the most important restrictions we apply are: 10- 2 < k/k' < 102, and 10-2 < kslA < 102. In particular, using eq.(4.33) and eq.(4.34), this implies T < S 5 T,, which, using the limit k, k' < 1, implies that the largest of kS and k'S must be smaller than 5 T. The upper bound on the CP violating sector particles tells us that there is a small part of the parameter space which we are going to explore, in which the particles responsible for the production of baryons are non relativistic. In that case, we can approximately extend the result found in eq.(4.92), with the purpose of having order of magnitude estimates, in the following way. From eq.(4.50), it is clear that baryon production is Boltzmann suppressed if the particles are non relativistic. In that case, 109 the CP violating charge will be generated by the scattering of these particle in the region where the induced mass on the particles is of the order of the critical temperature, as this is the condition of maximum CP violating interaction compatible with not being Boltzmann suppressed. Having observed this, it is easy to approximately extend the result of eq.(4.92) to the non relativistic case, by taking the vevs of the fields at the a value such that the induced mass is of the order of the critical temperature. In fig.3, we show the induced EDM as a function of kS, for several values of the critical temperature T, for k'S = 100, on top, and k'S = 500, at the bottom, with the couplings g, g' chosen as explained above, in order to fulfil the baryogenesis requirement. Notice that 500 GeV is roughly the limit that LHC will put on SU(2) doublets. We choose the maximum CP violating phase. The horizontal lines represent the present constraint on EDM de < 1.7 x 10-2 7 e cm at 95% CL [37], and the future expected one [39, 40, 41] of order 10-3 1 e cm. A few features are worth to be noted. We see that, for fixed T, the EDM decreases as we decrease kS, for kS light enough. This is due the fact that, reducing kS, we reduce both the EDM and the produced quantity of baryons. However, the loss in the production of baryons is compensated with a much smaller, compared to the decrease in kS, increase in the couplings g, g', so that, the baryon abundance can remain constant, while the EDM decreases. For large S, the EDM decreases both because the mass of the particles in the loops becomes heavier and heavier, and also because the mixing becomes more and more suppressed. The maximum is located at the point where kS k'S. In that case, in fact, the mass matrix has a diagonal piece roughly proportional to the identity, so, even though the diagonal elements are much larger than the off diagonal ones, mixing is much enhanced, and so is the EDM, which is proportional to the mixing. The lower and upper limit on the value of kS, as well as the minimum temperature, are dictated by the restrictions k'S < 5 T, kS < 5 T, and k/k' > 10- 2 . Now, as we decrease the critical temperature, the resulting EDM tends to decrease. In fact, decreasing the temperature much enhances the baryon production. This allows to decrease the couplings g, g', which explains why the induced EDM decreases. We 110 verify that increasing the hierarchy between g and g' does not change the results a lot, as the decrease in baryons production requires to make the other couplings large. Increasing the value of k'S up to the maximum allowed value of 1.1 TeV (since T.m a - v) does not produce any relevant change in the result, because this raises the minimum temperature, so that baryogenesis requires larger couplings, which forces the EDM not to decrease relevantly. It is only once the CP violating phase is lowered to less than 10-2 that a small experimentally unreachable region is opened up around k'S = 500 GeV and kS < 1 GeV. The same region is opened also enlarging the allowed hierarchy between the couplings to above 5 x 102. However, clearly, the presence at the same time of a large hierarchy and of a very small CP phase requires some additional structure on the model to explain the reason for their presence. The main conclusion we can draw from combining the analysis on baryogenesis and EDM is that, at present, the most natural region of the parameter space is perfectly allowed, however, improvements in the determination of the EDM, are going to explore the entire viable region of the parameter space, so that absence of signal, would result in effectively ruling out the model, at least if no further structure is added. 4.5 Comments on Dark Matter and Gauge Coupling Unification The proposed model finds its main motivation in stabilizing the weak scale through the requirement of attaining baryons in our universe. However, further than this, it is clear that the model provides two other interesting phenomenological aspects: gauge coupling unification and dark matter. Here, we just briefly introduce these aspects, and we postpone a more detailed discussion to future work. Thanks to the two doublets we have inserted in our model, which have the same quantum numbers as higgsinos in the MSSM, gauge coupling unification works much better than in the standard model. As it was shown in [11], gauge coupling unification 111 with the standard model plus higgsinos works roughly as well as the MSSM at two loops level, with the only possible problem being the fact that the unification scale is a bit low at around 1015 GeV. The problem from proton decay can however be avoided with some particular model at the GUT scale [11, 43]. We expect that 2-loop gauge coupling unification works quite well also in this model, with only small corrections coming from the presence of the singlet scalar S. Concerning the Dark Matter relic abundance, the lightest of the newly introduced particles is stable, and so it provides a natural candidate for Dark Matter. Estimating the relic abundance is quite complex, as it depends on the composition of the particle, and on its annihilation and coannihilations rate. However, if our newly introduced Yukawa couplings are not very close to their upper limit (of order one), and if the lightest particle is mostly composed of the two doublets, then its relic abundance is very similar to the one of pure higgsino dark matter in split supersymmetry [44, 45, 46, 47, 48, 49], requiring the doublets to be around the TeV scale. In the case of singlet dark matter, we expect the singlet to give the right abundance for a much lighter mass, as it is naturally much less interacting. However, estimates become much more difficult, and we postpone the precise determination of the relic abundance to future work. 4.6 Conclusions In this study, we have addressed the solution to the electroweak hierarchy problem in the context of the landscape, following a recent model proposed in [11]. We have shown that it is possible to connect the electroweak scale to a hierarchically small scale at which a gauge group becomes strong by dimensional transmutation. The assumption is that we are in a "friendly neighborhood" of the landscape in which only the relevant parameters of the low energy theory are effectively scanned. We then realize the model in such a way that there is a fragile, though necessary, feature of the universe which needs to be realized in our universe in order to sustain any sort of life. As a natural continuation of Weinberg's "structure principle", the fragile 112 feature is the presence of baryons in the universe, which are a necessary ingredient for the formation of clumped structures. We assume that, in the friendly neighborhood of the landscape in which we should be, baryogenesis is possible only through the mechanism of electroweak baryogenesis. Then, in order to produce a first order phase transition strong enough so that sphalerons do not wash out the produced baryon density, we develop a new mechanism to implement the electroweak phase transition. We introduce a new gauge sector which becomes strong at an exponentially small scale through dimensional transmutation. We couple this new sector to a singlet S which is then coupled to the higgs field. The electroweak phase transition occurs as the new gauge sector becomes strong, and produces a chiral condensation. This triggers a phase transition for the singlet S, which then triggers the phase transition for the higgs fields. In order to preserve baryon number, we need the phase transition to be strong enough, and this is true only if the higgs mass is comparable to the QCD scale of the strong sector. This solves the hierarchy problem. In order to provide the necessary CP violation, we introduce 2 SU(2) doublets t with hypercharge ±1/2, and a gauge singlet s. When we require the model to describe our world, the model leads to falsifiable predictions. The main result of the study in this chapter is the computation of the produced baryon number, for which we obtain: PB_ (2 x 10 3 x100 6) (9gg(g2 g'2 )kk' sin(0)) () (S (4.93) The requirement that this baryon abundance should cope with the observed one leads to a lower bound on a combination of the product of the CP violating phase, the new couplings, the S vev, and the critical temperature Tc. We infer that Gauge Coupling Unification, and the right amount of Dark Matter relic abundance are easily achieved in this model. We study in detail the induced electron Electric Dipole Moment (EDM), and we find that, at present, the most natural region of the parameter space of the theory is 113 allowed. However, soon in the future improvement in the EDM experiments will be sensitive the entire viable region of the parameter space of the model, so that absence of a signal would result in practically ruling out the model. 4.7 Appendix: Baryogenesis in the Large Velocity Approximation The method we have used in the main part of this work is based on some approximations that fail in limit of very fast wall speed: v, 1. Even though this regime seems to be disfavored by actual computations of the wall speed[31], it is worth to try to estimate the result even in this case. In this appendix we are going to do an approximate computation in the large wall velocity approximation. In this new regime, calculations become much more complicated, as the assumption of local thermodynamical equilibrium begins to fail, and the baryon number tends to be produced in the region of the wall, in the so called local baryogenesis scenario. We follow the treatment of [51]. In this approach, the phase transition is treated from an effective field theory point of view. A good way of looking at the high temperature phase of the unbroken phase is imagining it as discretized in a lot of cells, whose side is given by the typical size of the weak sphaleron barrier crossing configuration of the gauge field: (awT) - 1. Concentrating on each cell, we have that the thermal energy in a cell is in general much larger than the energy necessary to create a gauge field oscillation capable of crossing the barrier, and this means that most of the energy is in oscillations with smaller wavelength than (. So, these configurations cross the barrier at energy far above the one of the sphaleron configuration, and so their rate has nothing to do with the sphaleron rate, and this explain why their rate per unit volume is of order 4 (In the case m2 < 0, electroweak symmetry breaking has already occurred outside the bubble, but this discussion still roughly applies). We can parameterize the configuration in one cell with one variable T, which depends only on time. Obviously, this is a very rough approximation, but 114 this is at least a beginning. We try to describe the dynamics of the configuration near crossing the barrier, at a maximum of the energy, that we fixe to be at r = 0. Near this point, we can write the following Lagrangian, which is very similar to the one in [51]: L(T, *)= cl 2 + C2 2C 2 + C3b (4.94) 2(3 where ci are dimensionless parameters depending on the different possible barrier crossing trajectories, while b will be determined shortly. Since the point r = 0 is a maximum of the energy, no odd powers of r can appear. The appearance of the odd power in i can be understood because, at one loop level, the CP violating mass matrix of the particles in the CP violating sector induces an operator which contains the term TrF,,F/". This can be seen in the following way, in an argument similar to the one shown in [47]. We can imagine to do a chiral rotation on these field of amplitude equal to the CP violating phase. Because of the anomaly, this will induce the operator: (4.95) = 16w 9222TrFFWv where g2 is the SU(2) weak coupling, and F,, is the SU(2) field strength, and where v 2 sin()) Arg(Det(M)) = Arcsin (kk'S + (gg kkIS2sin()2 + (gg'v2 + kkS2Cos())2 (4.96) Promoting the vevs of h and S to the actual fields, we find the operator we were look- ing for. Now, the term TrF,,vFi contains a term proportional to the time derivative of the Chern Simons number, and so the operator action (4.94) with a term proportional to b, must contribute to the effective with b = g 16r2 , explaining the reason for the presence of the term in i The equation of motion for r is: = C2_ C1 2 27-- 115 C3 92g4 c 1 16wr2 (4.97) If the wall is fast enough,we can solve it in the impulse approximation, to get: nAi = c 3 cl 16w2 g (4.98) This kick to A-? makes the distribution of velocities of barrier crossing configurations asymmetric, leading to a production of baryons with respect to antibaryons. Now, this kick will be very inefficient in changing the distribution of baryons, unless the kick is larger than the typical speed o a generic configuration would have if it crossed the barrier in the absence of the wall. So, we require: A > '0. The fraction of configurations which satisfy this requirement is proportional to Ai, but is very difficult to estimate. Following [51], we just say that it is equal to f A, where f is our "ignorance" coefficient. So, we finally get: nB f -( 3 (4.99) where we have reabsorbed the constants ci into f. We finally get: PB _afW3 ss f a4455 92 6 22 AO 167r (4.100) There are a lot of heavy approximations which suggest that this estimate is very rough [51]: from the value of the coefficient ci, to the approximation of restricting to one degree of freedom, to the impulse approximation in solving the differential equation, and to the estimate of the fraction of configurations influenced by the kick. All these suggests that we should take eq.(4.100) with f baryon production, as the authors of [51] suggest. 116 1, as at most an upper limit on the de/(e cm) 1. x 10-27 _ I 7 - 1. x 10 28 1. x 10- 29 1. x 1030 1. x10 - 31 . T=25 GeV . -5 10 1 S 50 100 500 1000 de/(e cm) 1. x 10- 27 i k'S=500 GeV 1. x 10 28 1.xl 0-29 T=200 GeV 1.x 1 0-30 1.x 10- 31 I, c K 10 20 50 100 200 500 1000 Figure 4-3: Top, the predicted EDM given the constraint from Baryogenesis satisfied, as a function of kS for kS = 100 GeV, g' = gl/, maximum CP violating phase, and T, = 200 GeV (solid line), 100 GeV (long dashed), 25 GeV (short dashed). Bottom, the same for k'S = 500 GeV, and T = 200 GeV (solid line), 150 GeV (long dashed), 110 GeV (short dashed). The horizontal lines represent the present electron EDM constraint de < 1.7 x 10-2 7 e cm at 95% CL [37], and the expected improvement up to de < 103 1e cm [39, 40, 41]. 117 Bibliography [1] R. Barbieri and A. Strumia, "What is the limit on the Higgs mass?," Phys. Lett. B 462 (1999) 144, hep-ph/9905281. [2] L. Giusti, A. Romanino and A. Strumia, "Natural ranges of supersymmetric signals," Nucl. Phys. B 550 (1999) 3, hep-ph/9811386. [3] D. N. Spergel et al. [WMAP Collaboration], "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Determination of Cosmological Parameters," Astrophys. J. Suppl. 148 (2003) 175, astro-ph/0302209. [4] S. Perlmutter et al. [Supernova Cosmology Project Collaboration], "Measurements of Omega and Lambda from 42 High-Redshift Supernovae," Astrophys. J. 517 (1999) 565, astro-ph/9812133. [5] R. Bousso and J. Polchinski, "Quantization of four-form fluxes and dynamical neutralization of the cosmological constant," JHEP 0006 (2000) 006, hep-th/0004134. [6] S. B. Giddings, S. Kachru and J. Polchinski, "Hierarchies from fluxes in string compactifications," Phys. Rev. D 66 (2002) 106006, hep-th/0105097. [7] G. Curio and A. Krause, "G-fluxes and non-perturbative stabilisation of heterotic M-theory," Nucl. Phys. B 643 (2002) 131, hep-th/0108220. [8] A. Maloney, E. Silverstein and A. Strominger, "De Sitter space in noncritical string theory," hep-th/0205316. [9] S. Kachru, R. Kallosh, A. Linde and S. P. Trivedi, "De Sitter vacua in string theory," Phys. Rev. D 68, 046005 (2003), hep-th/0301240. 118 [10] L. Susskind, "The anthropic landscape of string theory," hep-th/0302219. [11] N. Arkani-Hamed, S. Dimopoulos and S. Kachru, "Predictive landscapes and new physics at a TeV," hep-th/0501082. [12] M. R. Douglas, "The statistics of string / M theory vacua," JHEP 0305 (2003) 046, hep-th/0303194. [13] G. Dvali and A. Vilenkin, "Cosmic attractors and gauge hierarchy," Phys. Rev. D 70 (2004) 063501, hep-th/0304043. [14] S. Ashok and M. R. Douglas, "Counting flux vacua," JHEP 0401 (2004) 060, hep-th/0307049. [15] F. Denef and M. R. Douglas, "Distributions of flux vacua," JHEP 0405 (2004) 072, hep-th/0404116. [16] A. Giryavets, S. Kachru and P. K. Tripathy, "On the taxonomy of flux vacua," JHEP 0408 (2004) 002, hep-th/0404243. [17] J. P. Conlon and F. Quevedo, "On the explicit construction and statistics of Calabi-Yau flux vacua," JHEP 0410 (2004) 039, hep-th/0409215. [18] 0. DeWolfe, A. Giryavets, S. Kachru and W. Taylor, "Enumerating flux vacua with enhanced symmetries," JHEP 0502 (2005) 037, hep-th/0411061. [19] J. Kumar and J. D. Wells, "Landscape cartography: A coarse survey of gauge group rank and stabilization of the proton," Phys. Rev. D 71 (2005) 026009, hep-th/0409218. [20] G. Dvali, "Large hierarchies from attractor vacua," hep-th/0410286. [21] R. Blumenhagen, F. Gmeiner, G. Honecker, D. Lust and T. Weigand, "The statistics of supersymmetric D-brane models," Nucl. Phys. B 713 (2005) 83, hep- th/0411173. 119 [22] F. Denef and M. R. Douglas, "Distributions of nonsupersymmetric flux vacua," JHEP 0503 (2005) 061, hep-th/0411183. [23] S. Weinberg, "Anthropic Bound On The Cosmological Constant," Phys. Rev. Lett. 59 (1987) 2607. [24] P. C. Schuster and N. Toro, "Is dark matter heavy because of electroweak symmetry breaking? Revisiting heavy neutrinos," hep-ph/0506079. [25] D. Bodeker, Phys. Lett. B426 (1998) 351, hep-ph/9801430; P. Arnold and L.G. Yaffe, hep-ph/9912306; P. Arnold, Phys. Rev. D62 (2000) 036003; G.D. Moore and K. Rummukainen, Phys. Rev. D61 (2000) 105008; G.D. Moore, Phys. Rev. D62 (2000) 085011. [26] A. D. Sakharov, "Violation Of CP Invariance, C Asymmetry, And Baryon Asymmetry Of The Universe," Pisma Zh. Eksp. Teor. Fiz. 5 (1967) 32 [JETP Lett. 5 (1967 SOPUA,34,392-393.1991 UFNAA,161,61-64.1991) 24]. [27] A. Riotto, "Theories of baryogenesis," hep-ph/9807454. [28] A. Riotto and M. Trodden, "Recent progress in baryogenesis," Ann. Rev. Nucl. Part. Sci. 49 (1999) 35, hep-ph/9901362. [29] M. B. Gavela, P. Hernandez, J. Orloff, O. Pene and C. Quimbay, "Standard model CP violation and baryon asymmetry. Part 2: Finite temperature," Nucl. Phys. B 430, 382 (1994), hep-ph/9406289. [30] K. Rummukainen, M. Tsypin, K. Kajantie, M. Laine and M. E. Shaposhnikov, "The universality class of the electroweak theory," Nucl. Phys. B 532, 283 (1998), hep-lat/9805013. [31] M. Dine, R. G. Leigh, P. Y. Huet, A. D. Linde and D. A. Linde, "Towards the theory of the electroweak phase transition," Phys. Rev. D 46 (1992) 550, hep-ph/9203203. 120 [32] P. Huet and A. E. Nelson, "Electroweak baryogenesis in supersymmetric models," Phys. Rev. D 53 (1996) 4578, hep-ph/9506477. [33] A. Riotto, "Towards a Nonequilibrium Quantum Field Theory Approach to Electroweak Baryogenesis," Phys. Rev. D 53 (1996) 5834, hep-ph/9510271; A. Riotto, "Supersymmetric electroweak baryogenesis, nonequilibrium field theory and quantum Boltzmann equations," Nucl. Phys. B 518 (1998) 339, hep-ph/9712221; A. Riotto, "The more relaxed supersymmetric electroweak baryogenesis," Phys. Rev. D 58 (1998) 095009, hep-ph/9803357. [34] P. Huet and E. Sather, "Electroweak baryogenesis and standard model CP violation," Phys. Rev. D 51, 379 (1995), hep-ph/9404302. [35] M. Joyce, T. Prokopec and N. Turok, "Nonlocal electroweak baryogenesis. Part 1: Thin wall regime," Phys. Rev. D 53 (1996) 2930, hep-ph/9410281. [36] X. Zhang and B. L. Young, "Effective Lagrangian approach to electroweak baryogenesis: Higgs mass limit and electric dipole moments of fermion," Phys. Rev. D 49 (1994) 563, hep-ph/9309269; [37] B. C. Regan, E. D. Commins, C. J. Schmidt and D. DeMille, "New limit on the electron electric dipole moment," Phys. Rev. Lett. 88 (2002) 071805. [38] D. Chang, W. F. Chang and W. Y. Keung, "Electric dipole moment in the split supersymmetry models," Phys. Rev. D 71 (2005) 076006, hep-ph/0503055. [39] D. Kawall, F. Bay, S. Bickman, Y. Jiang and D. DeMille, "Progress towards measuring the electric dipole moment of the electron in metastable PbO," AIP Conf. Proc. 698 (2004) 192. [40] S. K. Lamoreaux, "Solid state systems for electron electric dipole moment and other fundamental measurements," nucl-ex/0109014. [41] Y. K. Semertzidis, "Electric dipole moments of fundamental particles," hepex/0401016. 121 [42] J. J. Hudson, B. E. Sauer, M. R. Tarbutt and E. A. Hinds, "Measurement of the electron electric dipole moment using YbF molecules," Phys. Rev. Lett. 89 (2002) 023003, hep-ex/0202014. [43] R. Mahbubani, L. Senatore, in preparation [44] N. Arkani-Hamed and S. Dimopoulos, "Supersymmetric unification without low energy supersymmetry and signatures for fine-tuning at the LHC," hep- th/0405159. [45] G. F. Giudice and A. Romanino, "Split supersymmetry," Nucl. Phys. B 699 (2004) 65 [Erratum-ibid. B 706 (2005) 65], hep-ph/0406088. [46] A. Pierce, "Dark matter in the finely tuned minimal supersymmetric standard model," Phys. Rev. D 70 (2004) 075006, hep-ph/0406144. [47] N. Arkani-Hamed, S. Dimopoulos, G. F. Giudice and A. Romanino, "Aspects of split supersymmetry," Nucl. Phys. B 709 (2005) 3, hep-ph/0409232. [48] L. Senatore, "How heavy can the fermions in split SUSY be? A study on gravitino and extradimensional LSP," Phys. Rev. D 71, 103510 (2005), hep-ph/0412103. [49] A. Masiero, S. Profumo and P. Ullio, "Neutralino dark matter detection in split supersymmetry scenarios," Nucl. Phys. B 712 (2005) 86 [arXiv:hep-ph/0412058]. [50] see for example E. W. Kolb and M. S. Turner, The Early Universe, Perseus Publishing, 1990 [51] A. Lue, K. Rajagopal and M. Trodden, "Semi-analytical approaches to local electroweak baryogenesis," Phys. Rev. D 56 (1997) 1250, hep-ph/9612282. 122 Chapter 5 Limits on non-Gaussianities from WMAP data We develop a method to constrain the level of non-Gaussianity of density perturbations when the 3-point function is of the "equilateral" type. Departures from Gaussianity of this form are produced by single field models such as ghost or DBI inflation and in general by the presence of higher order derivative operators in the effective Lagrangian of the inflaton. We show that the induced shape of the 3-point function can be very well approximated by a factorizable form, making the analysis practical. We also show that, unless one has a full sky map with uniform noise, in order to saturate the Cramer-Rao bound for the error on the amplitude of the 3-point function, the estimator must contain a piece that is linear in the data. We apply our technique to the WMAP data obtaining a constraint on the amplitude fuil of "equilateral" non-Gaussianity: -366 < fNqUil.< 238 at 95% C.L. We also apply our technique to constrain the so-called "local" shape, which is predicted for example by the curvaton and variable decay width models. We show that the inclusion of the linear piece in the estimator improves the constraint over those obtained by the WMAP team, to -27 < fLcal < 121 at 95% C.L. 123 5.1 Introduction Despite major improvements in the quality of cosmological data over the last few years, there are very few observables that can characterize the early phases of the Universe, when density perturbations were generated. Deviations from a purely Gaussian statistics of density perturbations would be a very important constraint on models of early cosmology. In single field slow-roll inflation, the level of non-Gaussianity is sharply predicted [1, 2] to be very small, less than 10-6. This is quite far from the present experimental sensitivity. On the other hand, many models have recently been proposed with a much higher level of non-Gaussianity, within reach of present or forthcoming data. For nearly Gaussian fluctuations, the quantity most sensitive to departures from perfect Gaussianity is the 3-point correlation function. In general, each model will give a different correlation between the Newtonian potential modes1 : (~(kl)(k2)-D(k3)) = (27r)3 63 (kl + k2 + k 3 )F(k l, k 2, k 3 ) . (5.1) The function F describes the correlation as a function of the triangle shape in momentum space. The predictions for the function F in different models divide quite sharply into two qualitatively different classes as a consequence of qualitatively different ways of producing correlations among modes [4]. The first possibility is that the source of density perturbations is not the inflaton but a second light scalar field a. In this case non-Gaussianities are generated by the non-linear relation between the fluctuation cr of this field and the final perturbation 1 we observe. This non-linearity is local as it acts when the modes are much outside the horizon; schematically we have 4(x) = A a(x) + B(6U2(x) - (6a2)) +.... 1 Even starting from a purely Gaussian Even with perfectly Gaussian primordial fluctuations, the observables, e.g. the temperature anisotropy, will not be perfectly Gaussian as a consequence of the non-linear relation between pri-5 mordial perturbations and what we will eventually observe. These effects are usually of order 10 (see for example [3]) and thus beyond (but not much) present sensitivity. In the following we will disregard these contributions. 124 do, the quadratic piece introduces a 3-point function for F(kl, k2, k 3 )= fl 2 4 (3 1 2 + where A is the power spectrum normalization, ((kl)( kl 3 , k3k 1 1 of the form + 3) k3 k2 k3 (5.2) (k 2)) = (27r)3 53(kl + k 2 )A - which for the moment has been taken as exactly scale invariant, and where flL is proportional to B. Examples of this mechanism are the curvaton scenario [5] and the variable decay width model [6], which naturally give rise to fLal greater than 10 and 5, respectively. The second class of models are single field models with a non-minimal Lagrangian, where the correlation among modes is created by higher derivative operators [7, 8, 9, 10]. In this case, the correlation is strong among modes with comparable wavelength and it decays when we take one of k's to zero keeping the other two fixed. Although different models of this class give a different function F, all these functions are qualitatively very similar. We will call this kind of functions equilateral: as we will see, the signal is maximal for equilateral configurations in Fourier space, whereas for the local form (5.2) the most relevant configurations are the squeezed triangles with one side much smaller than the others. The strongest constraint on the level of non-Gaussianity comes from the WMAP experiment. The collaboration analyzed their data searching for non-Gaussianity of the local form (5.2), finding the data to be consistent with purely Gaussian statistics and placing limits on the parameter -58 < fal f al [11]: < 134 at 95% C.L. (5.3) The main purpose of the study in this chapter is to perform a similar analysis searching for non-Gaussianities of the equilateral form. We can extend the definition of f 'alin eq. (5.2) to a generic function F by setting the overall normalization on equilateral configurations: F(k, k, k) = fNL 125 6A2 O6 .k (5.4) In this way, two different models with the same fNL will give the same 3-point function for equilateral configurations. For equilateral models in particular, the overall amplitude will be characterized by fquil.. Indirect constraints on fqUl. have been obtained starting from the limit of eq. (5.3) in [4], resulting in fNqLUil.l 800; we will see that a dedicated analysis is, as expected, much more sensitive. Many of the equilateral models naturally predict sizeable values of the parameter fNLUi':ghost inflation and DBI inflation [8, 9] tend to have fNLuil' - 100, tilted ghost inflation fqUil. > 1 [10], while slow roll inflation with higher derivative coupling typically give fNUil < 1 [7]. In Section 5.2 we discuss the main idea of the analysis. We will see that an optimal analysis is numerically very challenging for a generic form of the function F, but simplifies dramatically if the function F is factorizable in a sense that will be defined below. As all the equilateral forms predicted in different models are (qualitatively and quantitatively) very similar, our approach will be to choose a factorizable function that well approximates this class. The analysis is further complicated by the breaking of rotational invariance: a portion of the sky is fully masked by the presence of the galaxy and moreover each point is observed a different number of times. In Section 5.3 we look for an optimal estimator for fNL. We discover that, unlike the rotationally invariant case, the minimum variance estimator contains not only terms which are cubic in the temperature fluctuations, but also linear pieces. These techniques are used to analyze the WMAP data in Section 5.4. The result is that the data are compatible with Gaussian statistics. The limit for the equilateral models is -366 < fu.il < 238 at 95% C.L. (5.5) We also obtain a limit on fLal -27 < fOCal < 121 at 95% C.L. (5.6) which is a slight improvement with respect to (5.3) as a consequence of the abovementioned additional linear piece in the estimator. 126 5.2 A factorizable equilateral shape It is in principle straightforward to generalize the analysis for non-Gaussianities from the local model to another one. We start assuming that the experimental noise on the temperature maps is isotropic and that the entire sky is observed (no Galactic and bright source mask); we will relax these assumptions in the next Section. In this case it can be proved that an optimal estimator £ for fNL exists [12, 13], that is an estimator which saturates the Cramer-Rao inequality and thus gives the strongest possible constraint on the amplitude of the non-Gaussian signal. The estimator E is a sum of terms cubic in the temperature fluctuations, each term weighted by its signal to noise ratio: .E (5.7) (allma 2m 2 a1 3m 3 ) alNmal2 m2 al3 m3 imC 1 C12Cl all 12m2 3m £ Given the assumptions, the power spectrum CI (which is the sum of the CMB signal and noise) is diagonal in Fourier space; N is a normalization factor which makes the estimator unbiased. From rotational invariance we can simplify the expression introducing the Wigner 3j symbols to I £ = 11 'E ( lN imi 12 13 m2 m3 l l B11 12 13 1) < Cl111 a a al a m (5.8) 3 , (5.8) where Bl 1 21 . is the angle-averaged bispectrum which contains all the information about the model of non-Gaussianity we are considering. If B1112 13 is calculated for fNL = 1, then the normalization factor N is given by N= (B112 3 )2 1l1213C, C2Ci (5.9) (59) We now have to relate the angle-averaged bispectrum to the underlying correlation among 3d modes F(kl, k 2, k 3). After some manipulations following [14], the estimator 127 takes the form E= (5.10) 00 N S I d2 Ylml (n)Y 2m 2 ()Y/ 3m 3 r2 dr jll (klr)jl2(k 2 r)jl3(k 3 r) Cl C2 13 (n) limi 0 J 2k32dk---k 2 kF(kl, k2 , k3 )A (kl) A T (k )AL 2 (k3 ) aLmal2m2alm3 where AT (k) is the CMB transfer function which relates the aim to the Newtonian potential 1(k): d 3k aim = 47ri T (2)3 (k)J(k)Ylm(k). (5.11) Equation (5.10) is valid for any shape of the 3-point function F. Unfortunately this expression is computationally very challenging. The sums over m can be taken inside the integrals and factorized, but we are still left with a triple sum over I of an integral over the sphere. The calculation time grows as N5 x2 ls where the number of pixels Npixels is of order 3 x 106 for WMAP. This approach is therefore numerically too demanding. As noted in [14], a crucial simplification is possible if the function F is factorizable as a product of functions of k, k2 and k3 or can be written as a sum of a small number of terms with this property. In this case the second line of (5.10) becomes factored as the product of functions of each 1 separately, so that now also the sum over can be done before integrating over the sphere. For example if we assume that F(kl, k2, k3) = f(kl)f -- ' =N 2 (k2)f 3 (k 3), ~J 2$f dr Jfr2dr 0 H1 | i=1 limi the estimator simplifiesto Ym,(k). 1alam, ji(kr)fi(k)T(k)C; aimiimi(fn) (5.12) The calculation is obviously much faster now: it now scales like N3 2es and it is dominated by going back and forth between real and spherical harmonics space. From expression (5.2) we see that this simplification is possible for the local shape, and it was indeed used for the analysis of the WMAP data in [11, ?]. 128 Unfortunately, none of the "equilateral models" discussed in the Introduction predicts a function F which is factorizable (see some explicit expressions in [4]), so that it is not easy to perform an optimal analysis for a particular given model. However, all these models give 3-point functions which are quite similar, so that it is a very good approximation to take a factorizable shape function F which is close to the class of functions we are interested in and perform the analysis for this shape. In the limit k - 0 with k 2 and k 3 fixed, all the equilateral functions diverge as k- 1 [4] (while the local form eq. (5.2) goes as k1 3 ). An example of a function which has this behavior, is symmetric in k1, k2 and k 3, and is a sum of factorizable functions (and is homogeneous of order k - 6 , see below) is 2 F(kl,k2, k3) = fNLU'-6A2 ( kkk 2 k2 1 k3 k2 3 , 2 3 ,2 3 + (5perm. (5.14) where the permutations act only on the last term in parentheses. The mild divergence for kl - 0 is achieved through a cancellation among the various terms. In figure 5-1, we compare this function with the local shape. The dependence of both functions under a common rescaling of all k's is fixed to be oc k - 6 by scale invariance, so that we can factor out k 6 for example. Everything will now depend only on the ratios k 2 /kl and k 3 /k l , which fix the shape of the triangle in momentum space. For each shape we plot F(1, k2 /kl, k 3 /kl)(k 2/k 1) 2 (k 3/kl) 2 ; this is the relevant quantity if we are interested in the relative importance of different triangular shapes. The square of this function in fact gives the signal to noise contribution of a particular shape in momentum space [4]. We see that for the function (5.14), the signal to noise 2 Eq. (5.14) can be derived as follows. In order to make the divergence of F mild in the squeezed limit we can use at the numerator a quantity which goes to zero in the same limit. The area of the triangle does the job, going like k for kl - 0. The area can be expressed purely in terms of the sides through Heron's formula [16], A = s(s - ki)(s - k 2 )(s - k3 ), where s = (kl I +k 2 +k 3 ) is the semiperimeter. The first s in the square root is irrelevant for our purposes, since it goes to a constant in the squeezed limit; we will therefore omit it. Also we want a sum of factorizable functions, so we get rid of the square root by considering A2 . In conclusion, a function with all the features stated above is (s - kl)(s - k )(s - k ) F(kwhich, exactly once to expanded, eq. (5.14). reduces 2 which, once expanded, reduces exactly to eq. (5.14). 129 3 (5.13) is concentrated on equilateral configurations, while squeezed triangles with one side much smaller than the others are the most relevant for the local shape. In figure 5-2 we study the equilateral function predicted both in the presence of higher-derivative terms [7] and in DBI inflation [9]. In the second part of the figure we show the difference between this function and the factorizable one used in our analysis. We see that the relative difference is quite small. The same remains true for other equilateral shapes (see [4] for the analogous plots for other models). In [4], a "cosine" between different shapes was defined which quantifies how different is the signal given by two distributions. The cosine is calculated from the scalar product of the functions in fig. 5-1. We can think about this cosine as a sort of correlation coefficient: if the cosine is close to 1 the two shapes are very difficult to distinguish experimentally and an optimal analysis for one of them is very good also for the other. On the other hand, a small cosine means that, once non-Gaussianities are detected, there is a good chance to distinguish the two functions and that an optimal analysis for one shape is not very good for the other. The cosine between our template shape and the functions predicted by equilateral models is very close to one (0.98 with the ghost inflation [8] prediction and 0.99 for higher derivative/DBI models [7, 9]). This means that the error introduced in the analysis by the use of the factorizable shape instead of the correct prediction for a given equilateral model is at the percent level. On the other hand, as evident from figure 5-1, our template shape is quite different from the local model - the cosine is merely 0.41. All these numbers are obtained in 3 dimensions assuming that we can directly measure the fluctuations with a 3d experiment like a galaxy survey. The CMB anisotropies are a complicated projection from the 3d modes. This makes it more difficult to distinguish different shapes, although the picture remains qualitatively the same. Let us proceed with the study of our template shape. To further simplify the 130 Factorizable eq. shape 0.8 0.6 0.4 0.2 o Local shape 6 4 2 1 0.8 Figure 5-1: Plot of the function F(l, k2/k1, k3/kl)(k2/kl?(k3/kl)2 for the equilateral shape used in the analysis (top) and for the local shape (bottom). The functions are both normalized to unity for equilateral configurations ~ = ~ = 1. Since F(k1, k2, k3) is symmetric in its three arguments, it is sufficient to specify it for k1 2:: k2 2:: k3, so ~ ~ ~ ~ 1 above. Moreover, the triangle inequality says that no side can be longer than the sum of the other two, so we only plot F in the triangular region 1 - ~ ~ ~ ~ ~ ~ 1 above, setting it to zero elsewhere. 131 estimator in eq. (5.10), we define the functions a,(r) (2) - 2 +OO 2 r+O 2 r+OO at(r) 2 T dk k (5.15) kr A dk k-l 1(k)j l (kr)A +00 (5.16) AT dk/\ (k)jl(kr), (5.18) We use this strange ordering of the functions to keep the notation compatible with [11, ?], where the functions a and p were introduced for the analysis of the local shape. To evaluate , we start from the spherical harmonic coefficients of the map aim and calculate the four maps A(r,h)- Zt atr)Yim(h)a'm B(r, ) ~Cl2 )Yjlm (n)alm (519) Im C(r,f)= / Cl Im cYim(h)alm D(r, ) - Im ( () Ylm()alm (5.20) Im Now the estimator £ is given by NJ2J 2h £=-N8| r2dr d2 nr A(r, )B(r, 22 2+ )D( rn))r(rf)D(r' (r, )D(r, n)j )] ) (5.21) and the normalization N can be calculated from (5.9) using the explicit form Bl l = :/(211+ 1)(212 + 1)(213 + 1)(11 0 0 12 0 13) 6 r2 dr [-a-,,(r)312(r)/311 (r) + (2 perm.) -26 11 (r)612 (r)&13 (r) + The functions a, 1l (5.23) (r)'Y12 (r)61 3 (r) + (5 perm.)] , y and 6 used in the analysis can be obtained numerically start- ing from the transfer functions A[(k), which can be computed given a particu- 132 lar cosmological model with publicly available software as CMBFAST [17]. Some plots of these functions are given in fig. 5-3, where we choose values of r close to 0 - (conformal time difference between recombination and the present), as TR these give the largest contribution to the estimator. The oscillatory behavior in- duced by the transfer function A (k) is evident, with a peak at 1 - 200. The factors of 1(1+ 1) and (21 + 1) in the P and 6 functions, respectively, can be un- derstood from the behavior at low l's, in the Sachs-Wolferegime. Here the transfer function can be approximated by AT(k) = -jl (k(0o that fo 1(i(o - R) oc fo 0 dk k -1 j (k(ro - dk j (k(ro - TR)) R)) /3, and we then see R)) oc 1/(1(1 + 1)) and 1(to - TR) oc OC1/(21 + 1). From these expressions we also see that the function 3 (the only one which is dimensionless) is very similar to the the Sachs-Wolfe regime,Cmb 5.3 - 2 0 dk k-' A j2(k(o - Clmb's as, in TR)). Optimal estimator for fNL There are two quite general experimental complications that make the estimator £ defined in the last Section non-optimal. First, foreground emission from the Galactic plane and from isolated bright sources contaminates a substantial fraction of sky, which must be masked out before the analysis [18]. This projection which can be accomplished by giving very large noise to regions that are affected by foregrounds, besides reducing the available amount of data, has the important effect of breaking rotational invariance, so that the signal covariance matrix clm m12 2 of the masked sky becomes non-diagonal in multipole space. Second, the noise level is not the same in different regions of the sky, because the experiment looks at different regions for different amounts of time. This implies that also the noise covariance matrix C1osel12 is non-diagonal in multipole space. The importance of the anisotropy of the noise was pointed out in [11], as it caused an increase of the variance of the estimator £ at high 's. For these reasons, it is worth trying to generalize the estimator £. We consider estimators for fNL (for a given shape of non-Gaussianities) which contain both trilinear 133 and linear terms in the alm's. In the rotationally invariant case, a linear piece in the estimator can only be proportional to the monopole, which is unobservable. However, in the absence of rotational invariance, linear terms can be relevant. In this class, it is straightforward to prove that the unbiased estimator with the smallest variance is given by Eii(a) = ((ama 2m 2 a 3 m 3 )i C11mI,14 m 2m 2,l 5m 4 5 13m3 ,6m6 al 4 m 4 alm 5 aAf4 ) limi -3 1? t~m (allmlal2m2 al 3m 3)l Ci,12m2 C13 3 ,14 m4a4m 4 ) where N is a normalization factor: N= E m C- 1 (alml al2 ml2 a 3m3 ) 1 llml,lm4 12m2,15m5 C-1 (a 4 m4 a 5 m5 al6 m6 ) (5.25) lam3,16m6 limi The averages (...)1 are taken with fNL = 1. As we stressed, the covariance matrix C = Cc mb + Cnoise contains the effects of the mask projection and the anisotropic noise and it is thus non-diagonal in multipole space. It turns out that the estimator Elinsaturates the Cramer-Rao bound, i.e. it is not only optimal in its class, but it is also the minimum variance unbiased estimator among all the possible ones. To prove this, we expand the probability distribution in the limit of weak non-Gaussianity (which is surely a good approximation for the CMB) using the Edgeworth expansion [19, 20, 21]. P(afNL) = (1-fNL a a a e-2 0 E (alil al2m2ailm3) alml aal 2m2 aalm )ea1r)N 2 limi 13 144 l4m415m5 al5m5 V/(27r)Ndet(C) (5.26) Now, following the same arguments of [13]which considered the same problem in the rotationally invariant case, the estimator is optimal (in the sense that its variance saturates the Cramer-Rao bound), if and only if the following condition is satisfied: dlogP(alfNL)= F(fNL)(in(a)- fNL) , d fNL 134 (5.27) where F(fNL) is a generic function3 of the parameter fNL. From the probabil- ity distribution (5.26) it is easy to check that the estimator dlog P(afNL)/dfNL Elin is proportional to in the limit of small fNL. We conclude that Olinis an optimal estimator for a nearly Gaussian distribution. We now want to make some approximations to the optimal estimator Ei£, to make it numerically easier to evaluate. The full inversion of the covariance matrix is computationally rather cumbersome (although doable as the matrix is with good approximation block diagonal [22]). We therefore approximate C-1a in the trilinear term of eq. (5.24) by masking out the sky before computing the alm's and taking C as diagonal: trilinear -a-l where (a,,mal12 m 2 ala3 m)l 1 E (ailmlal 2m2 al3 m3 )l l Ial, 12Mal_ , 2 imal m2am(. 12M2 ,Ni0f1101203 limi (5.28) is still given by the full sky expressions of the last Section. Once we have made this approximation for the trilinear term it is easy to prove that the choice for the linear term which minimizes the variance is linear 3 E N limi (allmlal 2 m 2 aa3 m 3 ) C11ml,l2m2 al 3 m 3 (5.29) C11012013 Note that we must not approximate the covariance matrix C in the numerator as diagonal. This would leave only a term proportional to the monopole a 00 (which is unobservable), as in the rotationally invariant case. The normalization factor N is given by N= Cl2Cl fsky C11213 (5.30) where fsky is the fraction of the sky actually observed. As shown in [23], this correctly takes into account the reduction of data introduced by the mask for multipoles much higher than the inverse angular scale of the mask. The accuracy of this approximation has been checked on non-Gaussian simulations for the local shape in [11]. 3 When eq. (5.27) is satisfied, the function F turns out to be the Fisher information for the parameter fNL, F(fNL) = ((dlog P(afNL)/dfNL) 2 ). 135 Let us try to understand qualitatively the effect of the linear correction. Take a large region of the sky that has been observed many times so that its noise level is low. This region will therefore have a small-scale power lower than average. Now, in a given realization depending on how the large scale modes look like, this longwavelength modulation of the small-scale power spectrum may be "misinterpreted" by the trilinear estimator as a non-Gaussian signal. Indeed, for the local shape, most of the signal comes precisely from the correlation between long wavelength modes and the small scale power [4]. On the other hand, for equilateral shapes, as we noted, the signal is quite low on squeezed configurations so that we expect this effect to be small. The linear term measures the correlation between a given map and the anisotropies in the power spectrum, thus correcting for this spurious signal. Clearly the spurious correlation is zero on average, but the effect increases the variance of the estimator. We will apply the linear correction of the estimator only for the local shape, since, as we will verify later, it gives a very small effect for equilateral shapes. Following the same steps as in the last Section to factorize the trilinear term in the estimator, we get an explicit expression for the linear piece in the local case: 3- J r dr Jd f 2 2 E (2SAB(fn, r) yl3,m3(fi) + SBB ('h, r) a1 (r) a13 m 3 , Y 3 m3 13 13m3 (n) (5.31) where the two maps SAB(a, r) and SBB(fi, r) are defined as SAB(,I,r) = E °Cl Yll,ml(~) C- Yl2,m2(f)(allmlal2m2) (5.32) limi SBB(nfr) = E limi / C11Ylr,ml1(n) l'12 l12m2(n)(allmla 2m2) (5.33) As discussed above, it is only the anisotropic part of the matrix (allml a12 m 2 ) that gives a contribution to the linear part of the estimator. This matrix can be decomposed as (anoise noise acmb) (allmla2m2) = (acmb allm1L=lm 2 +]-lmla2m2), 136 (534) where ab are the a's of a map generated with CMB signal only and then per- forming the mask projection, and amiseare the ones associated with a map generated with noise only and then performing the mask projection. Both of these two matrices have an anisotropic component. For the map (a mbal2mb2), this arises only because of n ise), it is generated both by the sky cut and noisea 12m2 the sky cut, while for the map \(allml 1 by the anisotropy of the noise power. We already discussed about the effect of the anisotropy of the noise in the previous paragraph; this effect turns out to be the most relevant. The sky cut gives a much smaller effect because, as we will explain more in detail in the next Section, it is mainly associated with the average value of the temperature outside of the sky cut, and this average value is subtracted out before the analysis. 5.4 Analysis of WMAP 1-year data We keep our methodology quite similar to the one used by the WMAP collaboration for their analysis of the local shape [11] in order to have a useful consistency check. We compare WMAP data with Gaussian Monte-Carlo realizations, which are used to estimate the variance of the estimator. We generate with HEALPix 4 a random CMB realization, with fixed cosmological parameters, at resolution nside=256 (786,432 pixels). The parameters are fixed to the WMAP best fit for a ACDM cosmology with power-law spectrum [24]: Qbh2 = 0.024, Qmh2 = 0.14, h = 0.72, T = 0.17, n, = 0.99. With these parameters, the present conformal time bination time is TR T0 is 13.24 Gpc and the recom- = 0.27 Gpc. A given realization is smoothed with the WMAP window functions for the Q1, Q2, V1, V2, W1, W2, W3 and W4 bands [25]. To each of these 8 maps we add an independent noise realization: for every pixel the noise is a Gaussian random variable with variance Uo2/Nbs, where Nobs is the number of observations of the pixel and u0 is a band dependent constant [26]. The maps are then combined to give a single map: we make a pixel by pixel average of the 8 maps weighted by the noise a2/Nobs We use this procedure because it is identical to that 4 See HEALPix website: http://www.eso.org/science/healpix/ 137 used in [11], thereby allowing direct comparisons between our results and those of the WMAP team. In future work, it can in principle be improved in two ways. First of all, for the window function to be strictly rather than approximately uniform across the sky, the weights of the eight input maps should be constant rather than variable from pixel to pixel. This is a very small effect in practice, since the eight Nbs-maps are very similar, making the weights close to constant. Second, the sensitivity on small scales can be improved by using -dependent weights [27]: for instance, at very high 1, most of the weight should be given to the W bands, since they have the narrowest beam. This would also have a small effect for our particular application, since our estimator uses only the first few hundred multipoles. We explicitly checked that this would not reduce the variance of our estimator appreciably. We apply the KpO mask to the final map, to cut out the Galactic plane and the known point sources [18]: this mask leaves the 76.8% of the pixels, fsky = 0.768. The average temperature outside the mask is then subtracted. On the resulting map we calculate the estimators defined in the preceding Sections. For the local shape we performed the analysis both with and without the linear piece discussed in Section 5.3. The quadratic maps of equations (5.32) and (5.33) used for the linear correction are calculated by averaging many ( 600) Monte-Carlo maps obtained with the same procedure above. We use HEALPix to generate and analyze maps at resolution level nside=256 (786,432 pixels). The integration over r is performed from 0 - 0.025 TR up to 0 - 2.5 TR with - 200 equidistant points, and then with another logarithmically spaced - 60 points up to the present epoch. Such a high resolution both in r spacing and in nside is necessary in order to reproduce the cancellation on squeezed triangles which occurs among the various terms in the equilateral shape (5.14). The computation of each fNL on a 2.0 GHz Opteron processor with 2 GB of RAM takes about 60 minutes for the local shape, and 100 minutes for the equilateral shape. We then apply exactly the same procedure to WMAP data. These maps are analyzed after template foreground corrections are applied, in order to reduce foreground signal leaking outside the mask, as described in [18]. 138 A useful analytic bound on the variance of the estimators for flocal and feqil. is obtained from the variance of the full sky estimator (with homogeneous noise) (5.8) with an fsky-correction which takes into account the reduction in available information from not observing the whole sky. Taking into account the normalization factor, one readily obtains D2 (5.35) C1 C1 /21C3 an = fsky 11<12<13 where B11 213 must be evaluated for fcal or fil. equal to 1. As we discussed, our approach is not strictly optimal as we are not inverting the full covariance matrix, so we expect Lan to be smaller than the actual standard deviation that we measure from our Monte-Carlos. In figure 5-4, we show the standard deviation of estimators for the local shape parameter fNal as a function of the maximum multipole analyzed. We compare the results of our Monte-Carlo simulations (with and without the linear correction) with the analytic bound discussed above. The results without linear correction are compatible with the analysis in [11]. We see that the addition of the linear piece reduces the variance divergence at high l's. The residual divergence is probably associated with the fact that we did not invert the full covariance matrix. estimator with the smallest variance is the one with linear correction at max The = 335, with a standard deviation of 37. The analytic approximation has an asymptotic value of 30 which should be considered the best possible limit with the present data. Our estimator thus extracts about (37/30)- 2 66% of the fal-information (inverse variance) from the WMAP data. The analysis of the data with lmax = 335 gives 47, so there is no evidence of deviation from pure Gaussianity and we can set a 2a limit of -27 < fINL < 121 at 95% C.L. (5.36) In fig. 5-5, we show a map SAB(n, r) for a radius around recombination calculated with noise only and for Imax= 370, as an illustrative example of the role of the linear piece. As we will clarify below, it is in fact the anisotropy of the noise that causes most of the contribution to the linear piece. The companion map SBB(n, r) is qualitatively 139 very similar to the map SAB. It is the anisotropy of the noise that gives rise to a nontrivial (i.e. non-uniform) SAB map, as is clear from the comparison in the same figure with a plot of the number of observations per pixel. This can help us in understanding the contribution of the linear piece associated with the anisotropy of the noise to the estimator. Let us consider a particular Gaussian Monte Carlo realization with a long wavelength mode crossing one of the regions with a high number of observations. The trilinear piece of the estimator will then detect a spurious non-Gaussian signal associated with the correlation between this long wavelength mode and the small short scale power of the noise (5). The maps SAB and SBB, because of their particular shape, will have a non-zero dot product with precisely the same long wavelength mode of the Monte Carlo map, and with an amplitude proportional to the anisotropy of the two point function of the noise, thus effectively subtracting the spurious signal detected by the trilinear piece and reducing the variance of the estimator. As the mask breaks rotational invariance, the CMB signal also gives non-uniform S-maps. However, the breaking of rotational invariance can be neglected far from the galaxy mask, so we expect that the contribution to the S-maps from the CMB is to first approximation constant outside the mask (and zero inside). In this approximation, the linear contribution of the estimator would be sensitive only to the average value outside the mask. However, given that we are constraining the statistical properties of temperature fluctuations, the average temperature in the region outside the mask has to be subtracted before performing the analysis; therefore we expect the linear correction associated with the CMB signal to be rather small. These expectations are in fact verified by our simulations. We checked that the maps S coming from the CMB signal are to first approximation constant outside the mask, with additional features coming from the patches used to mask out bright sources outside the Galactic plane. We then verified that if the average value of the temperature is not subtracted, the linear term coming from the CMB signal gives a very important reduction of the estimator variance, while its effect is almost negligible when the mean temperature is 5 The effect of this spurious signal obviously averages to zero among many Monte Carlo realizations but it increases the variance of the estimator. 140 subtracted, as we do in the analysis. Uil . The In figure 5-6, we study the standard deviation of the estimators for fNL estimator without linear corrections is quite close to the information-theoretical limit, so we did not add the linear corrections (which would require many averages analogous to the maps SAB and SBB). The reason why we have such a good behavior is that, as mentioned, here most of the signal comes from equilateral configurations. Thus the estimator is not very sensitive to the correlation between the short scale power and the (long wavelength) number of observations. Moreover the inversion of the covariance matrix is important only for the low multipoles, which give only a small fraction of the signal for fLil '. We find that the estimator with smallest standard deviation is at m,, = 405, with a standard deviation of 151. The analysis of the data with the same Im,, gives -64. We deduce a 2a limit -366 < fL u". < 238 at 95% C.L. (5.37) The given limit is approximately 2 times stronger than the limit obtained in [4], indirectly obtained starting from the limit on f al. The limit appears weaker than for the local case because, for the same fNL, the local distribution has larger signal to noise ratio than an equilateral one, as evident from fig. 5-1. This is not a physical difference but merely a consequence of our defining fNL at the equilateral configuration. To check that the two limits correspond approximately to the same "level of non-Gaussianity" we can define a quantity which is sensitive to the 3-point function integrated over all possible shapes of the triangle in momentum space. In this way it will be independent of which point we choose for normalization. We define this quantity, which we will call NG, directly in 3d as an integral of the square of the functions in fig. 5-1 NG (/E ( X2dzX2 3dx 2 he 3 F(1, X2, 2 3) o Aie Xi-3 1/2 (5.38) (5.38) The integration is restricted to the same triangular region as in fig.s 5-1 and 5-2: 141 1- 2 _< X 3 < X2 < 1. The measure of integration comes from the change of variables from the 3d wavevectors to the ratios x2 = k 2 /k1 and parametrically of order A/ 2 X3 = k 3 /kl. NG is · fNL (where A - 1.9.10 - 8 [24, 28]), which is the correct order of magnitude of the non-Gaussianity, analogous to the skewness for a single random variable. To better understand the meaning of the defined quantity, note that if we integrate inside the parentheses in eq. (5.38) over the remaining wavevector we get the total signal to noise ratio (or better the non-Gaussian to Gaussian ratio). This further integration would approximately multiply NG by the square root of the number of data. This means that to detect a given value of NG we need a number of data of order NG - 2 , in order to have a signal to noise ratio of order 1. This is clearly a good way to quantify the deviation of the statistics from pure Gaussianity. The 2c windows for fNL can be converted to constraints on NG (we define NG to have the same sign of fNL) 6 -0.006 < NGlOCal< 0.025 at 95% C.L. (5.39) at 95% C.L. (5.40) -0.016 < NGeqUil.< 0.010 Contrary to what one might think from a naive look at the limits on fNL,the maximum tolerated amount of non-Gaussian signal in a map is thus very similar for both shapes. So far in this chapter, we have neglected any possible dependence on the cosmological parameters. A proper analysis would marginalize over our uncertainties in parameters and this would increase the allowed range of fNL. In order to estimate what the effect of these uncertainties is, let us imagine that the real cosmological parameters are not exactly equal to the best fit ones. The cosmological parameters are determined mainly from the 2-point function of the CMB, Clmb, therefore the largest error bars are associated to those combinations of parameters which leave Clmb unchanged. In the limit in which the Cl mb's remain the same, also the variance of our estimator (which is computed with the best fit cosmological parameters for 6 Note that for the local model the integration in eq. (5.38) diverges like the log of the ratio of the minimum to the maximum scales in the integral. For the quoted numbers we chose a ratio of 1000. 142 a ACDM cosmology with power-law spectrum [24]) remains the same: in the weak non-Gaussian limit the variance just depends on the 2-point function. However, the combination of parameters which leave unchanged the Clmb's does not necessarily leave unchanged the bispectrum, therefore uncertainties in the determination of the cosmological parameters will have an influence on the expectation value of the esti- mator. The expectation value of the estimator would be: () = 1 1 N =j (5.41) 11C12B13 0 C1C12Ct3 where N, Bht213a and C1 are the same as in our analysis, computed with the best fit cosmological parameters, while Bl,1213 is the true bispectrum. Thus when varying parameters, the normalization N needs to be changed to make the estimator unbiased. The most relevant uncertainty is the reionization optical depth , which is cor- related with the uncertainty in the amplitude of the power spectrum &A. With the purpose of having a rough estimate, we can approximate the effect of reionization by a multiplicative factor e -T in front of the transfer function A\'(k) for 1 corresponding to scales smaller than the horizon at reionization. In order to keep the Clfmb's unchanged, at least at high l's, we then multiply A by e2 r. The bispectrum is proportional to A2 A(k) 3 and thus changes even if the Cl's do not. For the equilateral shape most of the signal comes from equilateral configurations with all the 3 modes inside the horizon at reionization; in this case () scales roughly as e. For the local shape the signal comes from squeezed configurations with one mode of much smaller wavelength than the others. Taking only 2 modes inside the horizon at reionization we get () oc e2 . If we consider the la error, r = 0.166+0.076[24], we find that the uncertainty in the reionization depth should correspond to an error on fNL of order 8% for the equilateral shape, and 15% for the local shape. These numbers translate directly into uncertainties in the allowed ranges we quote. In the future, if the error induced by the uncertainty in the cosmological parameters will become comparable to the variance of the estimator, a more detailed analysis will be required; however, at the moment 143 an analysis with fixed cosmological parameters is certainly good enough. 5.5 Conclusions We have developed a method to constrain the level of non-Gaussianity when the induced 3-point function is of the "equilateral" type. We showed that the induced shape of the 3-point function can be very well approximated by a factorizable form making the analysis practical. Applying our technique to the WMAP first year data we obtained -366 < fHq"l. < 238 at 95% C.L. (5.42) The natural expectation for this amplitude for ghost or DBI inflation is fuil. 100, below the current constraints but at a level that should be attainable in the future. The limit an experiment can set on fNL just scales as the inverse of the maximum the experiment can detect. As a result, in the case of WMAP, increased observing time will approximately decrease the error bars by 30% and 60% for 4 years and 8 years of observation. The increased angular resolution and smaller noise of Planck pushes the point where noise dominates over signal to 1 - 1500. This should result in a factor of 4 improvement on the present constraints. In addition polarization measurements by Planck can further reduce the range by an additional factor of 1.6 [29]. We also constrained the presence of a 3-point function of the "local" type, predicted for example by the curvaton and variable decay width inflation models, ob- taining -27 < fNL < 121 at 95% C.L. (5.43) We defined a quantity NG, which quantifies for any shape the level of nonGaussianity of the data, analogously to the skewness for a single random variable. The limits on NG are very similar for the two shapes, approximately NG] < 0.02 at 95% C.L. . We showed that unless one has a full sky map with uniform noise, the estimator 144 must contain a piece that is linear on the data in order to extract all the relevant information from the data and saturate the Cramer-Rao bound for the 3-point function measurement uncertainty. This correction is particularly important for the local shape and accounts for the improvement of our limits over that from the WMAP team. Moreover this correction goes a long way towards reducing the divergence in the variance of the estimator as Ima is increased. 145 Higher derivative 0.8 0.6 0.4 0.2 o 1 Difference 0.8 0.6 0.4 0.2 o 1 0.8 0.6 Figure 5-2: Top. Plot of the function F(l, k2/kl, k3/kJ)(k2/kl?(k3/kl? predicted by the higher-derivative [7] and the DBI models [9]. Bottom. Difference between the above plot and the analogous one (top of fig. 5-1) for the factorizable equilateral shape used in the analysis. 146 1(1+1)P x 109 ax 1011 4F I -1 -/ - - '% 2 100 200 300 .---------)-? . 1' 1 -2 :1 -2 -4 y x 1012 ,'' (21+1) x 1010 3 -I 4 'I 200 12 / 100 200 300 , / ,'100 200_ / -2 - i ro- 0.7TR--.--. -3-3 -8 r-70 To - 1.6 TR -1 To -1.3 TR -To -TR o - 0. 4 TR Figure 5-3: The functions al(r) (in units of Mpc-3 ), /31(r) (dimensionless), yl(r) (in units of Mpc-2), and 61(r) (in units of Mpc -1 ) are shown for various radii r as functions of the multipole number 1. The cosmological parameters are the same ones used in the analysis: 2 Q2bh = 0.024, Qmh2 = 0.14, h = 0.72, r = 0.17. With these parameters, the present conformal time ' 0 is 13.24 Gpc and the recombination time 147 TR is 0.27 Gpc. Aflocal 7n N L Iu 60 f I 50 40 30 20 i i 4 10 ·· I . 250 . . . . . 300 . 350 400 Imax Figure 5-4: Standard deviation for estimators of flmal as a function of the maximum used in the analysis. Lower curve: lower bound deduced from the full sky variance. Lower data points: standard deviations for the trilinear + linear estimator. Upper data points: the same for the estimator without linear term, for which the divergence at high 's caused by noise anisotropy had already been noticed in [11]. 148 Figure 5-5: Top: SAB(n, r) map for r around recombination calculated with noise only. Bottom: number of observations as a function of the position in the sky for the Ql band (the plot is quite similar for the other bands). A lighter color indicates points which are observed many times. 149 Afe q 250 200 f 150 i f~ 100 50 250 300 350 400 450 Ima Figure 5-6: Standard deviations for estimators of fjNL. as a function of the maximum 1 used in the analysis. Lower curve: lower bound deduced from the full sky variance. Data points: standard deviations for the estimator without linear term. 150 Bibliography [1] J. Maldacena, "Non-Gaussian features of primordial fluctuations in single field inflationary models," JHEP 0305, 013 (2003) [astro-ph/0210603]. [2] V. Acquaviva, N. Bartolo, S. Matarrese and A. Riotto, "Second-order cosmological perturbations from inflation," Nucl. Phys. B 667, 119 (2003) [astroph/0209156]. [3] P. "CMB 3-point functions generated by Creminelli and M. Zaldarriaga, non-linearities at recombination," Phys. Rev. D 70, 083532 (2004) [astro- ph/0405428]. [4] D. Babich, P. Creminelli and M. Zaldarriaga, "The shape of non-Gaussianities," JCAP 0408, 009 (2004) [astro-ph/0405356]. [5] D. H. Lyth, C. Ungarelli and D. Wands, "The primordial density perturbation in the curvaton scenario," Phys. Rev. D 67, 023503 (2003) [astro-ph/0208055]. [6] M. Zaldarriaga, "Non-Gaussianities in models with a varying inflaton decay rate," Phys. Rev. D 69, 043508 (2004) [astro-ph/0306006]. [7] P. Creminelli, "On non-Gaussianities in single-field inflation," JCAP 0310, 003 (2003) [astro-ph/0306122]. [8] N. Arkani-Hamed, P. Creminelli, S. Mukohyama and M. Zaldarriaga, "Ghost inflation," JCAP 0404, 001 (2004) [hep-th/0312100]. [9] M. Alishahiha, E. Silverstein and D. Tong, "DBI in the sky," Phys. Rev. D 70, 123505 (2004) [hep-th/0404084]. 151 [10] L. Senatore, "Tilted ghost inflation," Phys. Rev. D 71, 043512 (2005) [astroph/0406187]. [11] E. Komatsu et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Tests of Gaussianity," Astrophys. J. Suppl. 148, 119 (2003) [astroph/0302223]. [12] A. F. Heavens, "Estimating non-Gaussianity in the microwave background," astro-ph/9804222. [13] D. Babich, "Optimal Estimation of Non-Gaussianity," astro-ph/0503375. [14] L. M. Wang and M. Kamionkowski, "The cosmic microwave background bispectrum and inflation," Phys. Rev. D 61, 063504 (2000) [astro-ph/9907431]. [15] E. Komatsu, D. N. Spergel and B. D. Wandelt, "Measuring primordial nonGaussianity in the cosmic microwave background," astro-ph/0305189. [16] Heron of Alexandria, Metrica, Alexandria, circa 100 B.C. - 100 A.D. [17] U. Seljak and M. Zaldarriaga, "A Line of Sight Approach to Cosmic Microwave Background Anisotropies," Astrophys. J. 469, 437 (1996) [astro-ph/9603033]. [18] C. Bennett et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Foreground Emission," Astrophys. J. Suppl. 148, 97 (2003) [astroph/0302208]. [19] R. Juszkiewicz, D. H. Weinberg, P. Amsterdamski, M. Chodorowski and F. Bouchet, "Weakly nonlinear Gaussian fluctuations and the Edgeworth expansion," Astrophys. J. 442 (1995) 39. [20] A. Taylor and P. Watts, "Parameter information from nonlinear cosmological fields," Mon. Not. Roy. Astron. Soc. 328, 1027 (2001) [astro-ph/0010014]. [21] F. Bernardeau and L. Kofman, "Properties of the Cosmological Density Distribution Function," Astrophys. J. 443 (1995) 479 [astro-ph/9403028]. 152 [22] G. Hinshaw et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Angular Power Spectrum," Astrophys. J. Suppl. 148, 135 (2003) [astro-ph/0302217]. [23] E. Komatsu, "The Pursuit of Non-Gaussian Fluctuations in the Cosmic Microwave Background," astro-ph/0206039. [24] D. N. Spergel et al. [WMAP Collaboration], "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Determination of Cosmological Parameters," Astrophys. J. Suppl. 148, 175 (2003) [astro-ph/0302209]. [25] L. Page et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Beam Profiles and Window Functions," Astrophys. J. Suppl. 148, 39 (2003) [astro-ph/0302214]. [26] C. L. Bennett et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Preliminary Maps and Basic Results," Astrophys. J. Suppl. 148, 1 (2003) [astro-ph/0302207]. [27] M. Tegmark, A. de Oliveira-Costa & A.J.S. Hamilton, "A high resolution foreground cleaned CMB map from WMAP", Phys. Rev. D 68, 123523 (2003) [astroph/0302496]. [28] L. Verde et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Parameter Estimation Methodology," Astrophys. J. Suppl. 148, 195 (2003) [arXiv:astro-ph/0302218]. [29] D. Babich and M. Zaldarriaga, "Primordial Bispectrum Information from CMB Polarization," Phys. Rev. D 70, 083005 (2004) [astro-ph/0408455]. [30] K. M. Gorski, E. Hivon and B. D. Wandelt, "Analysis Issues for Large CMB Data Sets," astro-ph/9812350. 153 Chapter 6 The Minimal Model for Dark Matter and Unification Gauge coupling unification and the success of TeV-scale weakly interacting dark matter are usually taken as evidence of low energy supersymmetry (SUSY). However, if we assume that the tuning of the higgs can be explained in some unnatural way, from environmental considerations for example, SUSY is no longer a necessary component of any Beyond the Standard Model theory. In this chapter we study the minimal model with a dark matter candidate and gauge coupling unification. This consists of the SM plus fermions with the quantum numbers of SUSY higgsinos, and a singlet. It predicts thermal dark matter with a mass that can range from 100 GeV to around 2 TeV and generically gives rise to an electric dipole moment (EDM) that is just beyond current experimental limits, with a large portion of its allowed parameter space accessible to next generation EDM and direct detection experiments. We study precision unification in this model by embedding it in a 5-D orbifold GUT where certain large threshold corrections are calculable, achieving gauge coupling and b-T unification, and predicting a rate of proton decay just beyond current limits. 154 6.1 Introduction Over the last few decades the search for physics beyond the Standard Model (SM) has largely been driven by the principle of naturalness, according to which the parameters of a low energy effective field theory like the SM should not be much smaller than the contributions that come from running them up to the cutoff. This principle can be used to constrain the couplings of the effective theory with positive mass dimension, which have a strong dependence on UV physics. Requiring no fine tuning between bare parameters and the corrections they receive from renormalization means that the theory must have a low cutoff. New physics can enter at this scale to literally cut off the high-energy contributions from renormalization. In the specific case of the SM the effective lagrangian contains two relevant parameters: the higgs mass and the cosmological constant (c.c), both of which give rise to problems concerning the interpretation of the low energy theory. Any discussion of large discrepancies between expectation and observation must begin with what is known as the c.c. problem. This relates to our failure to find a well-motivated dynamical explanation for the factor of 10120between the observed c.c and the naive contribution to it from renormalization which is proportional to A4 , where A is the cutoff of the theory, usually taken to be equal to the Planck scale. Until very recently there was still hope in the high energy physics community that the c.c. might be set equal to zero by some mysterious symmetry of quantum gravity. This possibility has become increasingly unlikely with time since the observation that our universe is accelerating strongly suggests the presence of a non-zero cosmological constant [1, 2]. A less extreme example is the hierarchy between the higgs mass and the GUT scale which can be explained by SUSY breaking at around a TeV. Unfortunately the failure of indirect searches to find light SUSY partners has brought this possibility into question, since it implies the presence of some small fine-tuning in the SUSY sector. This 'little hierarchy' problem [3, 4] raises some doubts about the plausibility of low energy SUSY as an explanation for the smallness of the higgs mass. Both these problems can be understood from a different perspective: the fact 155 that the c.c. and the higgs mass are relevant parameters means that they dominate low energy physics, allowing them to determine very gross properties of the effective theory. We might therefore be able to put limits on them by requiring that this theory satisfy the environmental conditions necessary for the universe not to be empty. This approach was first used by Weinberg [5]to deduce an upper bound on the cosmological constant from structure formation, and was later employed to solve the hierarchy problem in an analogous way by invoking the atomic principle [6]. Potential motivation for this class of argument can be found in the string theory landscape. At low energies some regions of the landscape can be thought of as a field theory with many vacua, each having different physical properties. It is possible to imagine that all these vacua might have been equally populated in the early universe, but observers can evolve only in the few where the low energy conditions are conducive to life. The number of vacua with this property can be such a small proportion of the total as to dwarf even the tuning involved in the c.c. problem; resolving the hierarchy problem similarly needs no further assumptions. This mechanism for dealing with both issues simultaneously by scanning all relevant parameters of the low energy theory within a landscape was recently proposed in [7, 8]. From this point of view there seems to be no fundamental inconsistency with having the SM be the complete theory of our world up to the Planck scale; nevertheless this scenario presents various problems. Firstly there is increasing evidence for dark matter (DM) in the universe, and current cosmological observations fit well with the presence of a stable weakly interacting particle at around the TeV scale. The SM contains no such particle. Secondly, from a more aesthetic viewpoint gauge couplings do not quite unify at high energies in the SM alone; adding weakly interacting particles changes the running so unification works better. A well-motivated example of a model that does this is Split Supersymmetry [7], which is however not the simplest possible theory of this type. In light of this we study the minimal model with a finely-tuned higgs and a good thermal dark matter candidate proposed in [8], which also allows for gauge coupling unification. Although a systematic analysis of the complete set of such models was carried out in [9], the simplest one we study here was missed because 156 the authors did not consider the possibility of having large UV threshold corrections that fix unification, as well as a GUT mechanism suppressing proton decay. Adding just two 'higgsino' doublets' to the SM improves unification significantly. This model is highly constrained since it contains only one new parameter, a Dirac mass term for the doublets ('pu'), the neutral components of which make ideal DM candidates for 990 GeV< L < 1150 GeV (see [9] for details). However a model with pure higgsino dark matter is excluded by direct detection experiments since the degenerate neutralinos have unsuppressed vector-like couplings to the Z boson, giving rise to a spin-independent direct detection cross-section that is 2-3 orders of magnitude above current limits2 [10, 11]. To circumvent this problem, it suffices to include a singlet ('bino') at some relatively high energy ( 109 GeV), with yukawa couplings with the higgsinos and higgs, to lift the mass degeneracy between the 'LSP' and 'NLSP' 3 by order 100 keV [12], as explained in Appendix 6.8. The instability of such a large mass splitting between the higgsinos and bino to radiative corrections, which tend to make the higgsinos as heavy as the bino, leads us to consider these masses to be separated by at most two orders of magnitude, which is technically natural. We will see that the yukawa interactions allow the DM candidate to be as heavy as 2.2 TeV. There is also a single reparametrization invariant CP violating phase which gives rise to a two-loop contribution to the electron EDM that is well within the reach of next-generation experiments. Our chapter is organized as follows: in Section 6.2 we briefly introduce the model, in Section 6.3 we study the DM relic density in different regions of our parameter space with a view to constraining these parameters; we look more closely at the experimental implications of this model in the context of dark matter direct detection and EDM experiments in Sections 6.4 and 6.5. Next we study gauge coupling unification at two loops. We find that this is consistent modulo unknown UV threshold corrections, 1 Here 'higgsino' is just a mnemonic for their quantum numbers, as these particles have nothing to do with the SUSY partners of the higgs. 2 A model obtained adding a single higgsino doublet, although more minimal, is anomalous and hence is not considered here. 3 From here on we will refer to these particles and couplings by their SUSY equivalents without the quotation marks for simplicity. 157 however the unification scale is too low to embed this model in a simple 4D GUT. This is not necessarily a disadvantage since 4D GUTs have problems of their own, in splitting the higgs doublet and triplet for example. A particularly appealing way to solve all these problems is by embedding our model in a 5D orbifold GUT, in which we can calculate all large threshold corrections and achieve unification. We also find a particular model with b-r unification and a proton lifetime just above current bounds. We conclude in Section 6.7. 6.2 The Model As mentioned above, the model we study consists of the SM with the addition of two fermion doublets with the quantum numbers of SUSY higgsinos, plus a singlet bino, with the following renormalizable interaction terms: 1 I'JuTd + 1 MI'T 2 where Is is the bino, 'u,d s + AuIuThJs + AdAdhit s (6.1) are the higgsinos, h is the finely-tuned higgs. We forbid all other renormalizable couplings to SM fields by imposing a parity symmetry under which our additional particles are odd whereas all SM fields are even. As in SUSY conservation of this parity symmetry implies that our LSP is stable. The size of the yukawa couplings between the new fermions and the higgs are limited by requiring perturbativity to the cutoff. For equal yukawas this constrains A(Mz) = Ad(Mz) 0.88, while if we take one of the couplings to be small, say Ad(Mz) = 0.1 then Au(Mz) can be as large as 1.38. The above couplings allow for the CP violating phase 0 = Arg(MlA*\A), giving 5 free parameters in total. In spite of its similarity to the MSSM (and Split SUSY) weak-ino sector, there are a number of important differences which have a qualitative effect on the phenomenology of the model, especially from the perspective of the relic density. Firstly a bino-like LSP, which usually mixes with the wino, will generically annihilate less effectively in this model since the wino is absent. Secondly the new 158 yukawa couplings are free parameters so they can get much larger than in Split SUSY, where the usual relation to gauge couplings is imposed at the high SUSY breaking scale. This will play a crucial role in the relic density calculation since larger yukawas means greater mixing in the neutralino sector as well as more efficient annihilation, especially for the bino which is a gauge singlet. Our 3x3 neutralino mass matrix is shown below: MN= M1 Auv AdV A,,v 0 -- ie i Adv -Ae io 0 for v = 174 GeV, where we have chosen to put the CP violating phase in the /z term. The chargino is the charged component of the higgsino with tree level mass A. It is possible to get a feel for the behavior of this matrix by diagonalizing it perturbatively for small off-diagonal terms, this is done in Appendix 6.8. 6.3 Relic Abundance In this section we study the regions of parameter space in which the DM abundance is in accordance the post-WMAP 2a region 0.094 < Qdmh2 < 0.129 [2], where RQdm is the fraction of today's critical density in DM, and h = 0.72 ± 0.05 is the Hubble constant in units of 100 km/(s Mpc). As in Split SUSY, the absence of sleptons in our model greatly decreases the number of decay channels available to the LSP [13, 14]. Also similar to Split SUSY is the fact that our higgs can be heavier than in the MSSM (in our case the higgs mass is actually a free parameter), hence new decay channels will be available to it, resulting in a large enhancement of its width especially near the WW and ZZ thresholds. This in turn makes accessible neutralino annihilation via a resonant higgs, decreasing the relic density in regions of the parameter space where this channel is accessible. For a very bino-like LSP this is easily the dominant annihilation channel, allowing the bino density to decrease to an acceptable level. We use a modified version of 159 the DarkSUSY [15] code for our relic abundance calculations, explicitly adding the resonant decay of the heavy higgs to W and Z pairs. As mentioned in the previous section there are also some differences between our model and Split SUSY that are relevant to this discussion: the first is that the Minimal Model contains no wino equivalent (this feature also distinguishes this model from that in [16], which contains a similar dark matter analysis). The second difference concerns the size of the yukawa couplings which govern this mixing, as well as the annihilation cross-sectionto higgses. Rather than being tied to the gauge couplings at the SUSY breaking scale, these couplings are limited only by the constraint of perturbativity to the cutoff. This means that the yukawas can be much larger in our model, helping a bino-like LSP to both mix more and annihilate more efficiently. These effects are evident in our results and will be discussed in more detail below. We will restrict our study of DM relic abundance and direct detection in this model to the case with no CP violating phase ( = 0, 7r); we briefly comment on the general case in Section 6.5. Our results for different values of the yukawa couplings are shown in Figure 6-1 below, in which we highlight the points in the t-M1 plane that give rise to a relic density within the cosmological bound. The higgs is relatively heavy (Mhiggs= 160 GeV) in this plot in order to access processes with resonant annihilation through an s-channel higgs. As we will explain below the only effect this has is to allow a low mass region for a bino-like LSP with M1l Mhiggs/2. Notice that the relic abundance seems to be consistent with a dark matter mass as large as 2.2 TeV. Although a detailed analysis of the LHC signature of this model is not within the scope of this study, it is clear that a large part of this parameter space will be inaccessible at LHC. The pure higgsino region for example, will clearly be hard to explore since the higgsinos are heavy and also very degenerate. There is more hope in the bino LSP region for a light enough spectrum. While analyzing these results we must keep in mind that Qdm 10- 9GeV-2/(a)eff, where ()eff is an effective annihilation cross section for the LSP at the freeze out temperature, which takes into consideration all coannihilation channels as well as the thermal average [17]. It will be useful to approximate this quantity as the cross160 3000 2000 I .... .. annihilates too much annihilates too little ..... 1000 ... 500 1000 + (}=O, >'u=O.I, >'d=O.1 • (}=7r, >'u=O.I, >'d=O.1 • (}=O, >'u=0.88, >'d=0.88 • (}=7r, >.,,=0.88, >'d=0.88 • (}=O, >'u=1.38, • (}=7r, >',,=1.38, >'d=O.1 2000 1500 >'d=O.1 2500 3000 /l (GeV) Figure 6-1: Graph showing regions of parameter space consistent with WMAP. section for the dominant annihilation will help us build some intuition channel. Although rough, this approximation on the behavior of the relic density in different parts of the parameter space. We will not discuss the region close to the origin where the interpretation of the results become more involved due to large mixing and coannihilation. 6.3.1 Higgsino Dark Matter In order to get a feeling for the structure of Figure 6-1, it is useful to begin by looking at the regions in which the physics is most simple. This can be achieved by diminishing the the number of annihilation channels that are available to the LSP by taking the limit of small yukawa couplings. For B imation = 0, mixing occurs only on the diagonal M1 (see Appendix = Jl, to a very good approx- 6.8 and Figure 6-2), hence the region above the diagonal corresponds to a pure higgsino LSP with mass J-L. For Au = Ad = 0.1 the yukawa interactions are irrelevant and the LSP dominantly annihilates by t-channel neutral (charged) higgsino exchange to Z Z (WW) pairs. Charginos, which have a tree-level 161 Ann 200( 2000 100( u 1000 1 IVi 7MnN ---- I . I ' . . . . '--- .. 1f000 . .... ---- Afl - -- - I (GeV) S (GeV) (b) 0=7r (a) 0=0 Figure . 6-2: Gaugino fraction contours for A, = Ad = 0.88 and 0=0 (left),7r (right). mass /u and are almost degenerate with the LSP, coannihilate with it, decreasing the relic density by a factor of 3. This fixes the LSP mass to be around A= 1 TeV, giving rise to the wide vertical band that can be seen in the figure; for smaller /u the LSP over-annihilates, for larger /i it does not annihilate enough. Increasing the yukawa couplings increases the importance of t-channel bino exchange to higgs pairs. Notice that taking the limit M1 >> u makes this new interaction irrelevant, therefore the allowed region converges to the one in which only gauge interactions are effective. Taking this as our starting point, as we approach the diagonal the mass of the bino decreases, causing the t-channel bino exchange process to become less suppressed and increasing the total annihilation cross-section. This explains the shift to higher masses, which is more pronounced for larger yukawas as expected and peaks along the diagonal where the higgsino and bino are degenerate and the bino mass suppression is minimal. The increased coannihilation between higgsinos and binos close to the diagonal does not play a large part here since both particles have access to a similar t-channel diagram. Taking 0 = r makes little qualitative difference when either of the yukawas is small compared to M1 or ,/, since in this limit the angle is unphysical and can be rotated 162 away by a redefinition of the higgsino fields. However we can see in Figure 6-2 that for large yukawas the region above the diagonal M1 = changes to a mixed state, rather than being pure higgsino as before. Starting again with the large M1 limit and decreasing M1 decreases the mass suppression of the t-channel bino exchange diagram like in the 0 = 0 case, but the LSP also starts to mix more with the bino, an effect that acts in the opposite direction and decreases ()eff. This effect happens to outweigh the former, forcing the LSP to shift to lower masses in order to annihilate enough. With 0 = 7r and yukawas large enough, there is an additional allowed region for , < Mw. In this region the higgsino LSP is too light to annihilate to on-shell gauge bosons, so the dominant annihilation channels are phase-space suppressed. Furthermore if the splitting between the chargino and the LSP is large enough, the effect of coannihilation with the chargino into photon and on-shell W is Boltzmann suppressed, substantially decreasing the effective cross-section, and giving the right relic abundance even with such a light higgsino LSP. Although acceptable from a cosmological standpoint, this region is excluded by direct searches since it corresponds to a chargino that is too light. 6.3.2 Bino Dark Matter The region below the diagonal M1 = ,pcorresponds to a bino-like LSP. Recall that in the absence of yukawa couplings pure binos in this model do not couple to anything and hence cannot annihilate at all. Turning on the yukawas allows them to mix with higgsinos which have access to gauge annihilation channels. For Au = Ad = 0.1 this effect is only large enough when M1 and p are comparable (in fact when they are equal, the neutralino states are maximally mixed for arbitrarily small off-diagonal terms), explaining the stripe near the diagonal in Figure 6-1. Once pugets larger than - 1 TeV even pure higgsinos are too heavy to annihilate efficiently; this means that mixing is no longer sufficient to decrease the dark matter relic density to acceptable values and the stripe ends. Increasing the yukawas beyond a certain value (Au = 163 Ad = 0.88, which is slightly larger than their values in Split SUSY, is enough), makes t-channel annihilation to higgses become large enough that a bino LSP does not need to mix at all in order to have the correct annihilation cross-section. This gives rise to an allowed region which is in the shape of a stripe, where for fixed Ml the correct annihilation cross-section is achieved only for the small range of that gives the right t-channel suppression. As M1 increases the stripe converges towards the diagonal in order to compensate for the increase in LSP mass by increasing the cross-section. Once the diagonal is reached this channel cannot be enhanced any further, and there is no allowed region for heavier LSPs. In addition the cross-section for annihilation through an s-channel resonant higgs, even though CP suppressed (see Section 6.5 for details), becomes large enough to allow even LSPs that are very pure bino to annihilate in this way. The annihilation rate for this process is not very sensitive to the mixing, explaining the - 80 GeV. This line ends when Aigrows to apparent horizontal line at M1 = 21Mhiggs the point where the mixing is too small. As in the higgsino case, taking 0 = 7rchanges the shape of the contours of constant gaugino fraction and spreads them out in the plane (see Figure 6-2), making mixing with higgsinos relevant throughout the region. For small M1 , the allowed region starts where the mixing term is small enough for the combination of gauge and higgs channels not to cause over-annihilation. Increasing M1 again makes the region move towards the diagonal, where the increase in LSP mass is countered by increasing the cross-section for the gauge channel from mixing more. For either yukawa very large (Au = 1.38, Ad = 0.1), annihilation to higgses via t-channel higgsinos is so efficient that this process alone is sufficient to give bino-like LSPs the correct abundance. As M1 increases the allowed region again moves towards the diagonal in such a way as to keep the effective cross-section constant by decreasing the higgsino mass suppression, thus compensating for the increase in LSP mass. As we remarked earlier since Ad is effectively zero in this case, the angle 0 is unphysical and can be rotated away by a redefinition of the higgsino fields. 164 6.4 Direct Detection Dark matter is also detectable through elastic scatterings off ordinary matter. The direct detection cross-section for this process can be divided up into a spin-dependent and a spin-independent dominant. part; we will concentrate As before we restrict to B = significantly for intermediate The spin-independent a and 1r, on the former since it is usually we expect the result not to change values. interaction takes place through higgs exchange, VIa the yukawa couplings which mix higgsinos and binos. Since the only X~X~ h term in our model involves the product of the gaugino and higgsino fractions, the more mixed our dark matter is the more visible it will be to direct detection experiments. This effect can be seen in Fig 6-3 below. Current bound ----- 1e-42 , -------------------------------- IS """"" I - " ,,~ ~ 1e-45 ,: N \ 6 Proposed bound from next-generation experiments _ - .:::::-,- ---------------------- ------- --- ------- -: '2 o Q) C3 =' >:: 1e-48 • • >< b 0=0, >.,,=0.88, >',J=0.88 0=7r, >.,,=0.88, >',J=0.88 • 0=0, >.,,=1.38, >'rI=O.l • 0=7r, >.,,=0.1, >',J=0.1 Ie-51 1e-540 500 1000 1500 2000 2500 3000 Mx (GeV) Figure 6-3: Spin-independent part of dark matter direct detection cross-section. The cu~rent bound represents the CDMS limit [11], and, as an indicative value for the proposed bound from next generation experiments, we take the projected sensitivity of SuperCDMS phase B [18]. Although it seems like we cannot currently use this measure as a constraint, proportion of our parameter space will be accessible at next-generation Since higgsino LSPs are generally more pure than bino-type 165 the major experiments. ones, the former will escape detection as long as there is an order 100 keV splitting between its two neutral components. This is is necessary in order to avoid the limit from spin-independent direct detection measurements [12]. Also visible in the graph are the interesting discontinuities mentioned in [13], corresponding to the opening up of new annihilation channels at MLSP = 1/2Mhiggs through an s-channel higgs. We also notice a similar discontinuity at the top threshold from annihilation to tt; this effect becomes more pronounced as the new yukawa couplings increase. 6.5 Electric Dipole Moment Since our model does not contain any sleptons it induces an electron EDM only at two loops, proportional to sin(O) for 0 as defined above. This is a two-loop effect, we therefore expect it to be close to the experimental bound for 0(1) 0. The dominant diagram responsible for the EDM is generated by charginos and neutralinos in a loop and can be seen in Figure 6-4 below. This diagram is also present in Split SUSY where it gives a comparable contribution to the one with only charginos in the loop [19, 20]. f' f Figure 6-4: The 2-loop contribution to the EDM of a fermion f. The induced EDM is (see [19]): dw fd _ e where . a2mf 1 Im ((OLOR*)5 0 L 0 R*)g (r, r) 24 f 2i 3mx/ mx'l~ 2 8wr shMe 166 (6.2) G(r° r) = dzI idyY l diy I/ - ld dy yz(y+z/2) yZ(K + K) + ((Ky)(zyK3 [ (y - 3Ki)y 2y) n Ki] 3 4y(Ki+- 2(Ki y) 2 + y)y + Ki(Ki 2(Ki -- y) Y and 17= _r- +r-, oR = VNiexp r- 2 r mx _- ,oi OL=N3i NTMNN = diag(mxl, mx 2 , mx) with real and positive diagonal elements. The sign on the right-hand side of equation (6.2) corresponds to the fermion f with weak isospin i and f' is its electroweakpartner. In principle it should be possible to cross-correlate the region of our parameter space which is consistent with relic abundance measurements, with that consistent with electron EDM measurements in order to further constrain our parameters. However since the current release of DarkSUSY does not support CP violating phases and a version including CP violations seems almost ready for public release4 we leave an accurate study of the consequences of non-zero CP phase in relic abundance and direct detection calculations for a future work. We can still draw some interesting conclusions by estimating the effect of non-zero CP phase. Because there is no reason for these new contributions to be suppressed with respect to the CP-conserving ones (for of 0(1)), we might naively expect their inclusion to enhance the annihilation cross-section by around a factor of 2, increasing the acceptable LSP masses by V2 for constant relic abundance. This is discussed in greater detail in [21] (and [22] for direct detection) in which we see that this observation holds for most of the parameter space. We must note, however, that in particular small regions of the space the enhancement to the annihilation cross-section and the suppression to the elastic cross section can be much larger, justifying further investigation of this point in future 4 Private communication with one of the authors of DarkSUSY. 167 work. With this assumption in mind we see in Figure 6-5 that although the majority of our allowed region is below the current experimental limit of de < 1.7 at 95% C.L. [23], most of it will be accessible to next generation 10-27 e cm X EDM experiments. These propose to improve the precision of the electron EDM measurement by 4 orders of magnitude in the next 5 years, and maybe even up to 8 orders of magnitude, ing permitting fund- [24, 25, 26]. We also see in this figure that CP violation is enhanced on the diagonal where the mixing is largest. This is as expected since the yukawas that govern the mixing are necessary for there to be any CP violating phase at all. For the same reason, decoupling either particle sends the EDM to zero . 3000 1 x 10 27 • 0=0, Au=O.88, Ad=O.88 • O=7r, Au=O.88, Ad=O.88 , \ \ ~. 1.7 X 10-27 " 2000 > 0 v '--" ~ 1000 ....• 3 x 10-27 -. 1 X 10-28 500 1500 1000 J-1 2000 2500 3000 (GeV) Figure 6-5: Electron edm contours for 0 = 7r /2. The excluded region is bounded by the black contours. Note that CP violation was not included in the relic density calculation, and the dark matter plot is simply intended to indicate the approxi~ate region of interest for dark matter. 6.6 Gauge Coupling Unification In this section we study the running of gauge couplings in our model at two loops. The addition of higgsinos largely improves unification as compared to the 8M case, 168 but their effect is still not large enough and the model predicts a value for a,(Mz) around 9a lower than the experimental value of as(Mz) = 0.119i0.002 [27]. Moreover the scale at which the couplings unify is very low, around 1014 GeV, making proton decay occur much too quickly to embed in a simple GUT theory 5 . These problems can be avoided by adding the Split SUSY particle content at higher energies, as in [29], at the cost of losing minimality; instead we choose to solve this problem by embedding our minimal model in an extra dimensional theory. This decision is well motivated: even though normal 4D GUTs have had some successes, explaining the quark-lepton mass relations for example, and charge quantization in the SM, there are many reasons why these simple theories are not ideal. In spite of the fact that the matter content of the SM falls naturally into representations of SU(5), there are some components that seem particularly resistant to this. This is especially true of the higgs sector, the unification of which gives rise to the doublet-triplet splitting problem. Even in the matter sector, although b-- unification works reasonably well the same cannot be said for unification of the first two generations. In other words, it seems like gauge couplings want to unify while the matter content of the SM does not, at least; not to the same extent. This dilemma is easily addressed in an extra dimensional model with a GUT symmetry in the bulk, broken down to the SM gauge group on a brane by boundary conditions [30] since we can now choose where we put fields based on whether they unify consistently or not. Unified matter can be placed in the bulk whereas non-unified matter can be placed on the GUT-breaking brane. The low energy theory will then contain the zero modes of the 5D bulk fields as well as the brane matter. While solving many of the problems of standard 4D GUTs these extra dimensional theories have the drawback of having a large number of discrete choices for the location of the matter fields, as we shall see later. We will consider a model with one flat extra dimension compactified on a circle of radius R, with orbifolds S1 /(Z 2 x Z2), whose action is obtained from the identifications 5It is possible to evade the constraint from proton decay by setting some of the relevant mixing parameters to zero [28]. However we are not aware of any GUT model in which such an assertion is justified by symmetry arguments. 169 Z2 : - 2r R - y, Z2' :y -- 7r , y [0, 277r (6.3) where y is the coordinate of the fifth dimension. There are two fixed points under this action, at (0, w7R)and (rR/2, 37rR/2), at which are located two branes. We impose an SU(5) symmetry in the bulk and on the y = 0 brane; this symmetry is broken down to the SM SU(3) x SU(2) x U(1) on the other brane by a suitable choice of boundary conditions. All fields need to have definite transformation properties under the orbifold action - we choose the action on the fundamental to be ±P -- and on the adjoint, ±[P,A], for projection operators Pz = (+, +, +, +, +) and Pz, = (+, +, +,-,-). Aa for a = This gives SM gauge fields and their corresponding KK towers 1, ..., 12} (+, +) parity; and the towers Ad for = (13,..., 24} (+,-) parity, achieving the required symmetry-breaking pattern. By gauge invariance the unphysical fifth component of the gauge field, which is eaten in unitary gauge, gets opposite boundary conditions. 6 We still have the freedom to choose the location of the matter fields. In this model SU(5)-symmetric matter fields in the bulk will get split by the action of the Z' orbifold: the SM 5 for instance will either contain a massless d or a massless 1, with the other component only having massive modes. Matter fields in the bulk must therefore come in pairs with opposite eigenvalues under the orbifold projections, so for each SM generation in the bulk we will need two copies of 10 + 5. This provides us with a simple mechanism to forbid proton decay from Xand Y-exchange and also to split the color triplet higgs field from the doublet. To summarize, unification of SM matter fields in complete multiplets of SU(5) cannot be achieved in the bulk but on the SU(5) brane, while matter on the SM brane is not unified into complete GUT representations. 6 From an effective field theory point of view an orbifold is not absolutely necessary, our theory can simply be thought of as a theory with a compact extra dimension on an interval, with two branes on the boundaries. Because of the presence of the boundaries we are free to impose either Dirichlet or Neumann boundary conditions for the bulk field on each of the branes breaking the SU(5) to the SM gauge group purely by choice of boundary conditions and similarly splitting the multiplets accordingly. Our orbifold projection is therefore nothing more than a further restriction to the set of all possible choices we can make. 170 6.6.1 Running and matching We run the gauge couplings from the weak scale to the cutoff A by treating our model as a succession of effective field theories (EFTs) characterized by the differing particle content at different energies. The influence of the yukawa couplings between the higgsinos and the singlet on the two-loop running is negligible, hence it is fine to assume that the singlet is degenerate with the higgsinos so there is only one threshold from Mtop to the compactification scale 1/R, at which we will need to match with the full 5D theory. The SU(5)-symmetric bulk gauge coupling g5 can be matched on to the low energy couplings at the renormalization scale M via the equation 1 27rR -(M) - + Ai(M)+ Ai(MR) (6.4) The first term on the right represents a tree level contribution from the 5d kinetic term, Ai are similar contributions from brane-localized kinetic terms and Ai encode radiative contributions from KK modes. The latter come from renormalization of the 4D brane kinetic terms which run logarithmically as usual. To understand this in more detail let us consider radiative corrections to a U(1) gauge coupling in an extra dimension compactified on a circle with no orbifolds, due to a 5D massless scalar field [32]. Since 1/g' has mass dimension 1, by dimensional analysis we might expect corrections to it to go like A + m log A where m is some mass parameter in the theory. The linearly divergent term is UV sensitive and can be reabsorbed into the definition of g5 , whereas the log term cannot exist since there is no mass parameter in the theory. Hence the 5D gauge coupling does not run, and neither does the 4D gauge coupling. This can also be interpreted from a 4D point of view, where the KK partners of the scalar cut off the divergences of the zero mode. Since there is no distinction between the wavefunctions for even (cosine) and odd (sine) KK modes in the absence of an orbifold, and we know that the sum of their contributions must cancel the log divergence of the 4D massless scalar, each of these must give a contribution equal to -1/2 times that of the 4D massless scalar. 171 When we impose a Z2 orbifold projection and add two 3-branes at the orbifold fixed points, the scalar field must now transform as an eigenstate of this orbifold action and can either be even ((+, +), with Neuman boudary conditions on the branes), or odd ((-,-), with Dirichlet boundary conditions on the branes). This restricts us to a subset of the original modes and the cancellation of the log divergence no longer works. Since this running can only be due to 4D gauge kinetic terms localized on the branes, where the gauge coupling is dimensionless and can therefore receive logarithmic corrections, locality implies that the contribution from a tower of states on a particular brane can only be due to its boundary condition on that brane, with the total running equal to the sum of the contributions on each brane. In fact, it is only in the vicinity of the brane that imposing a particular boundary condition has any effect. As argued above, a (+, +) tower (excluding the zero mode) and a (-, -) tower must each give a total contribution equal to -1/2 times that of the zero mode, which corresponds to a coefficient of -1/4 to the running of each brane-localized kinetic term. Taking into account the contribution of the zero mode we can say that a tower of modes with + boundary conditions on a brane contributes +1/4 times the corresponding 4D coefficient, while a - boundary condition contributes -1/4 times the same quantity. This argument makes it explicit that the orbifold projection can be seen as a prescription on the boundary conditions of the fields in the extra dimension, which only affect the physics near each brane. Adding another orbifold projection as we are doing in this case also allows for towers with (+, -) and (-, +) boundary conditions which, from the above argument, both give a contribution of ±1/4 :F 1/4 = 0. The contribution of the (+, +) and (-, -) towers clearly remains unchanged. Explicitly integrating out the KK modes at one loop at the compactification scale allows us to verify this fact, and also compute the constant parts of the threshold corrections, which are scheme-dependent. In DR 7 we obtain [32]: 7 We use this renormalization scheme even though our theory is non-supersymmetric threshold corrections in this scheme contain no constant part [31]. 172 since 4D Ai = 1(MR) ((bs-21bq+8bF)Fe(MR)+ ( - 21bG + 8biF) F ) (6.5) with Fe(iR) = I-11 +00 Z= - log(r)- dt (t- 1 + t- log(MR), 1/ 2 ) Fo = - log(2) (6.6) +oo (3(it) - 1) ~ 0.02, 03(it) = en 2 n=--oo where bi, 'G 'F (,GF) are the Casimirs of the KK modes of real scalars (not including goldstone bosons), massive vector bosons and Dirac fermions respectively with even (odd) masses 2n/R ((2n + 1)/R). As explained above, the logarithmic part of the above expression is equal to exactly -1/2 times the contribution of the same fields in 4D [33, 34]. Since the compactification scale 1/R will always be relatively close to the unification scale A (so our 5D theory remains perturbative), it will be sufficient for us to use one loop matching in our two loop analysis as long as the matching is done at a scale M close to the compactification scale. As an aside, from equation (6.4) we can get: d dAi = bi - b2M M 8wr (6.7) where bi is shorthand for the combination (bS - 21bg + 8bF)/12 and biMM are the coefficients of the renormalization group equations below the compactification scale (see Appendix 6.9 for details). It is clear from this equation that it is unnatural to require Ai(1/R) < 1/(87r 2 ). The most natural assumption Ai(A) 1/(8wr2 ) gives a one-loop contribution comparable to the tree level term, implying the presence of some strong dynamics in the brane gauge sector at the scale A. We know that the 5D gauge theory becomes strong at the scale 247r3 /g52,so from naive dimensional analysis (NDA) (see for example [35]) we find that it is quite natural for A to coincide with the strong coupling scale for the bulk gauge group. 173 Running equation (6.7) to the compactification scale we obtain Ai(l/R) = i(A) + bi - bM 8 log(AR) (6.8) For AR > 1 the unknown bare parameter is negligible compared to the log-enhanced part, and can be ignored, leaving us with a calculable correction. 27rR/g2 we expect AR Using GUT = 87r2 /gUT - 100. Keeping this in mind, we shall check whether unification is possible in our model with AR in the regime where the bare brane gauge coupling is negligible. To this purpose we will impose the matching equation (6.4) at the scale A assuming Ai(A) = 0; we will then check whether the value of AR found justifies this approximation. In order to develop some intuition for the direction that these thresholds go in, we can analyze the one-loop expression (with one-loop thresholds) for the gauge couplings at Mz: aqI(M) = '' + 47rAi(AR)+ Aonv(AR) + log + ) (6.9) bsM are the SM beta function coefficients (see Appendix 6.9), /u is the scale of the = ( higgsinos and singlet, AConV 3-, 2, 0) are conversion factors from MS, in which the low-energy experimental values for the gauge couplings are defined, to DR [36]. Taking the linear combination (9/14)a- 1 - (23/14)a21 + a3 1 allows us to eliminate the A dependence as well as all SU(5)-symmetric terms, leaving 1 a (M-,) 9/14 al(M) 23/14 + 4(AR) + log conv +) 2 O( (6.10) lo where X = (9/14)X 1 - (23/14)X2 + X3 for any quantity X. Recall that the leading threshold correction from the 5D GUT is proportional to log(AR). The low-energy value of a3 is therefore changed by 174 Ja3(MZ) = 2 3 (Mz)) 27r log(AR) (6.11) We still have the freedom to choose the positions of the various matter fields. In order to determine the best setup for gauge coupling unification we need to keep in mind two facts: the first is that adding SU(5) multiplets in the bulk does not have any effect on a3(Mz); and the second is that b contains only contributions from (+, +) modes (in unitary gauge none of our bulk modes have (-,-) our SU(5) bulk multiplets are split into (+, +) and (+,-) boundary conditions; modes). As stated at the beginning of this section our 4D prediction for a 3 (Mz) is too low. Since fermions have a larger effect on running than scalars, this problem is most efficiently tackled by splitting up the fermion content of the SM into non-SU(5) symmetric parts in order to make b as positive as possible. Examining the particular linear combination that eliminated the dependence on A at one loop we find that one or more SU(5)-incomplete colored multiplets are needed in the bulk, or equivalently the weakly-interacting part of the same multiplet has to be on one of the branes. Since matter in the bulk is naturally split by the orbifold projections, this just involves separating the pair of multiplets whose zero modes make up one SM family. With this in mind we find that for fixed AR and , since separating different numbers of SM generations allows us to vary the low energy value of a3 anywhere from its experimental value to several as off, gauge coupling unification really does work in this model for some fraction of all available configurations. Although this may seem a little unsatisfactory from the point of view of predictivity, the situation can be somewhat ameliorated by further refining our requirements. For example, we can go some way towards explaining the hierarchy between the SM fermion masses by placing the first generation in the bulk, the second generation split between the bulk and a brane and the third generation entirely on a brane. This way, in addition to breaking the approximate flavor symmetry in the fermion sector we also obtain helpful factors of order 1/v'K_ between the masses of the different generations. The location of the higgs does not have a very large effect on unification, the simplest 175 choice would be to put it, as well as the higgsinos and singlet, on the SU(5)-breaking brane, where there is no need to introduce corresponding color triplet fields. This also helps to explain the hierarchy. In this model, which can be seen on the left-hand side SU(5) SU(3)xSU(2)xU(1) SU(5) H 52 102 10, 10, 102 5, I 5 I'0 10 5: 10, irR/2 nR/2 (a) Higgs on the broken brane (b) Higgs in the bulk and third generation on the SU(5) brane Figure 6-6: Matter content of the two orbifold GUT models we propose. of Figure 6-6, we also need to put our third generation and split second generation on the same brane in order for them to interact with the higgs. On the right is another model which capitalizes on every shred of evidence we have about GUT physics: we put the higgs in the bulk in this case (recall that the orbifold naturally gives rise to doublet-triplet splitting) so that we can switch the third generation to the SU(5)-preserving brane and obtain b-r unification (see Figure 6-7) without having analogous relationships for the other two generations8 . We also need to flip the positions of the first and second generations if we want to keep the suppression of the mass of the first generation with respect to the second. The low-energy values for a3 as a function of p in these two models can be seen in Figure 6-8 for different AR. Note that unification can be acheived in the regime 8 As explained in [37] the two yukawa couplings Ab and AT, run differently only below the compactification scale. Because of locality, the fields living on the SU(5) brane do not feel the SU(5) breaking until energies below the compactification scale; hence if they are unified at some high energy they keep being unified until this scale. 176 >.,,(Mz) 0.028 0.027 0.026 40' interval ' ~=------,-- 0.025 0.024 -- , --------,---======== -------------- AR=1oo AR=lO 4D 0.023 0.022 1000 2000 3000 p. (GeV) 4000 5000 Figure 6-7: Low energy prediction for Ab(Mz), as a function of the higgsino mass J.L, for the model with the biggs in the bulk, for AR = 10,100 and for the 4D model. 40' interval taken from [27). where AR » 1, justifying our initial assumption that the brane kinetic terms could be neglected. We see that although the dependence on J.L is very slight, small J.L seems to be preferred. However we cannot use this observation to put a firm upper limit on J.L because of the uncertainties associated with ignoring the bare kinetic terms on the branes. l)3(Mz) 0.125 :erval I 0.12 0.115 AR=lOO 0.11 AR=lO 0.105 --- 1000 0.095 2000 ------ 3000 4000 5000 l' (GeV) ................................................................................................... 4D Figure 6-8: Low energy prediction for Q:3(Mz), as a function of the higgsino mass J.L, for the model with the higgs on the brane (black line), and for the model with the higgs in the bulk (green line), for AR = 10,100, and for the 4D case (dashed line). Some typical values of 1/ R are shown. The second configuration, Figure 6-6(b), also gives proton decay through the mixing of the third generation with the first two. From our knowledge of the CKM matrix we infer that all mixing matrices will be close to the unit matrix, proton 177 decay will therefore be suppressed by off-diagonal elements. To minimize this suppression it is best to have an anti-neutrino and a strange quark in the final state. The proton decay rate for this process was computed in [38] and is proportional ((9412 ) to (I - n) [ I(Rt) 23 (Rti)l 3 (Ld) 3 12 where Rd,,u and Ld are the rotation matrices of the right-handed down-type and up-type quarks, and the left-handed down-type quarks, which are unknown. We assume that the 2-3 and the 1-3 mixing elements are 0.05 and 0.01 respectively, similar to the corresponding CKM matrix elements, giving a proton lifetime of -p(P - K+iP) _ 6.6 x 1038 years x 1014Ge V 4 x 10 35 years (6.12) for 1/R = 1.6 x 1013GeV. This is above the current limit from Super-Kamiokande of 1.9 x 1033 years at 90% C.L. [39, 40], although there are multi-megaton experiments in the planning stages that are expected to reach a sensitivity of up to 6 x 1034 years [40] . Given our lack of information about the mixing matrices involved9 , we see that there might be some possibility that proton decay in this model will be seen in the not-too-distant future. 6.7 Conclusion The identification of a TeV-scale weakly-interacting particle as a good dark matter candidate, and the unification of the gauge couplings are usually taken as indications of the presence of low-energy SUSY. However this might not necessarily be the case. If we assume that the tuning of the higgs mass can be explained in some other unnatural way, through environmental reasoning for instance, then new possibilities open up for physics beyond the SM. In this chapter we studied the minimal model consistent with current experimental limits, that has both a good thermal dark matter candidate and gauge coupling unification. 9 To this end we added to the SM two Experiments have only constrained the particular combination that appears in the SM as the CKM matrix. 178 higgsino-like particles and a singlet, with a singlet majorana mass of < 100 TeV in order to split the two neutralinos and so avoid direct detection constraints. Making the singlet light allowed for a new region of dark matter with mixed states as heavy as 2.2 TeV, well beyond the reach of the LHC and the generic expectation for a weakly interacting particle. Nevertheless we do have some handles on this model: firstly via the 2-loop induced electron EDM contribution which is just beyond present limits for CP angle of order 1, and secondly by the spin-independent direct detection cross section, both of which should be accessible at next-generation experiments. Turning to gauge coupling unification we saw that this was much improved at two loops by the presence of the higgsinos. A full 4D GUT model is nevertheless excluded by the smallness of the GUT scale - 1014GeV, which induces too fast proton decay. We embedded the model in a 5D orbifold GUT in which the threshold corrections were calculable and pushed ca in the right direction for unification (for a suitable matter configuration). It is very gratifying that such a model can help explain the pattern in the fermion mass hierarchy, give b-- unification, and predict a rate for proton decay that can be tested in the future. 6.8 Appendix A: The neutralino mass matrix We have a 3x3 neutralino mass matrix in which the mixing terms (see equation (6.2)) are unrelated to gauge couplings and are limited only by the requirement of perturbativity to the cutoff. It is possible to get a feel for the behavior of this matrix by finding the approximate eigenvalues and eigenvectors in the limit of equal and small off-diagonal terms (Av << M1 q±-/). The approximate eigenvalues and eigenvectors are shown in the table below: M2 M12 + 4A 2 v2 Gaugino fraction M 1 Ml,.4 1-2 L2 /~2 4 2 v2 + 2 2 A v -p)2 (Ml 0 2 2 2 V 179 for cos/) = If M1 << 1. the first eigenstate will be the LSP. In the opposite limit the compo- sition of the LSP is dependent upon the sign of cos(9). For cos(9) positive the second eigenstate, which is pure higgsino, is the LSP, while for cos(O) negative, the mixed third eigenstate becomes the LSP. We also see from the table that in the pure higgsino case, a splitting of order 100 keV (sufficient to evade the direct detection constraint) can be achieved with a singlet mass lighter than 109 GeV, where the upper limit corresponds to 0(1) yukawa couplings. 6.9 Appendix B: Two-Loop Beta Functions for Gauge Couplings The two-loop RGE for the gauge couplings in our minimal model is l d (--21r)idt1 = b M + (4/¢)2 [ 3 4j7rB jMa - di ] - d)(2 with P-function coefficients 104 18 5 44 6 5 14 12 11 9 2 25 bm __ -7 BMM= 6 10 (17 3 d= 10,2,2); 5 -26 d(3 d= 20'4') The running of the yukawa couplings is the same as in [9] but we will reproduce their RGEs here for convenience - we ignore all except the top yukawa coupling (we found that our two new yukawas do not have a significant effect). (47r) 2 At = At -3 E (~~~~~i=1lh[~~~ih+ 47rcii 180 t ++ (AU + Ad) withc= (17,3 ) The two-loop coupled RGEs can be solved analytically if we approximate the top yukawa coupling as a constant over the entire range of integration (see [36]for a study on the validity of this approximation). The solution is - (M) - 1 2M+ 2r 2 14 BM 47 W M A n(l+ b 47r I 181 G(A) lnM - 3 2 73 diA21nA t M Bibliography [1] S. Perlmutter et al. [Supernova Cosmology Project Collaboration], "Measurements of Omega and Lambda from 42 High-Redshift Supernovae," Astrophys. J. 517 (1999) 565, astro-ph/9812133. [2] D. N. Spergel et al. [WMAP Collaboration], "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Determination of Cosmological Parameters," Astrophys. J. Suppl. 148 (2003) 175, astro-ph/0302209. C. L. Bennett et al., "First Year Wilkinson Microwave Anisotropy Probe (WMAP) Observations: Preliminary Maps and Basic Results," Astrophys. J. Suppl. 148, 1 (2003), astro-ph/0302207. [3] R. Barbieri and A. Strumia, "What is the limit on the Higgs mass?," Phys. Lett. B 462, 144 (1999), hep-ph/9905281. [4] L. Giusti, A. Romanino and A. Strumia, "Natural ranges of supersymmetric signals," Nucl. Phys. B 550 (1999) 3, hep-ph/9811386. [5] S. Weinberg, "Anthropic Bound On The Cosmological Constant," Phys. Rev. Lett. 59 (1987) 2607. [6] V. Agrawal, S. M. Barr, J. F. Donoghue and D. Seckel, "The anthropic principle and the mass scale of the standard model," Phys. Rev. D 57, 5480 (1998), hepph/9707380. 182 [7] N. Arkani-Hamed and S. Dimopoulos, "Supersymmetric unification without low energy supersymmetry and signatures for fine-tuning at the LHC," JHEP 0506, 073 (2005) [arXiv:hep-th/0405159]. [8] N. Arkani-Hamed, S. Dimopoulos and S. Kachru, "Predictive landscapes and new physics at a TeV," arXiv:hep-th/0501082. [9] G. F. Giudice and A. Romanino, "Split supersymmetry," Nucl. Phys. B 699, 65 (2004) [Erratum-ibid. B 706, 65 (2005)], hep-ph/0406088. [10] M. W. Goodman and E. Witten, "Detectability Of Certain Dark-Matter Candidates," Phys. Rev. D 31, 3059 (1985). [11] D. S. Akerib et al. [CDMS Collaboration], "Limits on spin-independent WIMP nucleon interactions from the two-tower run of the Cryogenic Dark Matter Search," arXiv:astro-ph/0509259. [12] D. R. Smith and N. Weiner, "Inelastic dark matter," Phys. Rev. D 64, 043502 (2001), hep-ph/0101138. [13] A. Pierce, "Dark matter in the finely tuned minimal supersymmetric standard model," Phys. Rev. D 70 (2004) 075006, hep-ph/0406144. [14] A. Masiero, S. Profumo and P. Ullio, "Neutralino dark matter detection in split supersymmetry scenarios," Nucl. Phys. B 712, 86 (2005) [arXiv:hep-ph/0412058]. L. Senatore, "How heavy can the fermions in split SUSY be? A study on gravitino and extradimensional LSP," Phys. Rev. D 71, 103510 (2005) [arXiv:hepph/0412103]. [15] P. Gondolo, J. Edsjo, P. Ullio, L. Bergstrom, M. Schelke and E. A. Baltz, "Dark- SUSY: A numerical package for supersymmetric dark matter calculations," ,astroph/0211238; J. Edsjo, M. Schelke, P. Ullio and P. Gondolo, "Accurate relic densities with neutralino, chargino and sfermion coannihilations in mSUGRA," JCAP 0304, 001 (2003), hep-ph/0301106]; P. Gondolo and G. Gelmini, "Cosmic Abundances Of Stable Particles: Improved Analysis," Nucl. Phys. B 360, 145 (1991). 183 [16] A. Provenza, M. Quiros and P. Ullio, "Electroweak baryogenesis, large Yukawas and dark matter," JHEP 0510, 048 (2005) [arXiv:hep-ph/0507325]. [17] J. Edsjo and P. Gondolo, "Neutralino relic density including coannihilations," Phys. Rev. D 56, 1879 (1997), hep-ph/9704361]; [18] P. II L. Brink dark matter et al. [CDMS-II search: Collaboration], SuperCDMS," "Beyond the eConf C041213, CDMS- 2529 (2004) [arXiv:astro-ph/0503583]. Useful comparison can be done using the tools of http://dendera.berkeley.edu/plotter/entryform.html [19] D. Chang, W. F. Chang and W. Y. Keung, "Electric dipole moment in the split supersymmetry models," Phys. Rev. D 71 (2005) 076006, hep-ph/0503055. [20] N. Arkani-Hamed, S. Dimopoulos, G. F. Giudice and A. Romanino, "Aspects of split supersymmetry," Nucl. Phys. B 709, 3 (2005) [arXiv:hep-ph/0409232]. [21] P. Gondolo and K. Freese, "CP-violating effects in neutralino scattering and annihilation," JHEP 0207, 052 (2002), hep-ph/9908390. [22] T. Falk, A. Ferstl and K. A. Olive, "Variations of the neutralino elastic crosssection with CP violating phases," Astropart. Phys. 13, 301 (2000) [arXiv:hepph/9908311]. [23] B. C. Regan, E. D. Commins, C. J. Schmidt and D. DeMille, "New limit on the electron electric dipole moment," Phys. Rev. Lett. 88 (2002) 071805. [24] D. Kawall, F. Bay, S. Bickman, Y. Jiang and D. DeMille, "Progress towards measuring the electric dipole moment of the electron in metastable PbO," AIP Conf. Proc. 698 (2004) 192. [25] S. K. Lamoreaux, "Solid state systems for electron electric dipole moment and other fundamental measurements," nucl-ex/0109014. [26] Y. K. Semertzidis, "Electric dipole moments of fundamental arXiv:hep-ex/0401016. 184 particles," [27] S. Eidelman et al. [Particle Data Group], "Review of particle physics," Phys. Lett. B 592 (2004) 1. [28] I. Dorsner and P. F. Perez, "How long could we live?," Phys. Lett. B 625, 88 (2005) [arXiv:hep-ph/0410198]. [29] I. Antoniadis, A. Delgado, K. Benakli, M. Quiros and M. Tuckmantel, "Splitting extended supersymmetry," arXiv:hep-ph/0507192. [30] Y. Kawamura, "Gauge symmetry reduction from the extra space S(1)/Z(2)," Prog. Theor. Phys. 103, 613 (2000), hep-ph/9902423. A. Hebecker and J. MarchRussell, "A minimal S(1)/(Z(2) x Z'(2)) orbifold GUT," Nucl. Phys. B 613, 3 (2001), hep-ph/0106166. L. J. Hall and Y. Nomura, "Gauge coupling unification from unified theories in higher dimensions," Phys. Rev. D 65, 125012 (2002), hep-ph/0111068. G. Altarelli and F. Feruglio, "SU(5) grand unification in extra dimensions and proton decay," Phys. Lett. B 511, 257 (2001), hep-ph/0102301. L. J. Hall and Y. Nomura, "Grand unification in higher dimensions," Annals Phys. 306, 132 (2003), hep-ph/0212134. P. C. Schuster, "Grand unification in higher dimensions with split supersymmetry," hep-ph/0412263. A. B. Kobakhidze, "Proton stability in TeV-scale GUTs," Phys. Lett. B 514, 131 (2001) [arXiv:hepph/0102323]. [31] I. Antoniadis, C. Kounnas and K. Tamvakis, "Simple Treatment Of Threshold Effects," Phys. Lett. B 119, 377 (1982). [32] R. Contino, L. Pilo, R. Rattazzi and E. Trincherini, "Running and matching from 5 to 4 dimensions," Nucl. Phys. B 622, 227 (2002), hep-ph/0108102. [33] S. Weinberg, "Effective Gauge Theories," Phys. Lett. B 91, 51 (1980). [34] L. J. Hall, "Grand Unification Of Effective Gauge Theories," Nucl. Phys. B 178, 75 (1981). 185 [35] Z. Chacko, M. A. Luty and E. Ponton, "Massive higher-dimensional gauge fields as messengers of supersymmetry breaking," JHEP 0007, 036 (2000), hepph/9909248. [36] P. Langacker and N. Polonsky, "Uncertainties in coupling constant unification," Phys. Rev. D 47, 4028 (1993), hep-ph/9210235. [37] L. J. Hall and Y. Nomura, "A complete theory of grand unification in five dimensions," Phys. Rev. D 66, 075004 (2002), hep-ph/0205067. [38] A. Hebecker and J. March-Russell, "Proton decay signatures of orbifold GUTs," Phys. Lett. B 539, 119 (2002), hep-ph/0204037. [39] Y. Hayato et al. [Super-Kamiokande Collaboration], "Search for proton decay through p - anti-nu K+ in a large water Cherenkov detector," Phys. Rev. Lett. 83 (1999) 1529, hep-ex/9904020. [40] Y. Suzuki et al. [TITAND Working Group], "Multi-Megaton water Cherenkov detector for a proton decay search: TITAND (former name: TITANIC)," hepex/0110005. 186 Chapter 7 Conclusions I have reported on some of the research work I have done during my years as a graduate student. My research interest has focused mainly on the connections between Particle Physics and Cosmology, of which I think I have been able to explore many of the different aspects. Thanks to the very recent improvement in the experimental techniques in Observational Cosmology, the field of Cosmology has become a field where new observations can motivate new ideas from Theoretical Physics, as for example with the observation of Dark Matter, of the tyne anisotropy of the CMB, and recently of the acceleration of the universe. Viceversa it has become also a field where new ideas from Theoretical Physics can be applied to the early universe with the purpose either of testing them against some phases that we consider well understood, or of gaining through them some insight on phases less understood, as for example is the case of the inflationary phase. During my research work, I think I have been able to deepen most of these aspects. Concerning that part which deals with using new high energy theories to develop new possible models for the early universe, in chapter 2 I have shown a model for the early universe where a new kind of inflation is developed. This new inflationary model is based on the condensate of a ghost field, which is a new mechanism for obtaining a stable Lorentz violating state in quantum field theory which has been 187 understood only very recently. When applied to the inflationary phase, the ghost condensate predicts some new features, such as a large amount of non-Gaussianities, which, because of the large information this kind of signal carries on, if detected, it would both tell us with great confidence that inflation did occur in the past, and also that it occurred in a way very similar to the one of ghost condensation. This inflationary model is also the first one in which it is shown that the Hubble rate H can actually grow with time. The work on non-Gaussianities has found a natural development in the direct analysis of the data from the most advanced experiment on the CMB: the WMAP experiment, as I have shown in chapter 5. Experimental data are at the origin of Physics, and it is really emotional to be able to deal with them. The work on the analysis of the non-Gaussian signal in the WMAP data originates from a long theoretical work during which the generation of a non-Gaussian inflationary signal in the CMB and its importance as a possible smoking gun for inflation was progressively understood. In the very last few years, several inflationary models have emerged which are distinguishable from the standard slow roll scenario exactly for the predicted amount and characteristics of the non-Gaussian signal. After this theoretical work, my collaborators and I realized that in the WMAP data there was more information on the non-Gaussian signal than what had already been extracted. We therefore decided to develop a method to better exploit the experimental data, and we directly did the analysis. No evidence of a non-Gaussian signal was found, but we improved and completed the available limits, so that the limits we give are at present the best currently available. This last work represents to me that part of the field of Cosmology in which new theoretical ideas begin to be tested by new cosmological observations. In this thesis I have shown some other works in this direction, in which some other aspects of the history of the universe that we consider well understood are used to constraint models of Physics beyond the Standard Model. In chapters 3,4 and 6, I have shown how some observational constraints coming from Dark Matter relic abundance, Baryogenesis, or Big Bang Nucleosynthesis, can 188 be applied to test ideas motivated by the Landscape. The Landscape itself, istead, represent the most recent important idea in Theoretical Physics which is, at least partially, motivated by Cosmological Observations. In fact, the Landscape is a scenario which has recently emerged from String Theory in which the UV theory which describes our universe has a huge number of different low energy configurations, in which the parameters that we are able to observe scan from vacuum to vacuum. Together with the cosmological observation that the cosmological constant in our universe is not-null, and that it is roughly comparable to the upper limit it has in order for structures to be present in our universe, these theoretical considerations have led to the proposal that the reason why the cosmological constant (as well as the higgs mass) are small is not dynamical, as usually thought, but rather environmental, where by environmental we mean that their value should be such that the environment in our universe is such that physical observers can form in it. This has led to the proposal of several specific models, some of which I have then applied to the early universe and studied their constraint. In chapter 3, I have considered the case of Split Susy, the first of these theories, where it is proposed that the supersymmetric partners of the Standard Model are split into scalar, which are very heavy at an intermediate scale between the weak scale and the GUT scale, and the fermions, which are instead at the weak scale protected by chiral symmetry. The lightness of the higgs is then not explained dynamically, but it is tuned to be light for environmental reasons. In this model it is the fact that the lightest supersymmetric partner (LSP) should be a thermal relic and made up the dark matter of the universe that fixes the fermions at around the weak scale, with deep consequences for a possible detection at LHC. This would in fact not be true if the gravitino or another dark sector particle were the LSP, and I have studied in detail the dark matter relic abundance in this case, and its consequences on the spectrum of the weakly interacting fermionic partners. In the same direction, in chapter 4, I have studied a model which explains the tuning of the higgs through a mechanism in which baryons, which are necessary for in our universe there to be life, are present in our universe only if the higgs is 189 parametrically light. Baryons are produced at the electroweak phase transition in the so called electroweak baryogenesis scenario, and I have studied in detail the predicted baryon abundance. Taking this into account, the Physics associated with the CP violation then becomes very interesting when it is realized that the natural predicted electron electric dipole moment is parametrically just beyond current experimental limits, at a level reachable by next generation experiments. Eventually, in chapter 6, I have illustrated a study of the minimal model beyond the standard model which accounts for the dark matter in the universe and gauge coupling unification, where again the higgs is assumed to be light for environmental reasons. We have realized that the right amount of dark matter abundance can be achieved with particles of unusual large mass, while gauge coupling unification works extremely well, as we have explicitly shown embedding the model in a five dimensional theory. I think that thanks to these works I have been able to gain some good understanding of the field of Cosmology and Particle Physics. It is then very exciting to realize that in the very next few years, with the turning on of new experiments in Cosmology, such as the Planck satellite, and in Particle Physics as LHC, many of the ideas I have worked on during my studies for a PhD will be possibly tested, and our understanding of both the fields potentially revolutionized. A detection in the CMB of a tilt in the power spectrum of the scalar perturbations, or of a polarization induced by a background of gravitational waves, or of a non-Gaussian signal would shed light into the inflationary phase of the universe, and therefore also on the physics at very high energies. Combined with a definite improvement in the data on the Supernovae, expected in the very near future, these new experiments might shed light also on the great mystery of the present value of the Dark Energy, ruling out or giving further support to the idea of the Landscape. Similar in this will be the turning on of LHC which will probably help us understand if the solution to the hierarchy problem is to be solved in some natural way, as for example with TeV scale Supersymmetry, or if there is evidence of a tuning which points towards the presence of a Landscape. 190