Magnetic Field Evolution in Neutron Star Crusts Due to the Hall Effect and Ohmic Decay

, , and

© 2004. The American Astronomical Society. All rights reserved. Printed in U.S.A.
, , Citation Andrew Cumming et al 2004 ApJ 609 999 DOI 10.1086/421324

0004-637X/609/2/999

Abstract

We present calculations of magnetic field evolution by the Hall effect and ohmic decay in the crust of neutron stars (NSs). In accreting NSs, ohmic decay is always the dominant effect because of the large resistivity. In isolated NSs with relatively pure crusts, the Hall effect dominates ohmic decay after a time tswitch ≃ 104 yr B, where B12 is the magnetic field strength in units of 1012 G. We compute the evolution of an initial field distribution by ohmic decay and give approximate analytic formulae for both the surface and interior fields as a function of time. Because of the strong dependence of tswitch on B12, early ohmic decay can alter the currents down to the base of the crust for B ~ 1011 G, neutron drip for B ~ 1012 G, and near the top of the crust for B ≳ 1013 G. We then discuss magnetic field evolution by the Hall effect. Several examples are given to illustrate how an initial field configuration evolves. Hall-wave eigenfunctions are computed, including the effect of the large density change across the crust. We estimate the response of the crust to the magnetic stresses induced by Hall waves and give a detailed discussion of the boundary conditions at the solid-liquid interface. Finally, we discuss the implications for the Hall cascade proposed by Goldreich & Reisenegger.

Export citation and abstract BibTeX RIS

Please wait… references are loading.
10.1086/421324