Abstract.
In this paper we prove, under various conditions, the so-called Lojasiewicz inequality $ \| E' (u + \varphi) \| \geq \gamma|E(u+\varphi) - E(\varphi)|^{1-\theta} $, where $ \theta \in (0,1/2] $, and γ > 0, while $ \| u \| $ is sufciently small and φ is a critical point of the energy functional E supposed to be only C², instead of analytic in the classical settings. Here E can be for instance the energy associated to the semilinear heat equation $u_t = \Delta u - f(x,u) $ on a bounded domain $ \Omega \subset \mathbb{R}^N $. As a corollary of this inequality we give the rate of convergence of the solution u(t) to an equilibrium, and we exhibit examples showing that the given rate of convergence (which depends on the exponent θ and on the critical point through the nature of the kernel of the linear operator $ E'' (\varphi)) $ is optimal.
Similar content being viewed by others
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Haraux, A., Ali Jendoubi, M. & Kavian, O. Rate of decay to equilibrium in some semilinear parabolic equations. J.evol.equ. 3, 463–484 (2003). https://doi.org/10.1007/s00028-003-1112-8
Issue Date:
DOI: https://doi.org/10.1007/s00028-003-1112-8