To add on @Tendero, the expression ∑kxpk is sometimes called the "power p-norm" when p≥1. Most often, you can see mentions of the "squared ℓ2 norm" or "ℓ2 norm squared". For p=1, the exponent does not modify the computation, so it is just called the ℓ1 norm.
The use of the power is often more convenient mathematically and computationally: having a p-root (⋅√p) can be cumbersome in computing derivatives to find extrema.
When p≥1, it satisfies all the norm axioms. But when 0<p<1, the triangle inequality is not satisfied anymore, so it should not be called a norm. The correct denomination is a quasi-norm, with a modulus of concavity modulus K such that
ℓp(x+y)≤K(ℓp(x)+ℓp(y)).
When p=0, this is not a norm nor a quasi-norm anymore. It can be called cardinality function, sparsity, count index.
In signal processing where sparsity is considered useful, ℓ0 is usually the target to minimize: number of non-zero samples, number of taps for a filter.
However, it is quite intractable too (not differentiable). Under some theoretical conditions, the minimization of an ℓ0 penalty can be replaced by an ℓ1 penalty, the "last" convex ℓpp term.
However, more and more works address non-convex penalties (p<1) that better approximate ℓ0.
Finally, in a Bayesian context, the prior of a Laplacian distribution can be encapsulated in an ℓ1 penalty, as the Gaussian distribution can be encapsulated in an ℓ2 penalty, see for instance Why is Laplace prior producing sparse solutions?.