# Norms Induce Metrics

02 Nov 2014**Note:** We’ll only consider norms and metrics over Vector Spaces for this post.

In a previous post (“How different are metrics and norms?”) I claimed that norms induce metrics. I had heard this in passing in classes and read it in textbooks, so I took it to be true. However, I’ve been thinking about how the proof works out and decided to document it here.

## Definitions

We’ll use the same definitions as before:

A **norm** is a function that defines a length/size to each vector in a vector space. A norm, $p$, on a vector space, $U$ over a field $F$, is a function $p : U \to \R$ which satisfies the following for all $a \in F$, $u,v \in U$:

- $p(a v) = \vert a \vert p(v)$ (scalability)
- If $p(v) = 0$ then $v$ is the zero vector.
- $p( u + v ) \le p(u) + p(v)$ (triangle inequality)

A **metric** is a function that defines a distance between elements of a set. Formally a metric $d$, on a vector space, $V$, is a function $d : V \times V \to \R$ which satisfies the following for all $x,y,z \in V$:

- $d(x,y) \ge 0$ (non-negative)
- $d(x,y) = 0$ if and only if $x = y$
- $d(x,y) = d(y,x)$ (symmetry)
- $d(x,z) \le d(x,y) + d(y,z)$ (triangle inequality)

## Proof

Assume $V$ is a vector space over a field $F$, $p : V \to \R$ is a norm, and $u,v,w \in V$.

Define a function $d : V \times V \to \R$ by $d(u, v) = p( u - v )$. (I claim $d$ will then be a metric)

### Property 1.

By definition, $d(u,v) = p(u - v)$, and $p(t) \ge 0$ for all $t \in V$. So, $d(u,v) \ge 0$.

### Property 2.

*(forward)*
Assume $d(u,v) = 0$. Then $p(u-v) = 0$, by property 3 of norms this implies $u-v = 0$. Thus, $u = v$.

*(reverse)*
Assume $u = v$. Then $d(u,v) = p(u-v) = p(u - u) = p(0)$. Again by property 3 of norms $p(0) = 0$. Thus, $d(u,v) = 0$.

### Property 3.

We have,

### Property 4

Note that $p(-v) = p(-1 \cdot v) = \mid -1 \mid p(v) = p(v)$. We have,

So, $d(t,u) + d(u,v) \ge d(t,v)$.

Then, $d$ satisfies all four properties of being a metric. Therefore, it is a metric.

## First Attempts

Initially, I tried defining the metric as $d(u,v) = p(u) - p(v)$. I was thinking of the necessary symmetry, and since $p(u),p(v) \in \R$ they commute. This leads to *a few* problems in proving the properties (except non-negativity), including symmetry!

I took a pause from the proof and contemplated if this was a bad approach. I defaulted to coming up with an example in $\R^n$. Immediately I thought of the idea “the distance between two vectors should be the length of the difference (vector)”!

Looking back this definition makes a great deal of sense, and I’m not sure why I thought my original definition was a good idea…