Posted by **chris** on Wednesday, July 11, 2007 at 10:03pm.

Show that a subset W of a vector space V is a subspace of V if and only if the following condition holds: If u and v are any vectors in W and a and b are any scalars, then au + bv is in W.

## Answer This Question

## Related Questions

- Algebra - How would you prove this theorem: The column space of an m x n matrix ...
- Linear Algebra - 1/ Prove that the set V=R+ ( the set of all positive real ...
- Math Elementary Linear Algebra - determine whether or not the given set forms a ...
- math - Find an orthonormal basis for the subspace of R^3 consisting of all ...
- Linear Algebra - V = Mnxn is a space vector formed by all the square matrices. ...
- Linear Algebra - Vector in Space-Directions and Magnitudes Let a, b, c, d, e, & ...
- math - If A^TA is an invertible matrix, prove that the column vectors of A are ...
- math vectors - Use a specific example to prove that the cross product is also ...
- Linear Algebra - Prove that If a vector space is of dimension n and a set of ...
- Linear algebra - Use the properties of linear operations with vectors to show ...

More Related Questions