LaMoE: Integrating Sparse Mixture of Experts into LLaMA