Edge

Model Parallelism for Inference at edge

Explored distributing model components across different devices to enable running large models on edge with multiple smaller devices