Edge-LLM Inference With Cost-Aware Layer Allocation and Adaptive Scheduling | IEEE Journals & Magazine | IEEE Xplore