In the Shadow of Prompts: Adversarial Attacks and Model Cloning in Large Language Models