This paper shows that you can vastly improve a modelβs command-line (terminal) skills by carefully engineering the training data, not just by using a bigger model.
This paper shows, step by step, how to train a 1.36-billion-parameter science-focused language model directly from raw arXiv LaTeX files using only 2 A100 GPUs.
This paper teaches a computer to find buttons, text, and icons on screens so it can click and type in the right places, a skill called GUI grounding.