Inference
Inference in AI refers to the process of using a trained model to make predictions or decisions based on new data. Once an AI system has been trained, inference is what it does when you actually use it. For example, when you type a question into ChatGPT and it gives you an answer, that response is generated during inference.
You can think of training as the learning phase and inference as the doing phase. The model has already learned from a lot of data, and during inference, it applies what it knows to a new situation. In many AI applications, inference needs to be fast and efficient, especially in real-time tools like voice assistants or image recognition systems.