Diary
16th December, 2024
#####
- LLM context window is the amount of text an LLM can take as input and formulate a response for. Prompt engineering is tweaking the text in this context window.
- The longer the inpute in the context window, the higher the compute cost.
- Neural attention memory models (NAMMs) can be used to eliminate redundant tokens and improve performance.