Why would a model trained with data up to 2023 overlook a significant update on a widely-used platform like GitHub? Could it reflect selective data filtering during training, limitations on incorporating certain types of updates post-cutoff, or perhaps something specific to how different LLMs process and prioritize technical information?
What are your thoughts?
If the bulk of the corpus contains the old, outdated information, you're more likely to get that back from your prompt just because it's had more weightings applied to it than the newer info. Sometimes you can add extra emphasis to your prompt to pull out a prediction closer to what you want. For example, if an answer refers to the old version you can say, "That was correct for the previous version but I'm asking about the version of GitHub that supports LaTeX math rendering in Markdown after May 2022." Adding those extra words to your prompt might surface what you want -- if it's in there to begin with.