Daniel and I have talked over email, and to sum up the conversation, the problem is that he's fitting a linear trend, and he doesn't have very many data points.
You can see in the two attached graphics that when the searching neighborhood moves away from the large value, the extrapolated predictions actually go up. It's counter-intuitive that removing a large value causes predictions to increase, but you can get artifacts like this when you don't have many points.