How can one use 3k context size with Llama2? Trying out such a size makes the python process hog one core indefinitely.