We investigate whether in-context examples, widely used in decoder-only language models (LLMs), can improve embedding model performance in Retrieval tasks. Unlike in LLMs, naively prepending in-context examples (query-document pairs) to the target query at inference time does not work