Domain-adaptive fine-tune as orthogonal R@5 lift on top of MemPal raw #1249
nakata-app
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi MemPal team,
We've been using LongMemEval to evaluate a small open-source library
called
adaptmem, a 200-line hard-negative mining + contrastive fine-tune wrapper around
SentenceTransformers, and the numbers we got line up cleanly with
the work you've already published. Wanted to share back, see if
it's interesting.
What we measured
Same dataset (
longmemeval_s_cleaned.json), same encoder family(MiniLM-L6, ~90MB), run through your own
longmemeval_bench.py(monkey-patched to swap the encoder, zero changes to your eval logic).
Only the fine-tune step differs.
Three findings worth flagging:
Raw baseline R@5 = 0.966 matches your published number exactly.
Independent confirmation that your protocol is fully reproducible,
we didn't need any hints beyond the repo README.
FT-300 + raw mode: +5.6pt R@1, +1.4pt R@5. R@1 is where
contrastive fine-tuning moves the needle most, the model learns to
rank the right session first, not just in top-5.
FT-300 + hybrid_v4: +11pt R@1, +2.4pt R@5. Fine-tune and
hybrid retrieval stack orthogonally, each adds lift on top of the
other.
Possible integration shape
If interesting, a
mempal-adaptintegration could look like:"adapter": before ingestion, point adaptmem at the labelled-query
set (if available), it produces a domain-tuned encoder that mempal
then uses for embedding.
config load time.
We don't have strong feelings about the shape, happy to defer to
your design preferences. The point of this thread is just to put
the numbers in front of you and see whether there's a productive
conversation here.
Reproduce
Three committed result JSONs in
benchmarks/:results_minilm_baseline_400.json, raw protocol confirmation.results_ft100_400.json, self-contained FT-100 reproduce.results_ft300_direct.json, FT-300 reference run.Either outcome is fine
If this isn't a fit for mempal's direction, no problem, adaptmem
will keep on as a standalone tool. Just thought it was worth showing
the numbers and the integration sketch given how cleanly the
protocol confirmation came out.
Thanks again for the open work, the project structure made
independent reproduction straightforward.
Nakata
Beta Was this translation helpful? Give feedback.
All reactions