Robert Oschler
unread,Dec 2, 2024, 11:09:15 PM12/2/24Sign in to reply to author
Sign in to forward
You do not have permission to delete messages in this group
Either email addresses are anonymous for this group or you need the view member email addresses permission to view the original message
to Chrome Built-in AI Early Preview Program Discussions, Thomas Steiner, Chrome Built-in AI Early Preview Program Discussions, Robert Oschler
Funnily enough, things started working pretty good when I trimmed the prompt down to just this:
"
Below is a list of web pages that I have bookmarked related to:
${userQuery}
Please tell me what my bookmarks say about this.
Here are the bookmarks:
${documentText}"
If were going to speculate (wildly), it seems like Gemini Nano is an extraordinary beast. They managed to find a sweet spot of maximum power, as long as you use language that appears to map to a known, common cognition pattern (like a user asking about their bookmarks or about a document). So rather than trying to micro-manage the LLM's "thinking" through complex highly specific prompt instructions, simpler is better. I saw in your video how you used "logic boundaries" to corral the LLM, like when you told it specifically not to put text outside the synonym list and that an answer should appear only once, but that was a structured data context. So perhaps in more pure semantic use contexts like mine, "less is better" with Gemini Nano.
You can see how well Gemini Nano performed in the RAG search my Chrome extension performs. The search uses a subset of my Canary browser bookmarks as "grounding attributions" that are culled out during a cosine similarity search with the help of a locally loaded Jira embeddings model, managed by transformers.js. See the screenshot. The top text area is the LLM answer from the Prompt API, and the bottom text area contains the matching bookmarks with their summaries that were created by the Summarization API when the bookmark was added.
