Jar3d is a versatile agent for research intensive tasks that combines chain-of-reasoning, Meta-Prompting, and Agentic RAG techniques.
- It features integrations with popular providers and open-source models, allowing for 100% local operation given sufficient hardware resources.
Jar3d will introduce itself and help you refine your requirements. Once you have delivered all the required type /end
. Jar3d will pass your requirements on to the MetaExpert who will work with its team of experts to deliver on your requirements.
- Long-running research tasks, writing literature reviews, newsletters, etc.
- Potential adaptation for use with internal company documents, requiring no internet access.
- Can function as a research assistant or a local version of services like Perplexity.
- Market research
- Sourcing goods (i.e. Find me the chepaest A5 Wagyu beef from French suppliers)
For setup instructions, please refer to the Setup Jar3d guide.
Jar3d utilizes two powerful prompts written entirely in Markdown:
Both prompts incorporate adaptations of the Chain of Reasoning technique.
The Jar3d architecture incorporates aspects of Meta-Prompting, Agentic RAG, and an adaptation of Chain of Reasoning.
This system employs a sophisticated retrieval mechanism for conducting internet research. The process involves several steps, utilizing various tools and techniques to ensure comprehensive and relevant results.
- Utilizes the SERPER tool to find relevant web pages.
- Employs an LLM-executed search algorithm, expressed in natural language.
- Each iteration of the algorithm generates a search query for SERPER.
- SERPER returns a search engine results page (SERP).
- Another LLM call selects the most appropriate URL from the SERP.
- This process is repeated a predetermined number of times to compile a list of URLs for in-depth research.
- Employs LLM Sherpa as a document ingestor.
- Intelligently chunks the content from each URL in the compiled list.
- Results in a corpus of chunked text across all accumulated URLs.
- Embeds the chunked text using a locally hosted model from FastEmbed.
- Indexes embeddings in an in-memory FAISS vector store.
- Performs retrieval using a similarity search over the FAISS vector store.
- Utilizes cosine similarity between indexed embeddings and the meta-prompt (written by the meta-agent).
- Retrieves the most relevant information based on this similarity measure.
- Leverages FlashRank as a locally hosted re-ranking service.
- FlashRank uses cross-encoders for more accurate assessment of document relevance to the query.
- Selects a designated percentile of the highest-scoring documents from the re-ranked results.
- Passes this final set of retrieved documents to the meta-agent for further processing or analysis.