Talk to us

LlamaIndex Jan 23, 2024

LlamaIndex Newsletter 2024–01–23

Hello LlamaIndex Explorers 🦙,

Another exciting week at LlamaIndex, filled with vibrant community contributions and educational resources. Explore our array of new features, tutorials, guides, and demos, all tailored to enrich your experience with LlamaIndex.

Before delving into the updates, we have two significant announcements:

  • We’re thrilled to host our first in-person hackathon, set for February 2nd-4th. This is a fantastic opportunity to meet fellow RAG enthusiasts, collaborate, and compete for prizes totaling over $8000!
  • Don’t miss our webinar featuring Sehoon Kim and Amir Gholami, scheduled for Thursday at 9 am PT. They will introduce LLMCompiler, an agent compiler for parallel multi-function planning and execution.

We’re always excited to see your projects, articles, or videos. If you’ve created something you’re proud of, share it with us at news@llamaindex.ai. Also, remember to subscribe to our newsletter on our website to get all the latest news straight to your inbox.

🤩 The highlights:

  1. RankGPT: Introducing RankGPT leveraging GPT-3.5 and GPT-4 for top-tier document ranking and a novel sliding window technique for extensive context management. Notebook, Tweet.
  2. Composable Retrievers: An interface centralizing advanced retrieval and RAG techniques, enhancing RAG setups with IndexNodes for linking diverse retrievers and pipelines. Docs, Tweet.
  3. Advanced QA over Tabular Data Tutorial: A detailed guide to crafting query pipelines over tabular data, featuring Pandas, SQL, and Query Pipelines for an integrated few-shot, LLM, and custom function setup. Text-to-SQL, Text-to-Pandas.
  4. Long-Context Embedding Models: Explore models like M2-BERT-80M-32k-retrieval tackling the embedding chunking problem in RAG, with a focus on hybrid retrieval methods and hierarchical retrieval approaches. Guide.

✨ Feature Releases and Enhancements:

  • We have introduced RankGPT in our advanced module that utilizes GPT-3.5 and GPT-4 for efficient document ranking, featuring a unique sliding window strategy for handling large contexts. Notebook, Tweet.
  • We have launched Composable Retrievers which centralizes various advanced retrieval and RAG techniques into a versatile interface. It simplifies creating complex RAG setups by allowing you to define IndexNodes to link different retrievers or RAG pipelines. Docs, Tweet.
  • Anoop Sharma has introduced LlamaPack for Multi-Stock Ticker Analysis for analyzing various stock tickers with a single code line, enabling easy specification of tickers, time frames, and structured queries. LlamaPack, Tweet.
  • LlamaIndex.TS (LITS) supports streaming on all endpoints. Tweet.
  • We announced a new integration with Tonic Validate to allow simple access to LLM-powered evaluations. Blog post

🎥 Demo:

  • RAG-Maestro for ArXiv Research: Developed by Aymen Kallala, this web app utilizes RAG to efficiently search scientific concepts in ArXiv papers. It extracts keywords using RAKE, queries ArXiv for relevant papers, and offers on-the-fly indexing with in-line citations — a valuable tool for ML researchers navigating through ArXiv’s extensive library. Demo, GitHub Repo.

🗺️ Guides:

  • Guide to Advanced QA over Tabular Data which provides a comprehensive tutorial on creating sophisticated query pipelines over tabular data using Pandas or SQL, constructing a query DAG using our Query Pipelines, integrating few-shot examples, linked prompts, LLMs, custom functions, retrievers, and more. Text-to-SQL, Text-to-Pandas.
  • Guide to a Five-Part Series on Building a Full-Stack RAG Chatbot by Marco Bertelli, extensive tutorials covering every aspect of creating an RAG chatbot — from model selection and Flask backend setup to constructing the ChatEngine and optimizing the RAG pipeline.
  • Guide to Long-Context Embedding Models: The models, like M2-BERT-80M-32k-retrieval, offer a solution to the embedding chunking issue in RAG by grounding retrieval in broader semantic contexts. Learn about hybrid retrieval, combining chunk and document-level similarity, and other approaches like hierarchical retrieval.

✍️ Tutorials:

  • Wenqi tutorial on Democratizing LLMs: 4-bit Quantization for Optimal LLM Inference with LlamaIndex.
  • Andrej tutorial on FastAPI and LlamaIndex RAG: Creating Efficient APIs.
  • Lulia Brezeanu tutorial on Advanced Query Transformations to Improve RAG.
  • Akash Mathur in-depth tutorial on Advanced RAG: Query Augmentation for Next-Level Search using LlamaIndex.
  • Ryan Nguyen tutorial on Live Indexing for RAG: A Guide For Real-Time Indexing Using LlamaIndex and AWS.
  • Nipuna (Paragon AI) tutorial on Building a Full-Stack Complex PDF AI chatbot with LlamaIndex.

🏢 Calling all enterprises:

Are you building with LlamaIndex? We are working hard to make LlamaIndex, even more, Enterprise-ready and have sneak peeks at our upcoming products available for partners. Interested? Get in touch.