# Smart Retrieval Frontend The frontend is built using [Next.js](https://nextjs.org/) & [Vercel AI](https://github.com/vercel/ai) bootstrapped with [`create-llama`](https://github.com/run-llama/LlamaIndexTS/tree/main/packages/create-llama). ## Getting Started First, install the dependencies: ```bash npm install ``` Second, run the development server: ```bash npm run dev ``` Open [http://localhost:3000](http://localhost:3000) with your browser to see the result. You can start editing the page by modifying `app/page.tsx`. The page auto-updates as you edit the file. This project uses [`next/font`](https://nextjs.org/docs/basic-features/font-optimization) to automatically optimize and load Inter, a custom Google Font. ## Learn More To learn more about LlamaIndex, take a look at the following resources: - [LlamaIndex Documentation](https://docs.llamaindex.ai) - learn about LlamaIndex (Python features). - [LlamaIndexTS Documentation](https://ts.llamaindex.ai) - learn about LlamaIndex (Typescript features).