arXivFlow is a project that enables you to automatically fetch, process, and ingest the latest ArXiv research papers on any given topic on a daily basis. This daily retrieval supports continuous technological monitoring, ensuring that you stay up-to-date with emerging research and trends. The pipeline is orchestrated using Prefect for scheduling and seamless automation, and it stores the retrieved PDFs in a MinIO object storage system for efficient management and retrieval.
- Fetch ArXiv Papers: Automatically query the ArXiv API for research papers based on a topic and publication date.
- PDF Ingestion: Download the PDF files and store them in a MinIO bucket.
- Pipeline Orchestration: Use Prefect flows and tasks to schedule and manage the pipeline.
- Clone the repository
git clone https://github.com/Bessouat40/arXivFlow.git
cd arXivFlow
- Install the required packages
python3 -m pip install -r requirements.txt
You can run the pipeline as a scheduled flow using Prefect. For example, to run the pipeline daily at midnight, use the Prefect deployment approach or serve the flow directly (for testing purposes).
python3 -m main
You can now run Prefect flow inside a Docker container :
docker-compose up -d --build
Now you can access Prefect UI at localhost:4200.
Your flow will run every day at midnight.
The pipeline fetches articles based on a given topic and a target date (e.g., yesterday).
You can modify these parameters in your flow (in src/prefect/pipeline.py
).
The MinIOClient is configured with default credentials (minioadmin/minioadmin
) and an endpoint (localhost:9000
). The bucket name used is "llm-pdf
". Make sure your MinIO instance is running and accessible.
-
Python 3.11 (or compatible version)
-
MinIO: Make sure you have a running MinIO server. You can start one using Docker:
docker run -d --name minio_server \
-p 9000:9000 \
-p 9001:9001 \
-e MINIO_ROOT_USER=minioadmin \
-e MINIO_ROOT_PASSWORD=minioadmin \
minio/minio server /data --console-address ":9001"
-
Containerization with Docker: Create a Dockerfile to containerize the application and manage its dependencies.
-
Embedding Extraction: Use a model to extract and store embeddings from the PDFs for later semantic search.
-
Semantic Search: Implement a semantic search feature that leverages the stored embeddings to enable more accurate article search.