Replies: 7 comments 4 replies
-
In my view it is nice idea. It will help in building scalable pipeline (for example scaling individual component and putting it behind load balancer with caching support) There are multiple ways to achieve this but one quick hack I see to add operation mode I am exploring this space for Obsei, like ray, temporal, airflow etc. But recently come across Apache Beam's Pipeline IO implementation https://beam.apache.org/documentation/io/built-in/. From design point of view I liked it. SO main idea to add communication (rest, binary protocol, streaming etc) as a part of edge and node have to follow what edge is require. But yes it will make make tool quite bloated. I am curious to get more views around this topic :) |
Beta Was this translation helpful? Give feedback.
-
I agree with the URI first implementation. That will be the fastest way to make it happen ! |
Beta Was this translation helpful? Give feedback.
-
Yep, definitely a good idea and direction. We are exploring Ray for the purpose of parallelizing our pipelines and allowing distributed execution (#688). For our deployments, I could see a mix of Ray and Kubernetes to distribute our pipeline nodes to the right machines, but we are still in the phase of sorting out a good design here. open for any ideas and experience sharing regarding the above frameworks. From my perspective airflow is more suitable for batch jobs than real-time search queries. Not sure how Apache Beam fits in there, but I also mostly heard from people using it for bigger batch jobs with Spark & Co. |
Beta Was this translation helpful? Give feedback.
-
Moved it here to discussions as we want in future a clearer separation between these "early ideation & discussion" and actual issues/bugs/features. |
Beta Was this translation helpful? Give feedback.
-
On DataTalks.Club community, I asked question about developing scalable model serving ML Pipeline to @hanneshapke (Author of "Building Machine Learning Pipelines). He suggested interesting approach of using "combination between Apache Beam and Kubernetes". Quoting his response -
|
Beta Was this translation helpful? Give feedback.
-
I saw recently on HuggingFace that there is an API endpoint for hosted Readers. It could be an option for companies who don't need on-premise and can publish to HF (our case lol) Reference: |
Beta Was this translation helpful? Give feedback.
-
@guillim @tholor @oryx1729 I am currently trying to revamp Obsei. Hence working in similar direction. What do you think about it. I feel some portion can be also relevant for Haystack as well. Coping ticket content here, also I would love to hear feedback about my proposed design.
|
Beta Was this translation helpful? Give feedback.
-
Question
Hello, I am wondering if there is a way to outsource the Reader, or any other highly resource consuming component.
Additional context
The point is you can use a 2Go RAM server for 99% of any pipelines. It will fit most use-cases. However, the Reader part will require very specific hardware/RAM and can't scale. So instead of having to scale the Haystack server, we could simply have an outside service for running our Reader only. This external dedicated service could be connected as in the pipelines.yaml
Curious to read your thoughts
Beta Was this translation helpful? Give feedback.
All reactions