-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feasibility study: ditch farm and use pytorch #178
Comments
Noting that FARM project has transitioned to HAYSTACK: https://github.com/deepset-ai/haystack/ and that some OS-Climate members are also HAYSTACK sponsors/supporters. |
There are two alternatives that seems plausible here:
|
Let's not forget that we have a technical solution (disable multiprocessing) and a view to a technical solution (allocating appropriate amounts of SHM) to the train kpi-extraction problem. If there are OTHER problems with farm we want to solve, let's talk about those, but we should be able to dispose of the kpi-extaction problem very simply. |
The disable multiprocessing solution makes the training almost impractical, in one experiment it took around 1.5 days to train 80% only to run into kernel error. This was for training with 145 files. When we get more annotated files, it would be even more difficult. Allocating appropriate amounts of SHM seems like longer term solution to get working. One other obvious problem with FARM is, like @MichaelTiemannOSC pointed out, that it is no longer actively maintained. It has transitioned to farm-haystack and any further development will happen in that pypi package. In my above comment, I list two alternatives, one is using farm-haystack that is actively maintained or using transformers package which is the root package used by farm-haystack to build upon. |
We should absolutely get off of the unmaintained package. We could use farm-haystack but it's an added layer of dependency and I'm trying to sort out the cost/benefits over just using transformers package. |
Estimate the benefit, risk, and labor cost involved in removing the farm architecture and using simple PyTorch. We want to see if we do that, then does our workflow that involves quay image and Kubeflow pipelines become easy to understand, simplified, and less error-prone.
The text was updated successfully, but these errors were encountered: