-
Notifications
You must be signed in to change notification settings - Fork 1
Open
Description
from sul-dlss/speech-to-text#3
see if back of the envelope estimates could rule out some of our possible implementations, or if they all look close enough in cost to keep pursuing all paths.
this involves two orthogonal sets of choices: how to deploy, and how to measure cost of deployed services.
deployment choices (not all mutually exclusive, but some are):
- ECS, possibly with some rules around scaling, unclear whether we can scale from 0 to 1 or more GPU EC2 instances in the ECS cluster. may need to have at least one EC2 instance always up, but it might be ok for it to be non-GPU? is it possible to have heterogeneous mix of VM types in one ECS cluster?
- always on EC2 VM
- thin vs heavy whisper docker container (i.e. is the container built such that it contains the models it needs, or does it download them on startup? whisper models total out to about 13 GB, says @edsu)
- SageMaker vs docker container running whisper (concerns about SageMaker configurability and training data policy captured in [investigate/prototype] speech_to_text_generation_service approach 2: Explore AWS SageMaker speech-to-text#4)
cost estimate approach:
- cost calculator
- deploy and test w/ tagged resources (@edsu seemed to be leaning in this direction, using Add initial Docker container speech-to-text#9 as a starting point for investigation)
Metadata
Metadata
Assignees
Labels
No labels