Score:0

Which server specs to look for for a Scheduler in a data stack?

pk flag

I'm in the process of building a data stack for a small company : the choice has been made to have a UNIX server running as a "Scheduler". The goal of this scheduler is just to extract data from different applications and send it back to a in-cloud datawarehouse. Opposite flows are also expected in the near future.

In concrete terms the server will just host Airflow and run Python and Bash scripts. It's likely that the server will also host other Flasks apps for internal use only (data, monitoring etc.). The company in this case has really "small" flows : we're not talking of big data.

Now that I have to choose a host and hardware specs I'm a bit confused, what should I look for and what descriminants should I think of ? Is there any must-have option/capability or any that I should avoid at all costs ?

Thank you !

us flag
The server requirements are completely dependent on the application and how you use it. You can only get relevant information by performing your own benchmarking.
us flag
Does this answer your question? [Can you help me with my capacity planning?](https://serverfault.com/questions/384686/can-you-help-me-with-my-capacity-planning)
mangohost

Post an answer

Most people don’t grasp that asking a lot of questions unlocks learning and improves interpersonal bonding. In Alison’s studies, for example, though people could accurately recall how many questions had been asked in their conversations, they didn’t intuit the link between questions and liking. Across four studies, in which participants were engaged in conversations themselves or read transcripts of others’ conversations, people tended not to realize that question asking would influence—or had influenced—the level of amity between the conversationalists.