So, everyone wants to host their own (copy of a) large language model (LLM), but many academic institutions can't spin up multiple LLMs simultaneously, in perpetuity, nor do I believe the Scientific Funding Agencies in each country would want to pay for everyone to get a GPU cluster just to host 500+ copies of tomorrow's version of LLAMA-2(ish).
Are you aware of any effort proposing or planning to host LLMs for use by researchers in some shared infrastructure? After all, hosting the LLM costs the same per hour whether 1, 3 or 20 people are calling it, and at most academic institutions usage would be a little bursty.
Best, Amanda Stent
-- (they/she) Director, Davis Institute for AI Professor, Computer Science Colby College
Follow the Davis Institute for AI here https://web.colby.edu/davisinstitute/ Want to meet? Calendly - Amanda Stent https://calendly.com/amandastentcolby