Anyscale is a fully-managed Ray platform, on which you can build, deploy, and manage scalable AI and Python applicationsThis example goes over how to use LangChain to interact with Anyscale Endpoint.
Copy
##Installing the langchain packages needed to use the integrationpip install -qU langchain-community
question = "When was George Washington president?"llm_chain.invoke({"question": question})
With Ray, we can distribute the queries without asynchronized implementation. This not only applies to Anyscale LLM model, but to any other LangChain LLM models which do not have _acall or _agenerate implemented
Copy
prompt_list = [ "When was George Washington president?", "Explain to me the difference between nuclear fission and fusion.", "Give me a list of 5 science fiction books I should read next.", "Explain the difference between Spark and Ray.", "Suggest some fun holiday ideas.", "Tell a joke.", "What is 2+2?", "Explain what is machine learning like I am five years old.", "Explain what is artifical intelligence.",]
Copy
import ray@ray.remote(num_cpus=0.1)def send_query(llm, prompt): resp = llm.invoke(prompt) return respfutures = [send_query.remote(llm, prompt) for prompt in prompt_list]results = ray.get(futures)