OpenRouter Large Language Models
openrouter/router
Run any LLM (Large Language Model) with fal, powered by OpenRouter.
Inference
Commercial use
Streaming
Input
Additional Settings
Customize your input with more control.
Streaming
Result
Idle
What would you like to do next?
You will be charged based on the number of input and output tokens.
Logs
đź§ Using OpenRouter API
The OpenRouter API provides unified access to a wide range of large language models — including GPT, Claude, Gemini, and many others through a single API interface. Billed through fal.
🚀 Usage with OpenAI Client
pythonfrom openai import OpenAI import os client = OpenAI( base_url="https://fal.run/openrouter/router/openai/v1", api_key="not-needed", default_headers={ "Authorization": f"Key {os.environ['FAL_KEY']}", }, ) response = client.chat.completions.create( model="google/gemini-2.5-flash", messages=[ {"role": "user", "content": "Write a short story (under 200 words) about an AI that learns to dream. Use vivid sensory details and end with a surprising twist that makes the reader feel both awe and melancholy."}, ], ) print(response.choices[0].message.content)
đźšż Streaming Example
pythonfrom openai import OpenAI import os client = OpenAI( base_url="https://fal.run/openrouter/router/openai/v1", api_key="not-needed", default_headers={ "Authorization": f"Key {os.environ['FAL_KEY']}", }, ) stream = client.chat.completions.create( model="google/gemini-2.5-flash", messages=[ {"role": "user", "content": "Explain quantum computing in simple terms."}, ], stream=True, ) for chunk in stream: if chunk.choices and chunk.choices[0].delta: print(chunk.choices[0].delta.content, end="", flush=True)
📚 Documentation
For more details, visit the official docs:
- đź”— OpenRouter API Docs
- ⚡ fal.ai API Docs