/ mlflow / utils / model_catalog / anyscale.json
anyscale.json
  1  {
  2    "schema_version": "1.0",
  3    "models": {
  4      "HuggingFaceH4/zephyr-7b-beta": {
  5        "mode": "chat",
  6        "context_window": {
  7          "max_input": 16384,
  8          "max_output": 16384,
  9          "max_tokens": 16384
 10        },
 11        "pricing": {
 12          "input_per_million_tokens": 0.15,
 13          "output_per_million_tokens": 0.15
 14        },
 15        "capabilities": {
 16          "function_calling": false,
 17          "vision": false,
 18          "reasoning": false,
 19          "prompt_caching": false,
 20          "response_schema": false
 21        },
 22        "last_updated_at": "2026-04-24"
 23      },
 24      "codellama/CodeLlama-34b-Instruct-hf": {
 25        "mode": "chat",
 26        "context_window": {
 27          "max_input": 4096,
 28          "max_output": 4096,
 29          "max_tokens": 4096
 30        },
 31        "pricing": {
 32          "input_per_million_tokens": 1.0,
 33          "output_per_million_tokens": 1.0
 34        },
 35        "capabilities": {
 36          "function_calling": false,
 37          "vision": false,
 38          "reasoning": false,
 39          "prompt_caching": false,
 40          "response_schema": false
 41        },
 42        "last_updated_at": "2026-04-24"
 43      },
 44      "codellama/CodeLlama-70b-Instruct-hf": {
 45        "mode": "chat",
 46        "context_window": {
 47          "max_input": 4096,
 48          "max_output": 4096,
 49          "max_tokens": 4096
 50        },
 51        "pricing": {
 52          "input_per_million_tokens": 1.0,
 53          "output_per_million_tokens": 1.0
 54        },
 55        "capabilities": {
 56          "function_calling": false,
 57          "vision": false,
 58          "reasoning": false,
 59          "prompt_caching": false,
 60          "response_schema": false
 61        },
 62        "last_updated_at": "2026-04-24"
 63      },
 64      "google/gemma-7b-it": {
 65        "mode": "chat",
 66        "context_window": {
 67          "max_input": 8192,
 68          "max_output": 8192,
 69          "max_tokens": 8192
 70        },
 71        "pricing": {
 72          "input_per_million_tokens": 0.15,
 73          "output_per_million_tokens": 0.15
 74        },
 75        "capabilities": {
 76          "function_calling": false,
 77          "vision": false,
 78          "reasoning": false,
 79          "prompt_caching": false,
 80          "response_schema": false
 81        },
 82        "last_updated_at": "2026-04-24"
 83      },
 84      "meta-llama/Llama-2-13b-chat-hf": {
 85        "mode": "chat",
 86        "context_window": {
 87          "max_input": 4096,
 88          "max_output": 4096,
 89          "max_tokens": 4096
 90        },
 91        "pricing": {
 92          "input_per_million_tokens": 0.25,
 93          "output_per_million_tokens": 0.25
 94        },
 95        "capabilities": {
 96          "function_calling": false,
 97          "vision": false,
 98          "reasoning": false,
 99          "prompt_caching": false,
100          "response_schema": false
101        },
102        "last_updated_at": "2026-04-24"
103      },
104      "meta-llama/Llama-2-70b-chat-hf": {
105        "mode": "chat",
106        "context_window": {
107          "max_input": 4096,
108          "max_output": 4096,
109          "max_tokens": 4096
110        },
111        "pricing": {
112          "input_per_million_tokens": 1.0,
113          "output_per_million_tokens": 1.0
114        },
115        "capabilities": {
116          "function_calling": false,
117          "vision": false,
118          "reasoning": false,
119          "prompt_caching": false,
120          "response_schema": false
121        },
122        "last_updated_at": "2026-04-24"
123      },
124      "meta-llama/Llama-2-7b-chat-hf": {
125        "mode": "chat",
126        "context_window": {
127          "max_input": 4096,
128          "max_output": 4096,
129          "max_tokens": 4096
130        },
131        "pricing": {
132          "input_per_million_tokens": 0.15,
133          "output_per_million_tokens": 0.15
134        },
135        "capabilities": {
136          "function_calling": false,
137          "vision": false,
138          "reasoning": false,
139          "prompt_caching": false,
140          "response_schema": false
141        },
142        "last_updated_at": "2026-04-24"
143      },
144      "meta-llama/Meta-Llama-3-70B-Instruct": {
145        "mode": "chat",
146        "context_window": {
147          "max_input": 8192,
148          "max_output": 8192,
149          "max_tokens": 8192
150        },
151        "pricing": {
152          "input_per_million_tokens": 1.0,
153          "output_per_million_tokens": 1.0
154        },
155        "capabilities": {
156          "function_calling": false,
157          "vision": false,
158          "reasoning": false,
159          "prompt_caching": false,
160          "response_schema": false
161        },
162        "last_updated_at": "2026-04-24"
163      },
164      "meta-llama/Meta-Llama-3-8B-Instruct": {
165        "mode": "chat",
166        "context_window": {
167          "max_input": 8192,
168          "max_output": 8192,
169          "max_tokens": 8192
170        },
171        "pricing": {
172          "input_per_million_tokens": 0.15,
173          "output_per_million_tokens": 0.15
174        },
175        "capabilities": {
176          "function_calling": false,
177          "vision": false,
178          "reasoning": false,
179          "prompt_caching": false,
180          "response_schema": false
181        },
182        "last_updated_at": "2026-04-24"
183      },
184      "mistralai/Mistral-7B-Instruct-v0.1": {
185        "mode": "chat",
186        "context_window": {
187          "max_input": 16384,
188          "max_output": 16384,
189          "max_tokens": 16384
190        },
191        "pricing": {
192          "input_per_million_tokens": 0.15,
193          "output_per_million_tokens": 0.15
194        },
195        "capabilities": {
196          "function_calling": true,
197          "vision": false,
198          "reasoning": false,
199          "prompt_caching": false,
200          "response_schema": false
201        },
202        "last_updated_at": "2026-04-24"
203      },
204      "mistralai/Mixtral-8x22B-Instruct-v0.1": {
205        "mode": "chat",
206        "context_window": {
207          "max_input": 65536,
208          "max_output": 65536,
209          "max_tokens": 65536
210        },
211        "pricing": {
212          "input_per_million_tokens": 0.9,
213          "output_per_million_tokens": 0.9
214        },
215        "capabilities": {
216          "function_calling": true,
217          "vision": false,
218          "reasoning": false,
219          "prompt_caching": false,
220          "response_schema": false
221        },
222        "last_updated_at": "2026-04-24"
223      },
224      "mistralai/Mixtral-8x7B-Instruct-v0.1": {
225        "mode": "chat",
226        "context_window": {
227          "max_input": 16384,
228          "max_output": 16384,
229          "max_tokens": 16384
230        },
231        "pricing": {
232          "input_per_million_tokens": 0.15,
233          "output_per_million_tokens": 0.15
234        },
235        "capabilities": {
236          "function_calling": true,
237          "vision": false,
238          "reasoning": false,
239          "prompt_caching": false,
240          "response_schema": false
241        },
242        "last_updated_at": "2026-04-24"
243      }
244    }
245  }