app.py 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454
  1. from flask import Flask, request, jsonify, render_template_string
  2. from datetime import datetime
  3. import json
  4. import os
  5. import requests
  6. import traceback
  7. app = Flask(__name__)
  8. # Store recent requests in memory
  9. recent_requests = []
  10. MAX_REQUESTS = 50
  11. # Configuration with better defaults
  12. OLLAMA_ENDPOINT = os.getenv("OLLAMA_ENDPOINT", "http://localhost:11434")
  13. OLLAMA_MODEL = os.getenv("OLLAMA_MODEL", "llama2") # More common default
  14. VIDEO_FORMAT = os.getenv("VIDEO_FORMAT", "skip") # Skip videos by default for debugging
  15. def test_ollama_connection():
  16. """Test connection to Ollama and get available models"""
  17. try:
  18. # Try to get model list
  19. response = requests.get(f"{OLLAMA_ENDPOINT}/api/tags", timeout=5)
  20. if response.status_code == 200:
  21. models = response.json().get("models", [])
  22. model_names = [m.get("name", "") for m in models]
  23. print(f"✅ Connected to Ollama at {OLLAMA_ENDPOINT}")
  24. print(f"Available models: {model_names}")
  25. return True, model_names
  26. else:
  27. print(f"❌ Ollama responded with {response.status_code}: {response.text}")
  28. return False, []
  29. except Exception as e:
  30. print(f"❌ Cannot connect to Ollama at {OLLAMA_ENDPOINT}: {e}")
  31. return False, []
  32. def convert_gemini_to_ollama_generate(gemini_request):
  33. """Convert Gemini format to Ollama /api/generate format"""
  34. try:
  35. contents = gemini_request.get("contents", [])
  36. # Extract text parts and combine
  37. prompt_parts = []
  38. images = []
  39. for content in contents:
  40. parts = content.get("parts", [])
  41. for part in parts:
  42. if "text" in part:
  43. prompt_parts.append(part["text"])
  44. elif "inline_data" in part:
  45. inline = part["inline_data"]
  46. mime_type = inline.get("mime_type", "")
  47. data = inline.get("data", "")
  48. if VIDEO_FORMAT == "skip" and (
  49. mime_type.startswith("video/") or mime_type.startswith("image/")
  50. ):
  51. prompt_parts.append(
  52. f"[Media content ({mime_type}) was present but skipped for debugging]"
  53. )
  54. print(f"⏭️ Skipping media: {mime_type}")
  55. elif mime_type.startswith("image/"):
  56. images.append(data)
  57. print(f"🖼️ Adding image: {mime_type}")
  58. # Build Ollama request
  59. ollama_request = {
  60. "model": OLLAMA_MODEL,
  61. "prompt": " ".join(prompt_parts) if prompt_parts else "Hello",
  62. "stream": False,
  63. "options": {},
  64. }
  65. # Add images if present and not skipping
  66. if images and VIDEO_FORMAT != "skip":
  67. ollama_request["images"] = images
  68. # Add generation config
  69. gen_config = gemini_request.get("generationConfig", {})
  70. if "temperature" in gen_config:
  71. ollama_request["options"]["temperature"] = gen_config["temperature"]
  72. if "maxOutputTokens" in gen_config:
  73. ollama_request["options"]["num_predict"] = gen_config["maxOutputTokens"]
  74. return ollama_request
  75. except Exception as e:
  76. print(f"❌ Error converting to Ollama format: {e}")
  77. raise
  78. def convert_ollama_generate_to_gemini(ollama_response):
  79. """Convert Ollama /api/generate response to Gemini format"""
  80. try:
  81. print(
  82. f"🔄 Converting Ollama response: {json.dumps(ollama_response, indent=2)[:500]}..."
  83. )
  84. # Ollama /api/generate returns: {"response": "text", "done": true, ...}
  85. response_text = ollama_response.get("response", "")
  86. if not response_text:
  87. print(f"❌ No 'response' field in Ollama response")
  88. print(f"Full response keys: {list(ollama_response.keys())}")
  89. return {"error": "No response text from Ollama"}
  90. print(f"✅ Found response text: {response_text[:100]}...")
  91. # Convert to Gemini format
  92. gemini_response = {
  93. "candidates": [
  94. {
  95. "content": {"parts": [{"text": response_text}], "role": "model"},
  96. "finishReason": "STOP",
  97. "index": 0,
  98. }
  99. ],
  100. "usageMetadata": {
  101. "promptTokenCount": ollama_response.get("prompt_eval_count", 0),
  102. "candidatesTokenCount": ollama_response.get("eval_count", 0),
  103. "totalTokenCount": ollama_response.get("prompt_eval_count", 0)
  104. + ollama_response.get("eval_count", 0),
  105. },
  106. }
  107. return gemini_response
  108. except Exception as e:
  109. print(f"❌ Error converting Ollama response: {e}")
  110. print(f"Ollama response was: {ollama_response}")
  111. traceback.print_exc()
  112. return {"error": f"Conversion error: {str(e)}"}
  113. HTML_TEMPLATE = """
  114. <!DOCTYPE html>
  115. <html>
  116. <head>
  117. <title>Ollama Proxy Debug</title>
  118. <style>
  119. body { font-family: Arial, sans-serif; margin: 20px; background: #f5f5f5; }
  120. h1 { color: #333; }
  121. .config { background: #e3f2fd; padding: 15px; margin: 10px 0; border-radius: 5px; }
  122. .status { padding: 10px; margin: 10px 0; border-radius: 5px; }
  123. .status.connected { background: #c8e6c9; }
  124. .status.error { background: #ffcdd2; }
  125. .request {
  126. background: white;
  127. padding: 15px;
  128. margin: 10px 0;
  129. border-radius: 5px;
  130. box-shadow: 0 2px 4px rgba(0,0,0,0.1);
  131. }
  132. .timestamp { color: #666; font-size: 0.9em; }
  133. .method {
  134. display: inline-block;
  135. padding: 3px 8px;
  136. background: #4CAF50;
  137. color: white;
  138. border-radius: 3px;
  139. font-weight: bold;
  140. }
  141. .forwarded {
  142. display: inline-block;
  143. padding: 3px 8px;
  144. background: #2196F3;
  145. color: white;
  146. border-radius: 3px;
  147. font-size: 0.8em;
  148. margin-left: 10px;
  149. }
  150. .error-badge {
  151. display: inline-block;
  152. padding: 3px 8px;
  153. background: #f44336;
  154. color: white;
  155. border-radius: 3px;
  156. font-size: 0.8em;
  157. margin-left: 10px;
  158. }
  159. pre {
  160. background: #f4f4f4;
  161. padding: 10px;
  162. border-radius: 3px;
  163. overflow-x: auto;
  164. max-height: 300px;
  165. overflow-y: auto;
  166. }
  167. .clear-btn {
  168. background: #f44336;
  169. color: white;
  170. border: none;
  171. padding: 10px 20px;
  172. border-radius: 5px;
  173. cursor: pointer;
  174. margin: 10px 0;
  175. }
  176. .clear-btn:hover { background: #d32f2f; }
  177. </style>
  178. <script>
  179. function clearRequests() {
  180. fetch('/clear', { method: 'POST' })
  181. .then(() => location.reload());
  182. }
  183. // Auto-refresh every 5 seconds
  184. setTimeout(() => location.reload(), 5000);
  185. </script>
  186. </head>
  187. <body>
  188. <h1>🦙 Ollama Proxy Debug Interface</h1>
  189. <div class="config">
  190. <strong>Configuration:</strong><br>
  191. Ollama Endpoint: <strong>{{ endpoint }}</strong><br>
  192. Model: <strong>{{ model }}</strong><br>
  193. Video Format: <strong>{{ video_format }}</strong>
  194. </div>
  195. <div class="status {{ status_class }}">
  196. <strong>Ollama Status:</strong> {{ status_message }}<br>
  197. {% if models %}
  198. <strong>Available Models:</strong> {{ models|join(', ') }}
  199. {% endif %}
  200. </div>
  201. <p><strong>Send requests to:</strong> http://localhost:5000/webhook</p>
  202. <button class="clear-btn" onclick="clearRequests()">Clear All</button>
  203. <div id="requests">
  204. {% for req in requests %}
  205. <div class="request">
  206. <div>
  207. <span class="method">{{ req.method }}</span>
  208. <span class="timestamp">{{ req.timestamp }}</span>
  209. {% if req.forwarded %}
  210. <span class="forwarded">SENT TO OLLAMA</span>
  211. {% endif %}
  212. {% if req.error %}
  213. <span class="error-badge">ERROR</span>
  214. {% endif %}
  215. </div>
  216. <div><strong>Path:</strong> {{ req.path }}</div>
  217. {% if req.body %}
  218. <div><strong>Incoming (Gemini Format):</strong></div>
  219. <pre>{{ req.body }}</pre>
  220. {% endif %}
  221. {% if req.ollama_request %}
  222. <div><strong>Sent to Ollama:</strong></div>
  223. <pre>{{ req.ollama_request }}</pre>
  224. {% endif %}
  225. {% if req.ollama_response %}
  226. <div><strong>Ollama Raw Response:</strong></div>
  227. <pre>{{ req.ollama_response }}</pre>
  228. {% endif %}
  229. {% if req.response %}
  230. <div><strong>Final Response (Gemini Format):</strong></div>
  231. <pre>{{ req.response }}</pre>
  232. {% endif %}
  233. {% if req.error %}
  234. <div><strong>Error Details:</strong></div>
  235. <pre style="color: red;">{{ req.error }}</pre>
  236. {% endif %}
  237. </div>
  238. {% endfor %}
  239. </div>
  240. </body>
  241. </html>
  242. """
  243. @app.route("/")
  244. def index():
  245. connected, models = test_ollama_connection()
  246. status_class = "connected" if connected else "error"
  247. status_message = "Connected ✅" if connected else "Cannot connect ❌"
  248. return render_template_string(
  249. HTML_TEMPLATE,
  250. requests=reversed(recent_requests),
  251. endpoint=OLLAMA_ENDPOINT,
  252. model=OLLAMA_MODEL,
  253. video_format=VIDEO_FORMAT,
  254. status_class=status_class,
  255. status_message=status_message,
  256. models=models,
  257. )
  258. @app.route("/webhook", methods=["POST", "PUT", "PATCH"], defaults={"subpath": ""})
  259. @app.route("/webhook/<path:subpath>", methods=["POST", "PUT", "PATCH"])
  260. def webhook(subpath):
  261. """Accept requests and forward to Ollama /api/generate"""
  262. timestamp = datetime.now().strftime("%Y-%m-%d %H:%M:%S")
  263. full_path = request.full_path if request.query_string else request.path
  264. print(f"\n{'='*60}")
  265. print(f"[{timestamp}] INCOMING {request.method} {full_path}")
  266. print(f"{'='*60}")
  267. # Get request data
  268. try:
  269. gemini_request = request.get_json() if request.is_json else {}
  270. body_display = (
  271. json.dumps(gemini_request, indent=2) if gemini_request else "No JSON body"
  272. )
  273. except Exception as e:
  274. body_display = f"Error parsing JSON: {e}"
  275. gemini_request = {}
  276. print(f"Request body: {body_display[:300]}...")
  277. # Store request info
  278. req_info = {
  279. "timestamp": timestamp,
  280. "method": request.method,
  281. "path": full_path,
  282. "body": body_display,
  283. "forwarded": False,
  284. "response": None,
  285. "error": None,
  286. "ollama_request": None,
  287. "ollama_response": None,
  288. }
  289. try:
  290. if gemini_request:
  291. print(f"\n🔄 Converting to Ollama format...")
  292. # Convert to Ollama format
  293. ollama_request = convert_gemini_to_ollama_generate(gemini_request)
  294. req_info["ollama_request"] = json.dumps(ollama_request, indent=2)
  295. print(f"Ollama request: {json.dumps(ollama_request, indent=2)}")
  296. # Send to Ollama
  297. print(f"\n📤 Sending to Ollama: {OLLAMA_ENDPOINT}/api/generate")
  298. response = requests.post(
  299. f"{OLLAMA_ENDPOINT}/api/generate",
  300. json=ollama_request,
  301. headers={"Content-Type": "application/json"},
  302. timeout=120,
  303. )
  304. print(f"📥 Ollama response status: {response.status_code}")
  305. if response.status_code == 200:
  306. ollama_response = response.json()
  307. req_info["ollama_response"] = json.dumps(ollama_response, indent=2)
  308. print(f"✅ Ollama responded successfully")
  309. # Convert back to Gemini format
  310. gemini_response = convert_ollama_generate_to_gemini(ollama_response)
  311. if "error" in gemini_response:
  312. req_info["error"] = gemini_response["error"]
  313. req_info["response"] = json.dumps(gemini_response, indent=2)
  314. else:
  315. req_info["forwarded"] = True
  316. req_info["response"] = json.dumps(gemini_response, indent=2)
  317. print(f"✅ Conversion successful")
  318. recent_requests.append(req_info)
  319. if len(recent_requests) > MAX_REQUESTS:
  320. recent_requests.pop(0)
  321. return jsonify(gemini_response), 200
  322. else:
  323. error_text = response.text
  324. error_msg = f"Ollama returned {response.status_code}: {error_text}"
  325. print(f"❌ {error_msg}")
  326. req_info["error"] = error_msg
  327. req_info["ollama_response"] = error_text
  328. req_info["forwarded"] = True
  329. recent_requests.append(req_info)
  330. if len(recent_requests) > MAX_REQUESTS:
  331. recent_requests.pop(0)
  332. return (
  333. jsonify(
  334. {
  335. "error": {
  336. "message": error_text,
  337. "status": response.status_code,
  338. }
  339. }
  340. ),
  341. response.status_code,
  342. )
  343. else:
  344. req_info["error"] = "No JSON body received"
  345. recent_requests.append(req_info)
  346. if len(recent_requests) > MAX_REQUESTS:
  347. recent_requests.pop(0)
  348. return (
  349. jsonify({"status": "error", "message": "No JSON body to process"}),
  350. 400,
  351. )
  352. except Exception as e:
  353. error_msg = f"Exception: {str(e)}"
  354. print(f"❌ {error_msg}")
  355. traceback.print_exc()
  356. req_info["error"] = error_msg
  357. recent_requests.append(req_info)
  358. if len(recent_requests) > MAX_REQUESTS:
  359. recent_requests.pop(0)
  360. return jsonify({"error": {"message": error_msg}}), 500
  361. @app.route("/clear", methods=["POST"])
  362. def clear():
  363. """Clear all stored requests"""
  364. recent_requests.clear()
  365. return jsonify({"status": "cleared"}), 200
  366. @app.route("/test")
  367. def test_endpoint():
  368. """Test endpoint to verify Ollama connection"""
  369. connected, models = test_ollama_connection()
  370. return jsonify(
  371. {"connected": connected, "endpoint": OLLAMA_ENDPOINT, "models": models}
  372. )
  373. if __name__ == "__main__":
  374. print("🦙 Ollama Proxy Server Starting...")
  375. print(f"🔗 Ollama Endpoint: {OLLAMA_ENDPOINT}")
  376. print(f"🤖 Model: {OLLAMA_MODEL}")
  377. print(f"📹 Video Format: {VIDEO_FORMAT}")
  378. print("📍 Web UI: http://localhost:5000")
  379. print("📮 Webhook: http://localhost:5000/webhook")
  380. print("🧪 Test: http://localhost:5000/test")
  381. # Test connection on startup
  382. connected, models = test_ollama_connection()
  383. if not connected:
  384. print("\n⚠️ WARNING: Cannot connect to Ollama!")
  385. print("Please check:")
  386. print("1. Is Ollama running? (ollama serve)")
  387. print("2. Is it on the right port?")
  388. print("3. Set OLLAMA_ENDPOINT env var if different")
  389. app.run(host="0.0.0.0", port=5000, debug=True)