/ no_install_orchestrator.py
no_install_orchestrator.py
1 #!/usr/bin/env python3 2 """ 3 NO-INSTALL SWARM ORCHESTRATOR 4 Works on Kali without installing packages 5 """ 6 7 import subprocess 8 import sys 9 import json 10 import os 11 from datetime import datetime 12 13 # API Keys (already have) 14 GEMINI_KEY = "AIzaSyC9g4B4sY9xeaUntjNmN2MeWFyp5gL3_EM" 15 GROQ_KEY = "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" 16 17 def query_with_curl(api_name, prompt): 18 """Use curl to call APIs (no Python packages needed)""" 19 20 if api_name == "gemini": 21 # Try multiple Gemini endpoints 22 endpoints = [ 23 "https://generativelanguage.googleapis.com/v1beta/models/gemini-1.5-flash:generateContent", 24 "https://generativelanguage.googleapis.com/v1beta/models/gemini-pro:generateContent", 25 "https://generativelanguage.googleapis.com/v1/models/gemini-1.0-pro:generateContent" 26 ] 27 28 for endpoint in endpoints: 29 url = f"{endpoint}?key={GEMINI_KEY}" 30 cmd = [ 31 'curl', '-s', '-X', 'POST', 32 '-H', 'Content-Type: application/json', 33 '-d', json.dumps({ 34 "contents": [{ 35 "parts": [{"text": prompt}] 36 }], 37 "generationConfig": { 38 "maxOutputTokens": 200, 39 "temperature": 0.7 40 } 41 }), 42 url 43 ] 44 45 try: 46 result = subprocess.run(cmd, capture_output=True, text=True, timeout=10) 47 if result.returncode == 0 and '"text":' in result.stdout: 48 data = json.loads(result.stdout) 49 text = data.get('candidates', [{}])[0].get('content', {}).get('parts', [{}])[0].get('text', '') 50 if text: 51 return f"[Gemini] {text[:200]}..." 52 except: 53 continue 54 55 return "[Gemini: No working endpoint]" 56 57 elif api_name == "groq": 58 # Groq API via curl 59 cmd = [ 60 'curl', '-s', '-X', 'POST', 61 'https://api.groq.com/openai/v1/chat/completions', 62 '-H', f'Authorization: Bearer {GROQ_KEY}', 63 '-H', 'Content-Type: application/json', 64 '-d', json.dumps({ 65 "model": "llama3-8b-8192", 66 "messages": [{"role": "user", "content": prompt}], 67 "max_tokens": 200, 68 "temperature": 0.7 69 }) 70 ] 71 72 try: 73 result = subprocess.run(cmd, capture_output=True, text=True, timeout=10) 74 if result.returncode == 0: 75 data = json.loads(result.stdout) 76 text = data.get('choices', [{}])[0].get('message', {}).get('content', '') 77 return f"[Groq] {text[:200]}..." 78 except: 79 return "[Groq: Error]" 80 81 return f"[{api_name}: Not implemented]" 82 83 def main(): 84 print("š¤ NO-INSTALL AI ORCHESTRATOR") 85 print("=" * 50) 86 87 # Get task 88 if len(sys.argv) > 1: 89 task = " ".join(sys.argv[1:]) 90 else: 91 print("\nEnter your question:") 92 task = sys.stdin.readline().strip() or "Explain AI in simple terms" 93 94 print(f"\nš Task: {task}") 95 print("\nš Querying AI services...") 96 97 # Create subtasks 98 subtasks = [ 99 f"Explain: {task}", 100 f"Give examples of: {task}", 101 f"List key points about: {task}" 102 ] 103 104 results = [] 105 106 # Query each service 107 print("\n1. Querying Google Gemini...") 108 results.append(query_with_curl("gemini", subtasks[0])) 109 110 print("2. Querying Groq...") 111 results.append(query_with_curl("groq", subtasks[1])) 112 113 print("3. Querying Hugging Face...") 114 # Hugging Face via simple model 115 try: 116 hf_cmd = [ 117 'curl', '-s', '-X', 'POST', 118 'https://api-inference.huggingface.co/models/gpt2', 119 '-H', f'Authorization: Bearer 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', 120 '-H', 'Content-Type: application/json', 121 '-d', json.dumps({ 122 "inputs": subtasks[2], 123 "parameters": {"max_new_tokens": 100} 124 }) 125 ] 126 hf_result = subprocess.run(hf_cmd, capture_output=True, text=True, timeout=10) 127 if hf_result.returncode == 0: 128 data = json.loads(hf_result.stdout) 129 if isinstance(data, list): 130 hf_text = data[0].get('generated_text', '')[:200] 131 results.append(f"[Hugging Face] {hf_text}...") 132 else: 133 results.append("[Hugging Face: Model loading]") 134 else: 135 results.append("[Hugging Face: Error]") 136 except: 137 results.append("[Hugging Face: Failed]") 138 139 # Display results 140 print("\n" + "=" * 50) 141 print("š AI RESPONSES") 142 print("=" * 50) 143 144 for i, result in enumerate(results, 1): 145 print(f"\n{i}. {result}") 146 147 # Save to file 148 timestamp = datetime.now().strftime("%Y%m%d_%H%M%S") 149 filename = f"ai_results_{timestamp}.txt" 150 151 with open(filename, 'w') as f: 152 f.write(f"Task: {task}\n") 153 f.write(f"Time: {datetime.now()}\n\n") 154 for i, result in enumerate(results, 1): 155 f.write(f"{i}. {result}\n\n") 156 157 print(f"\nš¾ Results saved to: {filename}") 158 print("=" * 50) 159 160 if __name__ == "__main__": 161 main()