/ no_install_orchestrator.py
no_install_orchestrator.py
  1  #!/usr/bin/env python3
  2  """
  3  NO-INSTALL SWARM ORCHESTRATOR
  4  Works on Kali without installing packages
  5  """
  6  
  7  import subprocess
  8  import sys
  9  import json
 10  import os
 11  from datetime import datetime
 12  
 13  # API Keys (already have)
 14  GEMINI_KEY = "AIzaSyC9g4B4sY9xeaUntjNmN2MeWFyp5gL3_EM"
 15  GROQ_KEY = "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"
 16  
 17  def query_with_curl(api_name, prompt):
 18      """Use curl to call APIs (no Python packages needed)"""
 19      
 20      if api_name == "gemini":
 21          # Try multiple Gemini endpoints
 22          endpoints = [
 23              "https://generativelanguage.googleapis.com/v1beta/models/gemini-1.5-flash:generateContent",
 24              "https://generativelanguage.googleapis.com/v1beta/models/gemini-pro:generateContent",
 25              "https://generativelanguage.googleapis.com/v1/models/gemini-1.0-pro:generateContent"
 26          ]
 27          
 28          for endpoint in endpoints:
 29              url = f"{endpoint}?key={GEMINI_KEY}"
 30              cmd = [
 31                  'curl', '-s', '-X', 'POST',
 32                  '-H', 'Content-Type: application/json',
 33                  '-d', json.dumps({
 34                      "contents": [{
 35                          "parts": [{"text": prompt}]
 36                      }],
 37                      "generationConfig": {
 38                          "maxOutputTokens": 200,
 39                          "temperature": 0.7
 40                      }
 41                  }),
 42                  url
 43              ]
 44              
 45              try:
 46                  result = subprocess.run(cmd, capture_output=True, text=True, timeout=10)
 47                  if result.returncode == 0 and '"text":' in result.stdout:
 48                      data = json.loads(result.stdout)
 49                      text = data.get('candidates', [{}])[0].get('content', {}).get('parts', [{}])[0].get('text', '')
 50                      if text:
 51                          return f"[Gemini] {text[:200]}..."
 52              except:
 53                  continue
 54          
 55          return "[Gemini: No working endpoint]"
 56      
 57      elif api_name == "groq":
 58          # Groq API via curl
 59          cmd = [
 60              'curl', '-s', '-X', 'POST',
 61              'https://api.groq.com/openai/v1/chat/completions',
 62              '-H', f'Authorization: Bearer {GROQ_KEY}',
 63              '-H', 'Content-Type: application/json',
 64              '-d', json.dumps({
 65                  "model": "llama3-8b-8192",
 66                  "messages": [{"role": "user", "content": prompt}],
 67                  "max_tokens": 200,
 68                  "temperature": 0.7
 69              })
 70          ]
 71          
 72          try:
 73              result = subprocess.run(cmd, capture_output=True, text=True, timeout=10)
 74              if result.returncode == 0:
 75                  data = json.loads(result.stdout)
 76                  text = data.get('choices', [{}])[0].get('message', {}).get('content', '')
 77                  return f"[Groq] {text[:200]}..."
 78          except:
 79              return "[Groq: Error]"
 80      
 81      return f"[{api_name}: Not implemented]"
 82  
 83  def main():
 84      print("šŸ¤– NO-INSTALL AI ORCHESTRATOR")
 85      print("=" * 50)
 86      
 87      # Get task
 88      if len(sys.argv) > 1:
 89          task = " ".join(sys.argv[1:])
 90      else:
 91          print("\nEnter your question:")
 92          task = sys.stdin.readline().strip() or "Explain AI in simple terms"
 93      
 94      print(f"\nšŸ“ Task: {task}")
 95      print("\nšŸ”„ Querying AI services...")
 96      
 97      # Create subtasks
 98      subtasks = [
 99          f"Explain: {task}",
100          f"Give examples of: {task}",
101          f"List key points about: {task}"
102      ]
103      
104      results = []
105      
106      # Query each service
107      print("\n1. Querying Google Gemini...")
108      results.append(query_with_curl("gemini", subtasks[0]))
109      
110      print("2. Querying Groq...")
111      results.append(query_with_curl("groq", subtasks[1]))
112      
113      print("3. Querying Hugging Face...")
114      # Hugging Face via simple model
115      try:
116          hf_cmd = [
117              'curl', '-s', '-X', 'POST',
118              'https://api-inference.huggingface.co/models/gpt2',
119              '-H', f'Authorization: Bearer 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',
120              '-H', 'Content-Type: application/json',
121              '-d', json.dumps({
122                  "inputs": subtasks[2],
123                  "parameters": {"max_new_tokens": 100}
124              })
125          ]
126          hf_result = subprocess.run(hf_cmd, capture_output=True, text=True, timeout=10)
127          if hf_result.returncode == 0:
128              data = json.loads(hf_result.stdout)
129              if isinstance(data, list):
130                  hf_text = data[0].get('generated_text', '')[:200]
131                  results.append(f"[Hugging Face] {hf_text}...")
132              else:
133                  results.append("[Hugging Face: Model loading]")
134          else:
135              results.append("[Hugging Face: Error]")
136      except:
137          results.append("[Hugging Face: Failed]")
138      
139      # Display results
140      print("\n" + "=" * 50)
141      print("šŸ“Š AI RESPONSES")
142      print("=" * 50)
143      
144      for i, result in enumerate(results, 1):
145          print(f"\n{i}. {result}")
146      
147      # Save to file
148      timestamp = datetime.now().strftime("%Y%m%d_%H%M%S")
149      filename = f"ai_results_{timestamp}.txt"
150      
151      with open(filename, 'w') as f:
152          f.write(f"Task: {task}\n")
153          f.write(f"Time: {datetime.now()}\n\n")
154          for i, result in enumerate(results, 1):
155              f.write(f"{i}. {result}\n\n")
156      
157      print(f"\nšŸ’¾ Results saved to: {filename}")
158      print("=" * 50)
159  
160  if __name__ == "__main__":
161      main()